CN110569901B - Channel selection-based countermeasure elimination weak supervision target detection method - Google Patents

Channel selection-based countermeasure elimination weak supervision target detection method Download PDF

Info

Publication number
CN110569901B
CN110569901B CN201910838283.XA CN201910838283A CN110569901B CN 110569901 B CN110569901 B CN 110569901B CN 201910838283 A CN201910838283 A CN 201910838283A CN 110569901 B CN110569901 B CN 110569901B
Authority
CN
China
Prior art keywords
feature
network
image
region
characteristic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910838283.XA
Other languages
Chinese (zh)
Other versions
CN110569901A (en
Inventor
杨金福
单义
李明爱
武随烁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201910838283.XA priority Critical patent/CN110569901B/en
Publication of CN110569901A publication Critical patent/CN110569901A/en
Application granted granted Critical
Publication of CN110569901B publication Critical patent/CN110569901B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a method for detecting a countermeasure elimination weak supervision target based on feature channel selection, which is used for solving the problem of detection and positioning errors of the weak supervision target. Firstly, using weak supervision depth target detection as a bottom-layer framework, generating a candidate frame on training set data by adopting a selective search method, and using the candidate frame, a training set image and a corresponding image label as input of a weak supervision network; secondly, constructing a feature extraction network model by taking VGG16 as a basic network, and carrying out channel weighting selection on the obtained feature images in a feature channel compression mode, so as to excite image feature layers beneficial to classification and inhibit feature layers with interference on classification; then, obtaining a complete characteristic expression capable of expressing the image target by adopting a countermeasure elimination method as the input of a prediction network; and finally, training a prediction network according to the multitask cross entropy loss to realize target detection. The invention not only can more accurately position the position of the target object, but also can improve the accuracy of object identification.

Description

Channel selection-based countermeasure elimination weak supervision target detection method
Technical Field
The invention belongs to the technical field of target detection, and introduces a weak supervision target detection method based on countermeasure elimination.
Background
With the development of science and technology and the improvement of the intelligence level of human life, mobile robots gradually enter human production and life, and are widely applied to various industries. The target detection based on the mobile robot is widely applied to the fields of routing inspection, security protection, video monitoring search and the like. The deep learning target detection algorithm at the present stage generally needs a large amount of manually labeled data sets, which not only wastes a large amount of manpower, material resources and financial resources, but also causes the wrong labeled data to influence the robustness of the model and the detection precision of the model. The target detection algorithm based on weak supervision can complete target detection only by the image-level label, so that the method is more suitable for industrial production application environments.
In recent years, target detection algorithms based on deep learning have been developed. The method for realizing deep learning target detection based on the image-level label is also rapidly developed. 2016, bilen [1] The application of deep learning to the target detection algorithm of the image-level labels is firstly proposed by the people, and compared with the traditional weak supervision target detection algorithm, the detection precision on the public data set is obviously improved. 2016, diba [2] Et al propose a cascaded deep learning weak supervised target detection algorithm. The model is divided into two steps: (1) A convolutional neural network based on a first segment of the cascade network, for extracting a candidate region where a target object may exist; (2) And taking the candidate region output by the first section of the cascade network as the input of the second section of the cascade network, and finally detecting the target object by a multi-example learning method. 2017, tang [3] Et al propose a multi-instance loop-iterated deep-learning target detection network. Aiming at the problem that in the prediction stage of a target detection frame, after a non-maximum suppression strategy (NMS), the network inhibits the detection frame with lower classification score and accurate positioning, and retains the target with high target prediction classification score and poor positioning, the network provides a loop iteration method, and gradually iteratively updates the classification confidence coefficient of the object and the position of the detection frame. 2018, wang [4] The inventors propose a method for detecting a weakly supervised target by cooperative learning, which improves the detection accuracy by combining a strongly supervised learning network and a weakly supervised learning network. Using detection frame obtained from weak supervision target detection network as strong supervision learning networkInitializing input, reducing network parameters by utilizing a shared convolutional network, and updating the input of a strong supervision network through continuous cyclic training to improve the final precision. However, the position of the detection frame needs to be continuously updated and adjusted at the input end of the network model in the above network, and since the position regression is lacked in the weak supervision target detection compared with the strong supervision network, the problem of large target positioning error in the weak supervision target detection cannot be completely solved by continuously iterative updating.
The optimization method for the weak supervision target detection is a method for iterating the predicted candidate frame after the feature map of the target is obtained so as to optimize the position of the target detection frame. However, due to the lack of accurate position label information of the target in the weak supervision target detection, the problem of inaccurate positioning of the target detection frame is not well solved only by iterative optimization.
Reference:
1.H.Bilen and A.Vedaldi.Weakly supervised deep detection networks. In Proc.IEEE Int.Conf.Vis.Pattern Recognition.(CVPR),Pages 2846-2854,2016.
2.A.Diba,V.Sharma,A.Pazandeh,H.Pirsiavash,and L.Van Gool. Weakly supervised cascaded convolutional networks.In Proc.IEEE Int.Conf.Computer.Vis.Pattern Recognition.(CVPR),pages 5131-5139,2017.
3.P.Tang.X.Wang,X.Bai,and W.Liu.Multiple instance detection network with online instance classifier refinement.In Proc.IEEE Int. Conf.Computer.Vis.Pattern Recognition.(CVPR),pages 3059-3067, 2017.
4.J.Wang,J.Yao,Y.Zhang,et al.Collaborative Learning for Weakly Supervised Object Detection.IJCAI,2018.
disclosure of Invention
In order to overcome the defects in the prior art, the invention provides a method for detecting an confrontation elimination target based on feature channel selection. In the feature extraction network, a feature channel selection strategy is utilized to improve the weight of feature channels with forward promotion on classification, and then feature countermeasure elimination is utilized to extract more comprehensive features of each detected object, so that the problem of weak supervision target detection positioning error is solved. The model only uses image data with image level labels as the input of the network, and combines a channel feature selection method to extract features more useful for classifying the network as the input of a classifier. The invention not only can more accurately position the position of the target object, but also can improve the accuracy of object identification.
The invention is realized by adopting the following technical means:
a method for detecting a countermeasure elimination weak supervision target based on feature channel selection is characterized by comprising the following steps:
step 1: data pre-processing
And taking the images in the training set as input images, and preprocessing the input images, namely performing multi-scale transformation, horizontal turning and random cutting on the input images to obtain a preprocessed training set.
And 2, step: generating a target region candidate box
And (3) generating a candidate box set R for each image to be trained through a selective search algorithm on the images in the training set, namely generating a series of rectangular areas in which the target object may exist in the images. After the redundant candidate frames with the similarity and the contact ratio larger than a threshold value A (the value range is (0.5, 1)) are deleted, the coordinates of the remaining candidate frames are used as the input of the weak supervision target detection network to train the network model.
And 3, step 3: the whole network is divided into six parts: the system comprises a primary feature extraction network model, a salient region feature extraction network, a secondary region feature learning network, a comprehensive feature extraction layer, an interested region pooling layer and a global average pooling layer;
firstly, extracting the preliminary features of the image by using a preliminary feature extraction network model.
Among deep network models such as VGGNet, resNet, mobileNet and the like, a VGG16 model is preferably used as a preliminary feature extraction network model. And (2) taking the preprocessed training set obtained in the step (1) as the input of a primary feature extraction network, training a primary feature extraction network model, modifying a primary feature extraction network model structure file and a network parameter configuration file, finishing primary extraction of image features, and obtaining a primary feature extraction image.
And 4, step 4: and selecting a characteristic channel for the preliminary characteristic graph by using a salient region characteristic extraction network to obtain salient region characteristics.
Step 4.1: for the extracted H × W × C in step 3 1 Performing convolution operation on the preliminary feature map by 1 × 1 to obtain a feature image u with the size of H × W × C c . The transformation method is shown in formula (1)
Figure BDA0002192880920000041
Wherein u is c Is the output after the convolution is performed and,
Figure BDA0002192880920000042
which represents the kernel of a convolution with the original,
Figure BDA0002192880920000043
a feature map representing the input. H × W is the size of the input feature map, C 1 The number of the feature map channels before the convolution transformation is C, and the number of the feature map channels after the convolution transformation is C.
Step 4.2: output u from step 4.1 c Feature compression using global average pooling outputs a tensor z of size 1 × 1 × C c . Equation (2) for global average pooling is as follows:
Figure BDA0002192880920000044
wherein z is c ∈R c Is the feature tensor with output of 1 × 1 × C, u c Is the output of step 4.1, F sq The global average pooling operation is performed on the feature map, H multiplied by W represents the size of an input feature layer, and i, j are element coordinate indexes in the feature map.
Step 4.3: the output z in step 4.2 is compared c And performing characteristic channel selection operation to obtain a one-dimensional tensor s. The formula for channel selection is as follows:
s=Fex(w 2 δ(w 1 z c )) (3)
wherein s is a one-dimensional tensor output after network compression, W 1 ,W 2 And a weight parameter which needs to be continuously and iteratively optimized for the channel selection operation, wherein delta is a ReLU activation function, and Fex represents a nonlinear activation function.
Step 4.4: the feature vector s with the parameter weight obtained in the step 4.3 and the feature map u obtained in the step 4.1 are combined c And (3) multiplying corresponding feature maps to obtain a feature map x with weight distribution, wherein a specific calculation formula is as follows:
x=F scale (u c ,s) (4)
wherein, F scale Representing a eigenchannel multiplication. s represents the scalar obtained in step 4.3. u. u c As output in step 4.1.
Step 4.5: obtaining a salient region feature image M from the feature image x in the step 4.4 through a classifier of a salient region feature extraction network model A
M A =f(φ A ,x,y i ) (5)
Wherein M is A Representing the obtained characteristic region, f representing the classifier of the salient region characteristic extraction network model, preferably SDG phi A Representing the parameters of the network model, x being the characteristic diagram with weight distribution obtained in step 4.4, y i Is the label corresponding to the image.
And 5: output M of step 4.5 A Learning features other than the most salient region using a countervailing method
Figure BDA0002192880920000051
And obtaining a secondary feature map M through a secondary regional feature learning network B The method comprises the following steps:
step 5.1: output M of step 4.5 A By setting a threshold value t, the value range is (0.5, 1) to obtain the characteristic region R of the most significant part t
And step 5.2: the characteristic region R t Corresponding features are classified as 0, and a feature map with the removed features is obtained
Figure BDA0002192880920000052
Step 5.3: feature map obtained in step 5.2 is compared with a secondary area feature extraction network
Figure BDA0002192880920000053
And (4) further learning to obtain the features in the image, wherein the secondary region feature learning network and the salient region feature extraction network in the step 4 have the same network structure, but the input of the network is different, and the network parameters are not shared. Obtaining a secondary characteristic feature map M through a secondary region characteristic extraction network classifier B
Step 6: the salient region feature map M of the step 4.5 is processed A And the secondary region feature map M obtained by learning in the step 5.3 B A method for solving the maximum value of the corresponding pixel point to obtain a fused comprehensive characteristic graph M fuse
M fuse =max(M A ,M B ) (6)
Wherein M is fuse To obtain a comprehensive characteristic map after fusion. .
And 7: mapping the candidate region frame obtained in the step 2 to the comprehensive characteristic map M obtained in the step 6 fuse In the method, the interested region pooling is carried out on each candidate frame, and the candidate frames with different sizes are converted into the feature vector V with the same dimension c 。,
And step 8: corresponding the target frame to the feature vector V c After twice Global Average Pooling (GAP), obtaining probability values of corresponding categories by utilizing a softmax function;
and step 9: and training the whole network model through a total loss function L, wherein the input label of the training is a class label of the image in the data set. And reducing the difference between the predicted value and the data set label through continuous iterative training, and finishing the training when the model converges or reaches the preset training times.
Advantageous effects
In the existing weak supervision target detection method, the improvement of a target positioning frame is mostly based on iterative regression of the positioning frame, but the problem cannot be well solved because a network lacks a position label. The method for detecting the countermeasure elimination weak supervision target based on the feature selection improves the positioning problem of the target frame from the aspect of optimizing the feature extraction of the image, enables a network model to extract more comprehensive features of the target object, enables the subsequent positioning frame to completely cover the detection target, and can improve the accuracy of the network in classifying the object and improve the final detection positioning frame result.
Drawings
FIG. 1 is a general block diagram of a weakly supervised target detection approach based on countermeasure elimination;
FIG. 2 is a diagram of a basic feature extraction network architecture;
FIG. 3 is a diagram of a feature channel selection architecture;
FIG. 4 shows the results of part of the experiment.
Detailed Description
In order that those skilled in the art can better understand and use the present invention, the following technical solutions will be further described with reference to the accompanying drawings and specific embodiments.
1. The VOC data set was selected as the training and testing data set. Firstly, taking the images in the training set as input images, and preprocessing the input images, namely performing multi-scale transformation, horizontal turning and random cutting on the input images to obtain a preprocessed training set.
2. The images in the training set are subjected to a selective search algorithm to generate a candidate box set R, i.e., rectangular coordinates of the positions where the object may exist. After the redundant candidate boxes with the similarity and the coincidence degree larger than the set threshold value of 0.7 are deleted, the remaining candidate boxes are used as the input of the weak supervision target detection network to train the model. The overall network structure is shown in figure 1.
3. Among deep network models such as VGGNet, resNet, mobileNet and the like, a VGG16 model is preferably used as a preliminary feature extraction network model. The VGG network comprises 13 convolutional layers and four pooling layers, and the detailed network structure is shown in figure 2. And taking the preprocessed training set as the input of the convolutional neural network, training the convolutional neural network, further modifying the network structure file and the network parameter configuration file of the convolutional neural network, and finishing the extraction of the image characteristics.
4. The salient region feature extraction network model is shown in fig. 3, and includes 1 convolutional layer, 1 averaging pooling layer and weight distribution layer, and the detailed process is as follows. Carrying out convolution operation of 1 multiplied by 1 on the last layer of feature layer of the VGG network in 3 to change the number of channels of the feature map, and using the feature channel selection network to select the original H multiplied by W multiplied by C 1 The feature map of (2) is converted into a feature image of H × W × C. The parameters of the model are reduced by reducing the number of channels, so that the model is simplified and the efficiency of the model is improved. And performing feature compression on the feature map after convolution through a global average pooling layer, and outputting a tensor with the size of 1 × 1 × C. In order to distinguish the contributions of different feature image channels to the classification, the feature extraction capability is improved by learning the weight of each channel. And performing corresponding dot multiplication on the obtained weighted feature map with the parameters and the corresponding feature map. In this way, feature maps with different weights can be obtained.
5. Using the feature map in the step 4 to extract a network classifier by using the feature of the salient region to obtain a feature image M A . Obtaining the characteristic region R of the most significant part through a preset threshold value t t . By forming a characteristic region R t And (4) the corresponding characteristic is 0, and a characteristic graph after the characteristic is eliminated is obtained. The features in the graph are further learned by the feature map after elimination through the secondary regional feature extraction network, and the model structure of the secondary regional feature extraction network is the same as that of the salient region feature extraction network, as shown in fig. 3. But the input of the network is different and the parameters of the network model are not shared. Obtaining a feature image M by using a secondary regional feature network classifier B . Feature M before elimination by fusing features A To obtain the final characteristic diagram M fuse
6. Mapping the candidate region frame obtained in step 2 to the feature map M obtained in step 5 fuse Then, performing region-of-interest pooling operation on each candidate frame, and changing the candidate frames with different sizes into feature vectors with the same size. After two layers of tie pooling layers (GAP), obtaining corresponding classes through a softmax functionOther probability values.
7. Model training: the method comprises the steps that (1) data preprocessing (2) data are input into a network model for training (3), in the training process, the value of a loss function is continuously calculated to judge the error condition (4), the training error is reduced through reverse propagation (5), the training process is repeated until the model converges or reaches the set times, wherein a network overall loss function consists of three parts, namely a loss function of the significant region feature extraction network model classification, a classification loss function of a secondary region feature extraction network and the output classification loss of the whole in the final stage, and the loss function of the whole network is obtained through adding the three parts.
8. The training platform adopts NVIDIA GeForce GTX TITAN X GPU, the network building adopts a Pythrch frame, the batch-size is set to be 20, the initial learning rate is 0.001, the momentum is set to be 0.9, the weight attenuation is set to be 0.0005, and the optimization method adopts a gradient descent method.
9. For a given input image, firstly, extracting the features of different channels of the image by using a feature extraction network; and then weighting different characteristic channels through a characteristic selection network, enlarging the characteristic channels which are useful for classification by adopting a back propagation training model, and inhibiting the characteristic channels which interfere with the classification. Finally, the characteristic after the significant characteristic elimination is obtained by utilizing the characteristic countermeasure elimination network, the characteristics of the two parts are fused, and the model is trained by utilizing the total loss function
10. The results of some of the experiments of the present invention are shown in FIG. 4. From the results of fig. 4, it can be seen that for the object to be detected in the image, the positioning frame not only can well surround the detected object, but also has high confidence for the detected object.
Finally, it should be noted that: the above examples are only intended to illustrate the invention and do not limit the technical solutions described in the present invention. Therefore, although the present invention has been described in detail with reference to the above examples, it should be understood by those skilled in the art that the present invention may be modified and replaced with equivalents without departing from the spirit and scope of the invention, and all such modifications and improvements are intended to be included within the scope of the claims.

Claims (6)

1. A method for detecting an anti-elimination weak supervision target based on feature channel selection is characterized by comprising the following steps:
step 1: data pre-processing
Step 2: generating a target region candidate box
Generating a candidate frame set R for each preprocessed image through a selective search algorithm, namely generating a series of rectangular areas in which target objects possibly exist in the image, and only reserving candidate frames with similarity and contact ratio smaller than a threshold value A;
and step 3: an integral network model is constructed, and the model is divided into six parts: the system comprises a primary feature extraction network, a salient region feature extraction network, a secondary region feature learning network, a comprehensive feature extraction layer, an interested region pooling layer and a global average pooling layer;
the preliminary feature extraction network is used for extracting preliminary features of the image and is input into the data set preprocessed in the step 1;
the salient region feature extraction network is used for carrying out feature channel selection on the primary features of the image to obtain a salient region feature image M A
The secondary area feature learning network is used for obtaining a secondary feature map M of the image B And the network structure is the same as that of the salient region feature extraction network, and the difference is that the network parameters are not shared, and the input is M A Features learned by the countermeasure elimination method except for the most significant region
Figure FDA0002192880910000011
The comprehensive characteristic extraction layer is used for obtaining a comprehensive characteristic graph M fuse The method specifically comprises the following steps: characterizing the salient region to M A And a secondary region feature map M B Solving the maximum value of the corresponding pixel point to obtain a fused comprehensive characteristic graph M fuse
M fuse =max(M A ,M B ) (6)
Wherein,M fuse Obtaining a comprehensive characteristic diagram after fusion;
the region of interest pooling layer is used for mapping the candidate region frame obtained in the step 2 to a comprehensive characteristic map M fuse In the method, the interested region pooling is carried out on each candidate frame, and the candidate frames with different sizes are converted into the feature vector V with the same dimension c
The global average pooling layer (GAP) is used for corresponding the target box to the feature vector V c After twice Global Average Pooling (GAP), obtaining probability values of corresponding categories by utilizing a softmax function;
step 7, training the whole network model
Training the class probability predicted by the overall network model through an overall loss function L, training the overall network model through the back propagation of the difference value of the label value of the training data and the predicted value of the overall network model, and finishing the training when the model is converged;
and 8, acquiring data to be detected, and performing target detection by using the trained integral network model after the data to be detected is processed in the steps 1 and 2.
2. The method for detecting the countermeasure elimination weakly supervised target based on the feature channel selection as recited in claim 1, wherein: the pretreatment specifically comprises the following steps: and carrying out multi-scale transformation, horizontal turning and random cropping on the input image.
3. The method for detecting the confrontation elimination weak supervision target based on the feature channel selection as claimed in claim 1, characterized in that: the preliminary feature extraction network model is preferably a VGG16 model.
4. The method for detecting the confrontation elimination weak supervision target based on the feature channel selection as claimed in claim 1, characterized in that: the salient region feature extraction network is specifically as follows:
step 4.1: extracting H x W x C of network for preliminary feature extraction 1 Is subjected to a convolution operation of 1 x 1 to obtain a size ofH × W × C feature image u c The transformation method is shown in formula (1)
Figure FDA0002192880910000021
Wherein u is c Is the output after the convolution and is,
Figure FDA0002192880910000022
which represents the kernel of the convolution,
Figure FDA0002192880910000023
representing the input feature map, H × W is the size of the input feature map, C 1 The number of the characteristic image channels before the convolution transformation is C, and the number of the characteristic image channels after the convolution transformation is C;
step 4.2: the output u in step 4.1 is compared c Feature compression using global average pooling outputs a tensor z of size 1 × 1 × C c Equation (2) for global average pooling is as follows:
Figure FDA0002192880910000024
wherein z is c ∈R c Is the feature tensor with output of 1 × 1 × C, u c Is the output of step 4.1, F sq Representing the global average pooling operation of the feature map, H multiplied by W represents the size of an input feature layer, and i, j are element coordinate indexes in the feature map;
step 4.3: the output z in step 4.2 is compared c And performing characteristic channel selection operation to obtain a one-dimensional tensor s, wherein the formula of channel selection is as follows:
s=Fex(w 2 δ(w 1 z c )) (3)
wherein s is a one-dimensional tensor output after network compression, W 1 ,W 2 Weight parameters requiring continuous iterative optimization for channel selection operation, delta is ReLU activation function, fex represents nonlinear activationA function;
step 4.4: the feature vector s with the parameter weight obtained in the step 4.3 and the feature map u obtained in the step 4.1 are combined c And (3) multiplying corresponding feature maps to obtain a feature map x with weight distribution, wherein a specific calculation formula is as follows:
x=F scale (u c ,s) (4)
wherein, F scale Representing the eigen-channel multiplication, s representing the scalar obtained in step 4.3, u c Is the output of step 4.1;
step 4.5: obtaining a salient region feature image M from the feature image x in the step 4.4 through a classifier of a salient region feature extraction network model A
M A =f(φ A ,x,y i ) (5)
Wherein M is A The obtained characteristic region is represented, f represents a classifier of a salient region characteristic extraction network model, phi A Representing the parameters of the network model, x being the characteristic diagram with weight distribution obtained in step 4.4, y i Is the label corresponding to the image.
5. The method for detecting the countermeasure elimination weakly supervised target based on the feature channel selection as recited in claim 1, wherein: the characteristic diagram
Figure FDA0002192880910000036
Obtained by the following steps:
characterizing the salient region to image M A Obtaining the characteristic region R of the most significant part by setting a threshold value t t (ii) a Then the characteristic region R is divided into t Corresponding features are classified as 0 to obtain a feature map
Figure FDA0002192880910000037
6. The method for detecting the countermeasure elimination weakly supervised target based on the feature channel selection as recited in claim 1, wherein: the overall loss function L is specifically as follows:
Figure FDA0002192880910000031
wherein, y i Is the value of the corresponding label in the data set, phi c Predicting probability distribution for the final class label of the whole network model;
Figure FDA0002192880910000032
L A extracting loss function of network for salient region feature, y i Is the value of the corresponding tag in the data set,
Figure FDA0002192880910000033
representing the probability distribution of the corresponding labels predicted by the salient region feature network, wherein c is the number of the categories in the data set;
Figure FDA0002192880910000034
L B extracting the loss function of the network for the secondary area features, y i Is the value of the corresponding tag in the data set,
Figure FDA0002192880910000035
and c is the number of categories in the data set.
CN201910838283.XA 2019-09-05 2019-09-05 Channel selection-based countermeasure elimination weak supervision target detection method Active CN110569901B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910838283.XA CN110569901B (en) 2019-09-05 2019-09-05 Channel selection-based countermeasure elimination weak supervision target detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910838283.XA CN110569901B (en) 2019-09-05 2019-09-05 Channel selection-based countermeasure elimination weak supervision target detection method

Publications (2)

Publication Number Publication Date
CN110569901A CN110569901A (en) 2019-12-13
CN110569901B true CN110569901B (en) 2022-11-29

Family

ID=68777935

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910838283.XA Active CN110569901B (en) 2019-09-05 2019-09-05 Channel selection-based countermeasure elimination weak supervision target detection method

Country Status (1)

Country Link
CN (1) CN110569901B (en)

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111031312B (en) * 2019-12-16 2022-02-25 山东浪潮科学研究院有限公司 Image compression method for realizing attention mechanism based on network
CN111340096A (en) * 2020-02-24 2020-06-26 北京工业大学 Weakly supervised butterfly target detection method based on confrontation complementary learning
CN111723859A (en) * 2020-06-16 2020-09-29 江西中科九峰智慧医疗科技有限公司 Target positioning method and system based on weak tags
CN112150442A (en) * 2020-09-25 2020-12-29 帝工(杭州)科技产业有限公司 New crown diagnosis system based on deep convolutional neural network and multi-instance learning
CN112132856B (en) * 2020-09-30 2024-05-24 北京工业大学 Twin network tracking method based on self-adaptive template updating
CN112329793B (en) * 2020-11-05 2022-09-27 天津大学 Significance detection method based on structure self-adaption and scale self-adaption receptive fields
CN112686256B (en) * 2021-01-05 2023-10-20 中山大学 Training-free and unsupervised target co-location method, system and device
CN113158760B (en) * 2021-02-20 2023-08-15 温州大学 Weak supervision specific behavior recognition system based on multi-level labels
US20220309651A1 (en) * 2021-03-24 2022-09-29 Ping An Technology (Shenzhen) Co., Ltd. Method, device, and storage medium for semi-supervised learning for bone mineral density estimation in hip x-ray images
CN113095235B (en) * 2021-04-15 2023-10-27 国家电网有限公司 Image target detection method, system and device based on weak supervision and discrimination mechanism
CN113435385A (en) * 2021-07-08 2021-09-24 网易(杭州)网络有限公司 Face detection model training method, face detection model recognition device, face detection medium and face detection equipment
CN113705361A (en) * 2021-08-03 2021-11-26 北京百度网讯科技有限公司 Method and device for detecting model in living body and electronic equipment
CN113838130B (en) * 2021-08-30 2023-07-18 厦门大学 Weak supervision target positioning method based on feature expansibility learning
CN114842320A (en) * 2022-03-17 2022-08-02 江南大学 Robot target detection method and system based on DW-SEnet model
CN117437395A (en) * 2022-07-15 2024-01-23 马上消费金融股份有限公司 Target detection model training method, target detection method and target detection device
CN115432331A (en) * 2022-10-10 2022-12-06 浙江绿达智能科技有限公司 Intelligent classification dustbin
CN116385825B (en) * 2023-03-22 2024-04-30 小米汽车科技有限公司 Model joint training method and device and vehicle

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019136946A1 (en) * 2018-01-15 2019-07-18 中山大学 Deep learning-based weakly supervised salient object detection method and system

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019136946A1 (en) * 2018-01-15 2019-07-18 中山大学 Deep learning-based weakly supervised salient object detection method and system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Adversarial Complementary Learning for Weakly Supervised Object Localization";Xiaolin Zhang 等;《2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition》;20181116;全文 *
"Weakly Supervised Cascaded Convolutional Networks";Ali Diba 等;《2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)》;20171109;全文 *
弱监督学习下的视觉显著性目标检测算法;李策等;《计算机工程与设计》;20170516(第05期);全文 *

Also Published As

Publication number Publication date
CN110569901A (en) 2019-12-13

Similar Documents

Publication Publication Date Title
CN110569901B (en) Channel selection-based countermeasure elimination weak supervision target detection method
CN111583263B (en) Point cloud segmentation method based on joint dynamic graph convolution
CN111369572B (en) Weak supervision semantic segmentation method and device based on image restoration technology
CN107784288B (en) Iterative positioning type face detection method based on deep neural network
CN116579616B (en) Risk identification method based on deep learning
CN115439458A (en) Industrial image defect target detection algorithm based on depth map attention
Zeng et al. Steel sheet defect detection based on deep learning method
CN116010813A (en) Community detection method based on influence degree of fusion label nodes of graph neural network
CN117131348B (en) Data quality analysis method and system based on differential convolution characteristics
CN116561562B (en) Sound source depth optimization acquisition method based on waveguide singular points
Bode et al. Bounded: Neural boundary and edge detection in 3d point clouds via local neighborhood statistics
CN116977265A (en) Training method and device for defect detection model, computer equipment and storage medium
CN113673534B (en) RGB-D image fruit detection method based on FASTER RCNN
CN115880477A (en) Apple detection positioning method and system based on deep convolutional neural network
CN115661539A (en) Less-sample image identification method embedded with uncertainty information
CN114943741A (en) Visual SLAM method based on target detection and geometric probability in dynamic scene
Li Investigation of you only look once networks for vision-based small object detection
CN115063831A (en) High-performance pedestrian retrieval and re-identification method and device
CN114462490A (en) Retrieval method, retrieval device, electronic device and storage medium of image object
CN116778362B (en) Electric power tower acceptance component identification method based on field knowledge and deep learning
CN112364892B (en) Image identification method and device based on dynamic model
Rao et al. Markov random field classification technique for plant leaf disease detection
CN117437234B (en) Aerial photo ground object classification and change detection method based on graph neural network
CN116129280B (en) Method for detecting snow in remote sensing image
CN117152746B (en) Method for acquiring cervical cell classification parameters based on YOLOV5 network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant