CN116385466B - Method and system for dividing targets in image based on boundary box weak annotation - Google Patents
Method and system for dividing targets in image based on boundary box weak annotation Download PDFInfo
- Publication number
- CN116385466B CN116385466B CN202310494738.7A CN202310494738A CN116385466B CN 116385466 B CN116385466 B CN 116385466B CN 202310494738 A CN202310494738 A CN 202310494738A CN 116385466 B CN116385466 B CN 116385466B
- Authority
- CN
- China
- Prior art keywords
- pseudo
- segmentation
- pixel
- labels
- level
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 28
- 230000011218 segmentation Effects 0.000 claims abstract description 109
- 238000002372 labelling Methods 0.000 claims abstract description 46
- 238000012549 training Methods 0.000 claims abstract description 27
- 238000002790 cross-validation Methods 0.000 claims abstract description 9
- 238000001514 detection method Methods 0.000 claims abstract description 7
- 238000012795 verification Methods 0.000 claims abstract description 7
- 238000003062 neural network model Methods 0.000 claims description 7
- 238000012545 processing Methods 0.000 claims description 7
- 230000006870 function Effects 0.000 claims description 6
- 238000013528 artificial neural network Methods 0.000 claims description 5
- 230000008878 coupling Effects 0.000 claims description 4
- 238000010168 coupling process Methods 0.000 claims description 4
- 238000005859 coupling reaction Methods 0.000 claims description 4
- 238000010276 construction Methods 0.000 claims description 3
- 238000005457 optimization Methods 0.000 claims 2
- 230000004927 fusion Effects 0.000 abstract 1
- 238000010586 diagram Methods 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Geometry (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a method and a system for dividing targets in images based on weak labels of boundary boxes, which relate to the technical field of image division and comprise the following steps: acquiring an image data set, and marking a boundary box of the image data set; generating pixel-level pseudo labels with confidence on the basis of the boundary box labels and training a target segmentation model according to the pixel-level pseudo labels; optimizing pixel-level pseudo labels based on cross verification iteration and training a target segmentation model based on the pixel-level pseudo labels; and outputting the optimal model. Compared with a pixel-level manual labeling method, the method does not need time-consuming and labor-consuming pixel-level labeling, only uses simple bounding box labeling, and saves manpower and material resources; the segmentation performance of the model is improved through simultaneous learning of two tasks of an output graph and a boundary box and fusion of two global and local scale graphs; by the cross-validation pseudo-annotation noise detection method, false annotations in the pseudo-annotations are detected, influence on model training is reduced, and target segmentation accuracy in images is improved.
Description
Technical Field
The invention relates to the technical field of image segmentation, in particular to a method and a system for segmenting targets in an image based on weak annotation of a boundary box.
Background
At present, the object segmentation in the image is widely applied to image content understanding or scene analysis tasks, and has wide application value. In recent years, the deep convolutional neural network greatly improves the performance of target segmentation, and the target segmentation method in the existing image has the advantages that the method with better performance is a full supervision method based on deep learning, but a large amount of manually marked pixel-level data is needed, so that the marking working cost is too high, and the limitation of other application scenes is difficult to generalize.
Therefore, the object segmentation method in the image based on the weak annotation of the bounding box is provided, time-consuming pixel-level annotation data is not needed, only the bounding boxes of all objects in the image are needed to be annotated, and finally a object segmentation model is learned based on the bounding boxes of the objects to carry out object segmentation, so that the problem to be solved by the person skilled in the art is needed to be solved.
Disclosure of Invention
In view of the above, the invention provides a method and a system for dividing targets in an image based on weak labeling of a bounding box, which do not need time-consuming pixel-level labeling data, only need to label bounding boxes of all targets in the image, then generate multi-level pseudo labels, optimize the pseudo labels through a neural network, and finally learn a target dividing model, and in order to achieve the above purpose, the invention adopts the following technical scheme:
A method for dividing targets in an image based on weak annotation of a bounding box comprises the following steps:
acquiring an image data set, and marking a boundary box of the image data set;
Generating a pixel-level pseudo label with confidence based on the boundary box label;
Training a target segmentation model according to the pixel-level pseudo-labels;
Optimizing pixel-level pseudo labels based on cross verification iteration, and training a target segmentation model through the optimized pixel-level pseudo labels to obtain an optimal model;
And inputting the image to be segmented into the optimal model for target segmentation to obtain a segmentation result.
Optionally, performing bounding box labeling on the image dataset includes: and marking the minimum circumscribed rectangular frame of each target in the image data set as a boundary frame, merging the boundary frames of a plurality of targets into one boundary frame when the boundary frames of the targets are overlapped, wherein the targets are positioned in the boundary frame, and the background area is arranged outside the boundary frame.
Optionally, processing the bounding box label to obtain a specific position of the object in the bounding box.
Optionally, the specific steps of processing the bounding box label are as follows:
marking the input image and the boundary box through a color space;
Dividing the color space by using a GrabCut algorithm, and generating a pseudo mark according to a division result;
Calculating a first segmentation confidence according to consistency of segmentation results of pixels in the pseudo labels:
calculating a second segmentation confidence according to the area ratio of the minimum circumscribed rectangle to the boundary box mark in the segmentation result;
and coupling the first segmentation confidence and the second segmentation confidence to obtain the confidence of the pseudo-label.
Optionally, the training the target segmentation model according to the pixel-level pseudo-labeling includes:
acquiring an image frame, marking a boundary frame, and marking a pixel level pseudo mark;
Constructing a deep neural network model;
Training a target segmentation depth neural network model under the supervision of pixel-level pseudo labeling;
And constructing a loss function through the loss between the boundary box label and the predicted boundary box, the loss between the global segmentation result and the pseudo label, the loss between the local segmentation result and the pseudo label and the confidence level of the pseudo label to guide the neural network to learn.
Optionally, the optimizing the pixel-level pseudo-labeling based on the cross-validation iteration includes:
randomly dividing the dataset T and the pseudo-annotation S into two parts T 1,S1 and T 2,S2; respectively learning two network models D (T 1,S1) and D (T 2,S2);
The result W 1 is obtained by deducing the learned network model D (T 2,S2) on the data set T 1, the degree of difference between the pseudo-annotation S 1 and the deduced result W 1 is calculated, namely the result is regarded as noise, the result is marked as noise in S 1, noise detection is carried out in the pseudo-annotation S 2, and the noise is marked in S 2;
updating the pseudo labels S 1 and S 2, combining the two parts of data, and repeating the steps to update iteratively until no noise is detected.
Optionally, the method comprises the following steps: and updating the pixel-level pseudo-labels by the optimized pixel-level pseudo-labels, training the target segmentation depth neural network model under the supervision of the optimized pixel-level pseudo-labels, iterating for N rounds until the model converges, and outputting an optimal model.
Optionally, a target segmentation system in an image based on a weak label of a bounding box includes:
the acquisition module is used for: for acquiring an image dataset;
and the marking module is used for: the method comprises the steps of performing boundary box labeling on the image data set, and generating pixel-level pseudo labels with confidence on the basis of the boundary box labeling;
The segmentation model construction module: training a target segmentation model according to the pixel-level pseudo-labels;
Training module: optimizing pixel-level pseudo labels based on cross verification iteration, and training a target segmentation model through the optimized pixel-level pseudo labels to obtain an optimal model;
And an output module: and inputting the image to be segmented into the optimal model for target segmentation to obtain a segmentation result.
Compared with the prior art, the method and the system for dividing the target in the image based on the weak annotation of the bounding box have the following beneficial effects:
compared with a pixel-level manual labeling method, the method for dividing the target in the image based on the boundary box weak labeling does not need time-consuming and labor-consuming pixel-level labeling, only uses simple boundary box labeling, is easier to popularize in any application scene, and saves a large amount of manpower and material resources; the pixel-level pseudo-labeling segmentation results and the result consistency thereof generated in different modes are used for obtaining pseudo-labels with confidence coefficient and guiding neural network learning; the global and local scale segmentation results are predicted by learning and predicting the simultaneous learning of the target segmentation task and the boundary frame task, so that the segmentation performance of the model is improved; by the cross-validation pseudo-annotation noise detection method, false annotations in the pseudo-annotations are detected, influence on model training is reduced, and target segmentation accuracy in images is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are required to be used in the embodiments or the description of the prior art will be briefly described below, and it is obvious that the drawings in the following description are only embodiments of the present invention, and that other drawings can be obtained according to the provided drawings without inventive effort for a person skilled in the art.
Fig. 1 is a flowchart of a method for dividing targets in an image based on weak annotation of a bounding box.
Fig. 2 is a schematic diagram of bounding box labeling provided in the present invention.
FIG. 3 is a schematic diagram of pseudo label generation with confidence provided by the present invention.
Fig. 4 is a flowchart of cross-validation-based noise detection provided by the present invention.
Fig. 5 is a schematic diagram of a target segmentation model structure provided by the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The embodiment of the invention discloses a target segmentation method in an image based on a boundary box weak annotation, which comprises the following steps:
acquiring an image data set, and marking a boundary box of the image data set;
Generating a pixel-level pseudo label with confidence based on the boundary box label;
Training a target segmentation model according to the pixel-level pseudo-labels;
Optimizing pixel-level pseudo labels based on cross verification iteration, and training a target segmentation model through the optimized pixel-level pseudo labels to obtain an optimal model;
And inputting the image to be segmented into the optimal model for target segmentation to obtain a segmentation result.
Further, as shown in fig. 2, the image dataset is subjected to bounding box labeling, and the pixel-level labeling has the limitation that the labeling work cost is too high and is difficult to generalize to other application scenes. And adopting simple bounding box labeling, namely using the smallest circumscribed rectangular box of each object in the picture as a bounding box labeling, and merging into one bounding box if the bounding boxes of a plurality of objects are overlapped. Wherein, the red rectangle box in fig. 2 shows the bounding box labeling example used by the method.
Further, the bounding box labels are processed to obtain specific positions of the objects in the bounding box.
Further, the specific steps of processing the bounding box label are as follows:
marking the input image and the boundary box through a color space;
Dividing the color space by using a GrabCut algorithm, and generating a pseudo mark according to a division result;
Calculating a first segmentation confidence according to consistency of segmentation results of pixels in the pseudo labels:
calculating a second segmentation confidence according to the area ratio of the minimum circumscribed rectangle to the boundary box mark in the segmentation result;
and coupling the first segmentation confidence and the second segmentation confidence to obtain the confidence of the pseudo-label.
The region outside the bounding box must be the background, according to the labeling definition of the bounding box, where the object is present, but no specific location at the pixel level of the object is given. The method is characterized in that firstly, based on the labeling of the boundary box, a pseudo labeling is generated by using a segmentation result obtained by using a GrabCut algorithm.
For the input image I and the boundary box label B, the input image is respectively represented by LAB and RGB color spaces, two segmentation results S LAB and S RGB are obtained by respectively applying Grabut algorithm to the two color spaces, and a pseudo label S is generated on the basis of the two segmentation results, as shown in a formula (1). For each pixel i, if both the segmentation results are 1, the false label is 1, if both the segmentation results are 0, the false label is 0, and otherwise, the false label is 0.5 when the segmentation results are inconsistent.
In the pseudo labeling result S, a pixel having a value of 1 and 0 indicates that the two division results are identical, and a pixel having a value of 0.5 indicates that the two division results are not identical. The proportion of 1 to 0 in the pseudo-label can be used for measuring the consistency of two segmentation results, and the segmentation results with higher consistency are more reliable. The duty cycle of 1 and 0 in the pseudo-label can thus be taken as a confidence in the segmentation result:
The confidence coefficient of each segmentation result can be calculated, two segmentation results S LAB and S RGB are obtained for the LAB and RGB color spaces, the minimum circumscribed rectangle and the area ratio of the minimum circumscribed rectangle to the boundary box mark B are calculated respectively, the ratios C LAB and C RGB are obtained, the ratio is closer to 1, the confidence coefficient is higher, and otherwise, the confidence coefficient is lower.
By combining the results, the confidence coefficient of the pseudo label S can be obtained:
C=CLAB-RGB×CLAB×CRGB (4)
fig. 3 is a schematic diagram of pseudo label generation with confidence, which sequentially includes an image and a bounding box, an LAB segmentation result, an RGB segmentation result and a pseudo label, wherein a red rectangular box is the bounding box label, the LAB segmentation result and the RGB segmentation result respectively represent the segmentation result of the GrabCut algorithm under the LAB and the RGB color spaces, a yellow box is the minimum circumscribed rectangle of the segmentation result, the pseudo label and the confidence level thereof are generated based on the two segmentation results, and white, gray and black in the pseudo label are respectively represented as 1, 0.5 and 0.
Further, the training the target segmentation model according to the pixel-level pseudo-labeling includes:
Under the supervision of pixel-level pseudo labeling, a deep neural network model of target segmentation is trained, and the network structure is shown in fig. 4. The input data includes an image I, a bounding box label B, and a pixel-level pseudo label S.
(1) The input image I firstly obtains global features through an encoder, performs bounding box regression prediction on the basis of the global features to obtain a bounding box B, and performs global segmentation to obtain a global segmentation result S G*.
(2) And obtaining a rough position of the target through the predicted boundary box B, extracting image features in the boundary box B to obtain local features, and dividing based on the local features to obtain a local division result S L*.
(3) The loss function of model training includes three parts: loss L (B, B) between the labeled and predicted bounding boxes B, loss L (S, S G*) between the global segmentation result S G* and the pseudo-label S, loss L (S, S L*) between the local segmentation result S L* and the pseudo-label S. Meanwhile, the false labeling confidence coefficient C is counted into the loss function, and the larger the confidence coefficient is, the larger the weight of the false labeling confidence coefficient C in the loss function is. The total loss function is:
Loss=(L(B,B*)+L(S,SG *)+L(S,SL *))×C (5)
the invention simultaneously generates the segmentation result and the prediction boundary frame, so that the model is a model for multi-task collaborative learning, and the two tasks are simultaneously learned, thereby further improving the characteristic learning capacity of the neural network. The segmentation results are generated based on the global features and the local features, respectively, and are therefore multi-scale segmentation models.
Further, the optimizing pixel-level pseudo-labeling based on cross-validation iteration includes:
in the pixel-level false labeling, although the confidence coefficient implies the probability that the false labeling is correct, the false labeling with low confidence coefficient is not reliable, but the false labeling with high confidence coefficient still has the condition of labeling errors, namely labeling noise. The invention further adopts a marking noise elimination mechanism based on cross validation, and the general flow is shown in figure 5. Firstly, randomly dividing a data set T and a pseudo-label S into two parts T 1,S1 and T 2,S2, respectively learning two network models D (T 1,S1) and D (T 2,S2), The result W 1 is then inferred on the data set T 1 using the learned network model D (T 2,S2). Since the model of inferred W 1 is trained on dataset T 2 and pseudo-annotation S 2, inferred result W 1 and pseudo-annotation S 1 are relatively independent. By calculating the degree of difference between the pseudo-labeling S 1 and the inferred result W 1, noise is considered to be greater than a certain threshold, and the result is labeled as noise (the confidence level is set to 0) in S 1. Similarly, noise in the pseudo tag S 2 can be detected and marked as noise (confidence level set to 0) in S 2. after updating the pseudo tags S 1 and S 2, combining the two parts of data, and repeating the steps to update iteratively until no pseudo tag noise is detected.
Further, the method comprises the step of continuously iteratively optimizing the pseudo label through cross-validation-based pseudo label noise detection on the basis of the initial pixel-level pseudo label with confidence. After the reliable pseudo labels are obtained, inputting all image data and the pseudo labels with confidence to a multi-task multi-scale target segmentation model, and retraining to obtain the target segmentation model which is the optimal model.
Further, a system for segmenting a target in an image based on weak annotation of a bounding box comprises:
the acquisition module is used for: for acquiring an image dataset;
and the marking module is used for: the method comprises the steps of performing boundary box labeling on the image data set, and generating pixel-level pseudo labels with confidence on the basis of the boundary box labeling;
The segmentation model construction module: training a target segmentation model according to the pixel-level pseudo-labels;
Training module: optimizing pixel-level pseudo labels based on cross verification iteration, and training a target segmentation model through the optimized pixel-level pseudo labels to obtain an optimal model;
And an output module: and inputting the image to be segmented into the optimal model for target segmentation to obtain a segmentation result.
In the present specification, each embodiment is described in a progressive manner, and each embodiment is mainly described in a different point from other embodiments, and identical and similar parts between the embodiments are all enough to refer to each other. For the device disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant points refer to the description of the method section.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Claims (5)
1. The method for dividing the target in the image based on the weak annotation of the bounding box is characterized by comprising the following steps:
acquiring an image data set, and marking a boundary box of the image data set;
Generating a pixel-level pseudo label with confidence based on the boundary box label;
processing the boundary box label to obtain a specific position of a target in the boundary box;
The specific steps of processing the boundary box label are as follows:
marking the input image and the boundary box through a color space;
Dividing the color space by using a GrabCut algorithm, and generating a pseudo mark according to a division result;
Calculating a first segmentation confidence according to consistency of segmentation results of pixels in the pseudo labels:
calculating a second segmentation confidence according to the area ratio of the minimum circumscribed rectangle to the boundary box mark in the segmentation result;
Coupling the first segmentation confidence coefficient and the second segmentation confidence coefficient to obtain the confidence coefficient of the pseudo label;
Training a target segmentation model according to the pixel-level pseudo-labels;
Optimizing pixel-level pseudo labels based on cross verification iteration, and training a target segmentation model through the optimized pixel-level pseudo labels to obtain an optimal model;
the iterative optimization pixel-level pseudo-labeling based on cross-validation comprises:
randomly dividing the dataset T and the pseudo-annotation S into two parts T 1,S1 and T 2,S2; respectively learning two network models D (T 1,S1) and D (T 2,S2);
The result W 1 is obtained by deducing the learned network model D (T 2,S2) on the data set T 1, the degree of difference between the pseudo-annotation S 1 and the deduced result W 1 is calculated, namely the result is regarded as noise, the result is marked as noise in S 1, noise detection is carried out in the pseudo-annotation S 2, and the noise is marked in S 2;
Updating the pseudo labels S 1 and S 2, combining the two parts of data, and repeating the steps for iterative updating until no noise is detected;
And inputting the image to be segmented into the optimal model for target segmentation to obtain a segmentation result.
2. The method for segmenting the target in the image based on the weak annotation of the boundary box according to claim 1, wherein the step of carrying out the boundary box annotation on the image data set comprises the following steps: and marking the minimum circumscribed rectangular frame of each target in the image data set as a boundary frame, merging the boundary frames of a plurality of targets into one boundary frame when the boundary frames of the targets are overlapped, wherein the targets are positioned in the boundary frame, and the background area is arranged outside the boundary frame.
3. The method for segmenting the target in the image based on the weak annotation of the bounding box according to claim 1, wherein the training the target segmentation model according to the pixel-level pseudo annotation comprises:
acquiring an image frame, marking a boundary frame, and marking a pixel level pseudo mark;
Constructing a deep neural network model;
Training a target segmentation depth neural network model under the supervision of pixel-level pseudo labeling;
And constructing a loss function through the loss between the boundary box label and the predicted boundary box, the loss between the global segmentation result and the pseudo label, the loss between the local segmentation result and the pseudo label and the confidence level of the pseudo label to guide the neural network to learn.
4. A method of object segmentation in an image based on bounding box weak annotation according to claim 3, comprising: and updating the pixel-level pseudo-labels by the optimized pixel-level pseudo-labels, training the target segmentation depth neural network model under the supervision of the optimized pixel-level pseudo-labels, iterating for N rounds until the model converges, and outputting an optimal model.
5. A system for segmenting a target in an image based on weak annotation of a bounding box, comprising:
the acquisition module is used for: for acquiring an image dataset;
and the marking module is used for: the method comprises the steps of performing boundary box labeling on the image data set, and generating pixel-level pseudo labels with confidence on the basis of the boundary box labeling;
processing the boundary box label to obtain a specific position of a target in the boundary box;
The specific steps of processing the boundary box label are as follows:
marking the input image and the boundary box through a color space;
Dividing the color space by using a GrabCut algorithm, and generating a pseudo mark according to a division result;
Calculating a first segmentation confidence according to consistency of segmentation results of pixels in the pseudo labels:
calculating a second segmentation confidence according to the area ratio of the minimum circumscribed rectangle to the boundary box mark in the segmentation result;
Coupling the first segmentation confidence coefficient and the second segmentation confidence coefficient to obtain the confidence coefficient of the pseudo label;
The segmentation model construction module: training a target segmentation model according to the pixel-level pseudo-labels;
Training module: optimizing pixel-level pseudo labels based on cross verification iteration, and training a target segmentation model through the optimized pixel-level pseudo labels to obtain an optimal model;
the iterative optimization pixel-level pseudo-labeling based on cross-validation comprises:
randomly dividing the dataset T and the pseudo-annotation S into two parts T 1,S1 and T 2,S2; respectively learning two network models D (T 1,S1) and D (T 2,S2);
The result W 1 is obtained by deducing the learned network model D (T 2,S2) on the data set T 1, the degree of difference between the pseudo-annotation S 1 and the deduced result W 1 is calculated, namely the result is regarded as noise, the result is marked as noise in S 1, noise detection is carried out in the pseudo-annotation S 2, and the noise is marked in S 2;
Updating the pseudo labels S 1 and S 2, combining the two parts of data, and repeating the steps for iterative updating until no noise is detected;
And an output module: and inputting the image to be segmented into the optimal model for target segmentation to obtain a segmentation result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310494738.7A CN116385466B (en) | 2023-05-05 | 2023-05-05 | Method and system for dividing targets in image based on boundary box weak annotation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310494738.7A CN116385466B (en) | 2023-05-05 | 2023-05-05 | Method and system for dividing targets in image based on boundary box weak annotation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116385466A CN116385466A (en) | 2023-07-04 |
CN116385466B true CN116385466B (en) | 2024-06-21 |
Family
ID=86971138
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310494738.7A Active CN116385466B (en) | 2023-05-05 | 2023-05-05 | Method and system for dividing targets in image based on boundary box weak annotation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116385466B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117078698B (en) * | 2023-08-22 | 2024-03-05 | 山东第一医科大学第二附属医院 | Peripheral blood vessel image auxiliary segmentation method and system based on deep learning |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114169389A (en) * | 2021-10-22 | 2022-03-11 | 福建亿榕信息技术有限公司 | Class-expanded target detection model training method and storage device |
CN115761574A (en) * | 2022-10-27 | 2023-03-07 | 之江实验室 | Weak surveillance video target segmentation method and device based on frame labeling |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110298387A (en) * | 2019-06-10 | 2019-10-01 | 天津大学 | Incorporate the deep neural network object detection method of Pixel-level attention mechanism |
CN110796046B (en) * | 2019-10-17 | 2023-10-10 | 武汉科技大学 | Intelligent steel slag detection method and system based on convolutional neural network |
KR102508067B1 (en) * | 2021-02-08 | 2023-03-08 | 연세대학교 산학협력단 | Apparatus and Method for Generating Learning Data for Semantic Image Segmentation Based On Weak Supervised Learning |
CN112926681B (en) * | 2021-03-29 | 2022-11-29 | 复旦大学 | Target detection method and device based on deep convolutional neural network |
CN113673338B (en) * | 2021-07-16 | 2023-09-26 | 华南理工大学 | Automatic labeling method, system and medium for weak supervision of natural scene text image character pixels |
-
2023
- 2023-05-05 CN CN202310494738.7A patent/CN116385466B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114169389A (en) * | 2021-10-22 | 2022-03-11 | 福建亿榕信息技术有限公司 | Class-expanded target detection model training method and storage device |
CN115761574A (en) * | 2022-10-27 | 2023-03-07 | 之江实验室 | Weak surveillance video target segmentation method and device based on frame labeling |
Also Published As
Publication number | Publication date |
---|---|
CN116385466A (en) | 2023-07-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11429818B2 (en) | Method, system and device for multi-label object detection based on an object detection network | |
TWI742382B (en) | Neural network system for vehicle parts recognition executed by computer, method for vehicle part recognition through neural network system, device and computing equipment for vehicle part recognition | |
CN109977918B (en) | Target detection positioning optimization method based on unsupervised domain adaptation | |
CN109241913B (en) | Ship detection method and system combining significance detection and deep learning | |
CN109886066B (en) | Rapid target detection method based on multi-scale and multi-layer feature fusion | |
WO2020000390A1 (en) | Systems and methods for depth estimation via affinity learned with convolutional spatial propagation networks | |
CN114332135B (en) | Semi-supervised medical image segmentation method and device based on dual-model interactive learning | |
US10262214B1 (en) | Learning method, learning device for detecting lane by using CNN and testing method, testing device using the same | |
CN108830171B (en) | Intelligent logistics warehouse guide line visual detection method based on deep learning | |
Li et al. | A robust instance segmentation framework for underground sewer defect detection | |
CN111611643A (en) | Family type vectorization data obtaining method and device, electronic equipment and storage medium | |
CN113673338B (en) | Automatic labeling method, system and medium for weak supervision of natural scene text image character pixels | |
US20220180624A1 (en) | Method and device for automatic identification of labels of an image | |
CN113609896A (en) | Object-level remote sensing change detection method and system based on dual-correlation attention | |
CN113313166B (en) | Ship target automatic labeling method based on feature consistency learning | |
CN111738295B (en) | Image segmentation method and storage medium | |
CN116385466B (en) | Method and system for dividing targets in image based on boundary box weak annotation | |
CN115410059B (en) | Remote sensing image part supervision change detection method and device based on contrast loss | |
CN116740758A (en) | Bird image recognition method and system for preventing misjudgment | |
CN116071389A (en) | Front background matching-based boundary frame weak supervision image segmentation method | |
CN113657225B (en) | Target detection method | |
CN116957051A (en) | Remote sensing image weak supervision target detection method for optimizing feature extraction | |
CN116758421A (en) | Remote sensing image directed target detection method based on weak supervised learning | |
CN116258877A (en) | Land utilization scene similarity change detection method, device, medium and equipment | |
CN116681961A (en) | Weak supervision target detection method based on semi-supervision method and noise processing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant |