CN117437465A - Improved soft-NMS target detection method based on unbalanced data - Google Patents
Improved soft-NMS target detection method based on unbalanced data Download PDFInfo
- Publication number
- CN117437465A CN117437465A CN202311375631.7A CN202311375631A CN117437465A CN 117437465 A CN117437465 A CN 117437465A CN 202311375631 A CN202311375631 A CN 202311375631A CN 117437465 A CN117437465 A CN 117437465A
- Authority
- CN
- China
- Prior art keywords
- target
- target detection
- detection
- frames
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 125
- 238000012549 training Methods 0.000 claims abstract description 24
- 238000000034 method Methods 0.000 claims abstract description 16
- 238000012216 screening Methods 0.000 claims abstract description 7
- 230000015572 biosynthetic process Effects 0.000 claims description 4
- 238000011176 pooling Methods 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 4
- 238000003786 synthesis reaction Methods 0.000 claims description 4
- 230000004927 fusion Effects 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 10
- 238000005070 sampling Methods 0.000 description 6
- 230000001629 suppression Effects 0.000 description 5
- 230000004913 activation Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000009826 distribution Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000009776 industrial production Methods 0.000 description 1
- 230000005764 inhibitory process Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an improved soft-NMS target detection method based on unbalanced data, which comprises the following steps: step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target; training a target detection model based on a Faster RCNN; step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames; step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence; step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame; and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results. The invention redefines the updated formula of the confidence coefficient of the target detection frame, and effectively reduces the omission ratio of the tail class targets.
Description
Technical Field
The invention relates to the technical field of target detection of image processing, in particular to an improved soft-NMS target detection method based on unbalanced data.
Background
Object detection is an important research direction in the field of computer vision, and the object is to determine whether a specific object exists in an image or video and determine its category and position. In practical systems, unbalanced distribution data is widely used in various application fields of target detection, such as security monitoring, automatic driving, remote sensing, image segmentation, target tracking, and the like. An unbalanced distribution is a particular problem of unbalance in a large data background, meaning that most classes in a data set occupy a large number of samples, while few classes have only a small number of samples.
The common target detection model is based on balanced sample types in the training data set, so that most samples are trained excessively in the training process, and few samples are not trained sufficiently, so that the target detection model trained by unbalanced data is often biased to most samples with a large number of training samples, and the target detection model has poor performance on few samples with limited sample size, thereby reducing the overall performance of the target detection model, and particularly has poorer detection precision on small-size targets of few types. In practical application, even if a higher detection rate can be obtained on the whole, false detection of a few types of examples may cause serious consequences or high cost, such as false industrial process fault diagnosis, no hacker intrusion detection, remote sensing detection omission detection or false judgment of obstacles in automatic driving, etc., all belong to the false detection of few types, and great harm is caused to industrial production and network security.
The non-maximum suppression NMS algorithm and soft-NMS algorithm are technical methods for selecting detection frames frequently used in target detection algorithms, and when unbalanced data characteristics are not considered, the technical methods can cause missed detection and false detection of target detection due to excessive suppression of few target detection frames, so that the detection accuracy is reduced. Therefore, it is important to study non-maximum suppression NMS-related algorithms under unbalanced data.
Disclosure of Invention
(one) solving the technical problems
Aiming at the defects of the prior art, the invention provides an improved soft-NMS target detection method based on unbalanced data, which considers the long-tail data phenomenon widely existing in the actual application of target detection and redefines an updating formula of the confidence coefficient of a target detection frame, thereby protecting the tail type target detection frame from being excessively restrained and effectively reducing the omission ratio of the tail type target.
(II) technical scheme
In order to achieve the above purpose, the present invention provides the following technical solutions: an improved soft-NMS target detection method based on imbalance data, comprising the steps of:
step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target;
training a target detection model based on a Faster RCNN;
step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames;
step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence;
step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame;
and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results.
Preferably, in the first step, calculating the balance coefficient of each type of object includes the following operations:
according to the characteristics of unbalanced data, calculating a balance coefficient k (i) of each type of target, wherein the formula is as follows;
where i=1, 2, …, C is the number of target categories in the imbalance data; n (i) is the total number of class i targets.
Preferably, the training of the target detection model based on the fast RCNN in the second step includes the following operations:
s1, inputting an unbalanced data training set picture into a Regnet backbone network, and extracting feature graphs C2, C3, C4 and C5 of 4 different stages;
s2, inputting the characteristic graphs C2, C3, C4 and C5 of the 4 different stages into an FPN characteristic pyramid for fusion to obtain new characteristic graphs P2, P3, P4 and P5 with 4 different resolutions;
s3, extracting suggestion frames of the feature graphs P2, P3, P4 and P5 through an RPN (remote procedure network) area suggestion network;
s4, inputting the feature graphs P2, P3, P4 and P5 and the suggestion boxes into the RoI align for pooling synthesis, and then classifying and regressing through a fully connected network to obtain a detection result and corresponding loss;
and S5, repeating the steps from S1 to S4 after the training of one round is completed until the set iteration round is reached, and outputting and storing parameters in a network to obtain a target detection model.
Preferably, in the fifth step, the improved soft-NMS method based on unbalance data comprises the following operations:
(a) Selecting a target detection frame with highest confidence in each type of targets, marking as M, and adding the M into the set D;
(b) Calculating the cross ratio between the rest target detection frames and M, and processing the confidence s of the target detection frames according to the following formula j The threshold is set in a segmented threshold mode, and the formula is as follows:
where j=1, 2, …, B (i) -1, represents the j-th detection frame of the i-th detection target, and B (i) is the number of detection frames of the i-th detection target; u (U) Iou (M,b j ) Is the intersection ratio of the detection frame M and the jth detection frame; k (i) is the balance coefficient of the unbalanced data;and alpha < beta;
(c) And (c) if all the types of target detection frames obtained in the step (b) are empty, finishing the screening of the detection frames.
(III) beneficial effects
Compared with the prior art, the invention provides an improved soft-NMS target detection method based on unbalanced data, which has the following beneficial effects:
1. the invention provides an improved method for target detection aiming at unbalanced data characteristics. Compared with the prior art, the invention considers the phenomenon of unbalanced data widely existing in the practical application of target detection, the unbalanced data of most types of targets and few types of targets causes the reduction of the accuracy of the class identification of the target detection, and the omission ratio of the tail class target frame in the non-maximum value inhibition algorithm is high. The improved method provided by the invention can effectively reduce the omission ratio of tail class targets;
2. the invention provides an improved non-maximum suppression algorithm, wherein the algorithm introduces the balance coefficient of unbalanced data, redefines an updating formula of the confidence coefficient of a target detection frame, sets a threshold value in a sectional threshold value mode, reduces the influence of the unbalance of the data on the performance of the algorithm, and improves the application flexibility and the performance of the non-maximum suppression algorithm;
3. the improved soft-NMS target detection method based on unbalanced data has universality, can be popularized to the target detection field of other computer vision, and has better popularization.
Drawings
Fig. 1 is a training flow chart of the method of the present invention.
Detailed Description
For a better understanding of the objects, structures and functions of the present invention, the improved soft-NMS object detection method based on unbalance data of the present invention will be described in further detail with reference to the specific embodiments and the accompanying drawings.
Experiments the balanced CIFAR10, CIFAR100 and ImageNet2012 datasets may be downsampled using exponential decay to generate an imbalance dataset for the experiment, referred to as CIFAR10-LT, CIFAR100-LT and ImageNet-LT. CIFAR10-LT and CIFAR100-LT generate three different types of training sets according to imbalance rates {10,50,100 }. The imbalance ratio of ImageNet-LT was 256. The category of the maximum sample includes 1280 pictures, and only contains 5 pictures at least. The validation sets for all data sets are balanced.
As shown in fig. 1, the improved soft-NMS target detection method based on unbalanced data provided by the present invention comprises the following steps:
step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target;
training a target detection model based on a Faster RCNN;
step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames;
step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence;
step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame;
and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results.
Further, in step one, calculating the balance coefficient of each type of object includes the following operations:
according to the characteristics of unbalanced data, calculating a balance coefficient k (i) of each type of target, wherein the formula is as follows;
where i=1, 2, …, C is the number of target categories in the imbalance data; n (i) is the total number of class i targets.
Further, training the target detection model based on the Faster RCNN in the second step comprises the following operations:
s1, inputting an unbalanced data training set picture into a Regnet backbone network, and extracting feature graphs C2, C3, C4 and C5 of 4 different stages;
s2, inputting the characteristic graphs C2, C3, C4 and C5 of the 4 different stages into an FPN characteristic pyramid for fusion to obtain new characteristic graphs P2, P3, P4 and P5 with 4 different resolutions;
s3, extracting suggestion frames of the feature graphs P2, P3, P4 and P5 through an RPN (remote procedure network) area suggestion network;
s4, inputting the feature graphs P2, P3, P4 and P5 and the suggestion boxes into the RoI align for pooling synthesis, and then classifying and regressing through a fully connected network to obtain a detection result and corresponding loss;
and S5, repeating the steps from S1 to S4 after the training of one round is completed until the set iteration round is reached, and outputting and storing parameters in a network to obtain a target detection model.
Specifically, in the second step, a fast RCNN model is used to train a target detection model, which includes the following operations:
defining a bottleneck layer, wherein one branch sequentially passes through a group convolution layer with the convolution kernel size of 1x1, a regularized BN layer, a group convolution layer with the convolution kernel size of 3x3, a regularized BN layer, a convolution layer with the convolution kernel size of 1x1, a regularized BN layer and a relu activation function, the other branch is an identity mapping, and the outputs of the two branches are added to obtain the output of the bottleneck; the pictures in the training set are normalized and input into a regnet backbone network, the pictures are sequentially subjected to convolution layers with the convolution kernel size of 3x3, the step length of 2, the filling of 1 and the output channel of 48, a regularized BN layer, an activation function relu, a characteristic diagram C2 is obtained through 2 bottleck with the output channel of 96, a characteristic diagram C3 is obtained through 6 bottleck with the output channel of 192, a characteristic diagram C4 is obtained through 15 bottleck with the output channel of 432, and a characteristic diagram C5 is obtained through 2 bottleck with the output channel of 1008.
C5 is subjected to a convolution layer with a convolution kernel size of 1x1 and an output channel of 256 to obtain a feature map M5, and then is subjected to a convolution layer with a convolution kernel size of 3x3 to output as a feature map P5;
c4 is added with a characteristic diagram obtained by up-sampling of M5 with a sampling multiplying power of 2 through a convolution layer with a convolution kernel size of 1x1 and an output channel of 256 to obtain a characteristic diagram M4, and then the characteristic diagram P4 is output through a convolution layer with a convolution kernel size of 3x 3;
c3 is added with a characteristic diagram obtained by up-sampling M4 with a sampling multiplying power of 2 through a convolution layer with a convolution kernel size of 1x1 and an output channel of 256 to obtain a characteristic diagram M3, and then the characteristic diagram P3 is output through the convolution layer with a convolution kernel size of 3x 3;
c2 is added with a characteristic map obtained by up-sampling M3 with a sampling multiplying power of 2 through a convolution layer output channel with a convolution kernel size of 1x1 and an output channel of 256 to obtain a characteristic map M2, and then the characteristic map P2 is output through a convolution layer with a convolution kernel size of 3x 3.
Determining the proportion and the size of the anchors, and extracting suggestion boxes from the feature maps P2, P3, P4 and P5.
The RPN firstly carries out 3X3 convolution on the feature map obtained by the feature extraction network to fuse the feature map information; then, according to anchors; setting two groups of parallel 1X1 convolutions, and respectively classifying and regressing the feature images; the classification layer carries out two classifications on each anchor box, judges whether the anchor box belongs to the foreground or the background, and outputs a result for each anchor on the feature map; the regression layer predicts the offset between each anchor box and the real labeling frame, and comprises the offsets deltax and deltay of the center coordinates and the offsets deltaw and deltah of the width and the height, so that the regression layer outputs a result for each anchor on the feature map; and finally, integrating the results of the two branches, selecting an anchor box with a classification layer prediction result as a foreground, and adjusting the central point and the length and the width of the anchor box by using the offset obtained by the regression layer calculation.
Inputting the suggestion boxes output by each feature map and the RPN area suggestion network into the RoI align for pooling synthesis to obtain suggestion boxes of each feature map with the size of 7x7, merging and inputting the suggestion boxes into the full-connection network to obtain 6-channel and 24-channel outputs, representing classification and regression results, and calculating target detection loss.
After the training of one round is completed, repeating the steps until the set iteration round is reached, outputting and storing parameters in the network, and obtaining the FaterRCNN target detection model.
In step five, an improved soft-NMS method based on imbalance data, comprising the operations of:
(a) Selecting a target detection frame with highest confidence in each type of targets, marking as M, and adding the M into the set D;
(b) Calculating the cross ratio between the rest target detection frames and M, and processing the confidence s of the target detection frames according to the following formula j The threshold is set in a segmented threshold mode, and the formula is as follows:
where j=1, 2, …, B (i) -1, represents the j-th detection frame of the i-th detection target, and B (i) is the number of detection frames of the i-th detection target; u (U) Iou (M,b j ) Is the intersection ratio of the detection frame M and the jth detection frame; k (i) is the balance coefficient of the unbalanced data;and alpha < beta;
(c) And (c) if all the types of target detection frames obtained in the step (b) are empty, finishing the screening of the detection frames.
According to the improved soft-NMS target detection method based on unbalanced data, which is provided by the invention, the phenomenon of long-tail data widely existing in the actual application of target detection is considered, the balance coefficient of the unbalanced data is introduced, the updating formula of the confidence coefficient of the target detection frame is redefined, the threshold is set in a sectional threshold mode, the influence of the unbalance of the data on the algorithm performance is reduced, the target detection frame of the tail type is protected from being excessively restrained, and the problem of target omission of the tail type is reduced.
It will be understood that the invention has been described in terms of several embodiments, and that various changes and equivalents may be made to these features and embodiments by those skilled in the art without departing from the spirit and scope of the invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to the particular embodiment disclosed, but that the invention will include all embodiments falling within the scope of the appended claims.
Claims (4)
1. An improved soft-NMS target detection method based on imbalance data, comprising the steps of:
step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target;
training a target detection model based on a Faster RCNN;
step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames;
step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence;
step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame;
and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results.
2. The improved soft-NMS object detection method based on unbalanced data of claim 1, wherein in the first step, the balance coefficient of each class of objects is calculated, comprising the following operations:
according to the characteristics of unbalanced data, calculating a balance coefficient k (i) of each type of target, wherein the formula is as follows;
where i=1, 2, …, C is the number of target categories in the imbalance data; n (i) is the total number of class i targets.
3. The improved soft-NMS object detection method based on unbalanced data of claim 1, wherein training the object detection model based on Faster RCNN in the second step comprises the following operations:
s1, inputting an unbalanced data training set picture into a Regnet backbone network, and extracting feature graphs C2, C3, C4 and C5 of 4 different stages;
s2, inputting the characteristic graphs C2, C3, C4 and C5 of the 4 different stages into an FPN characteristic pyramid for fusion to obtain new characteristic graphs P2, P3, P4 and P5 with 4 different resolutions;
s3, extracting suggestion frames of the feature graphs P2, P3, P4 and P5 through an RPN (remote procedure network) area suggestion network;
s4, inputting the feature graphs P2, P3, P4 and P5 and the suggestion boxes into the RoI align for pooling synthesis, and then classifying and regressing through a fully connected network to obtain a detection result and corresponding loss;
and S5, repeating the steps from S1 to S4 after the training of one round is completed until the set iteration round is reached, and outputting and storing parameters in a network to obtain a target detection model.
4. The improved soft-NMS object detection method based on unbalanced data according to claim 1, wherein in the fifth step, the improved soft-NMS method based on unbalanced data comprises the following operations:
(a) Selecting a target detection frame with highest confidence in each type of targets, marking as M, and adding the M into the set D;
(b) Calculating the cross ratio between the rest target detection frames and M, and processing the confidence s of the target detection frames according to the following formula j The threshold is set in a segmented threshold mode, and the formula is as follows:
where j=1, 2, …, B (i) -1, represents the j-th detection frame of the i-th detection target, and B (i) is the number of detection frames of the i-th detection target; u (U) Iou (M,b j ) Is the intersection ratio of the detection frame M and the jth detection frame; k (i) is the balance coefficient of the unbalanced data;and alpha < beta;
(c) And (c) if all the types of target detection frames obtained in the step (b) are empty, finishing the screening of the detection frames.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311375631.7A CN117437465B (en) | 2023-10-23 | 2023-10-23 | Improved soft-NMS target detection method based on unbalanced data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311375631.7A CN117437465B (en) | 2023-10-23 | 2023-10-23 | Improved soft-NMS target detection method based on unbalanced data |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117437465A true CN117437465A (en) | 2024-01-23 |
CN117437465B CN117437465B (en) | 2024-06-07 |
Family
ID=89556256
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311375631.7A Active CN117437465B (en) | 2023-10-23 | 2023-10-23 | Improved soft-NMS target detection method based on unbalanced data |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117437465B (en) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104134080A (en) * | 2014-08-01 | 2014-11-05 | 重庆大学 | Method and system for automatically detecting roadbed collapse and side slope collapse of road |
CN110348329A (en) * | 2019-06-24 | 2019-10-18 | 电子科技大学 | Pedestrian detection method based on video sequence interframe information |
CN111882546A (en) * | 2020-07-30 | 2020-11-03 | 中原工学院 | Weak supervised learning-based three-branch convolutional network fabric defect detection method |
CN114078106A (en) * | 2020-08-06 | 2022-02-22 | 沈阳中科数控技术股份有限公司 | Defect detection method based on improved Faster R-CNN |
CN114359199A (en) * | 2021-12-28 | 2022-04-15 | 集美大学 | Fish counting method, device, equipment and medium based on deep learning |
CN114821271A (en) * | 2022-05-19 | 2022-07-29 | 平安科技(深圳)有限公司 | Model training method, image description generation device and storage medium |
CN115346135A (en) * | 2022-08-19 | 2022-11-15 | 陕西航天技术应用研究院有限公司 | Optical remote sensing image ship target identification method based on convolutional neural network |
CN116561622A (en) * | 2023-04-25 | 2023-08-08 | 广西师范大学 | Federal learning method for class unbalanced data distribution |
-
2023
- 2023-10-23 CN CN202311375631.7A patent/CN117437465B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104134080A (en) * | 2014-08-01 | 2014-11-05 | 重庆大学 | Method and system for automatically detecting roadbed collapse and side slope collapse of road |
CN110348329A (en) * | 2019-06-24 | 2019-10-18 | 电子科技大学 | Pedestrian detection method based on video sequence interframe information |
CN111882546A (en) * | 2020-07-30 | 2020-11-03 | 中原工学院 | Weak supervised learning-based three-branch convolutional network fabric defect detection method |
CN114078106A (en) * | 2020-08-06 | 2022-02-22 | 沈阳中科数控技术股份有限公司 | Defect detection method based on improved Faster R-CNN |
CN114359199A (en) * | 2021-12-28 | 2022-04-15 | 集美大学 | Fish counting method, device, equipment and medium based on deep learning |
CN114821271A (en) * | 2022-05-19 | 2022-07-29 | 平安科技(深圳)有限公司 | Model training method, image description generation device and storage medium |
CN115346135A (en) * | 2022-08-19 | 2022-11-15 | 陕西航天技术应用研究院有限公司 | Optical remote sensing image ship target identification method based on convolutional neural network |
CN116561622A (en) * | 2023-04-25 | 2023-08-08 | 广西师范大学 | Federal learning method for class unbalanced data distribution |
Also Published As
Publication number | Publication date |
---|---|
CN117437465B (en) | 2024-06-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110084292B (en) | Target detection method based on DenseNet and multi-scale feature fusion | |
CN108776772B (en) | Cross-time building change detection modeling method, detection device, method and storage medium | |
CN110991311B (en) | Target detection method based on dense connection deep network | |
CN110033481B (en) | Method and apparatus for image processing | |
CN110826379B (en) | Target detection method based on feature multiplexing and YOLOv3 | |
KR102328734B1 (en) | Method for automatically evaluating labeling reliability of training images for use in deep learning network to analyze images, and reliability-evaluating device using the same | |
CN113269073A (en) | Ship multi-target tracking method based on YOLO V5 algorithm | |
US20120134586A1 (en) | Device with datastream pipeline architecture for recognizing and locating objects in an image by detection window scanning | |
CN110147707B (en) | High-precision vehicle identification method and system | |
CN109902576B (en) | Training method and application of head and shoulder image classifier | |
CN115410100A (en) | Small target detection method and system based on unmanned aerial vehicle image | |
CN114078197A (en) | Small sample target detection method and device based on support sample characteristic enhancement | |
CN111414882B (en) | Infrared target detection method based on multi-level resolution balance network | |
CN116363532A (en) | Unmanned aerial vehicle image traffic target detection method based on attention mechanism and re-parameterization | |
CN114120138A (en) | Method, device, equipment and medium for detecting and identifying remote sensing image target | |
CN113343989A (en) | Target detection method and system based on self-adaption of foreground selection domain | |
CN113887455B (en) | Face mask detection system and method based on improved FCOS | |
CN111553184A (en) | Small target detection method and device based on electronic purse net and electronic equipment | |
CN113454649B (en) | Target detection method, apparatus, electronic device, and computer-readable storage medium | |
CN111222575A (en) | KLXS multi-model fusion method and system based on HRRP target recognition | |
CN110399868B (en) | Coastal wetland bird detection method | |
CN117437465B (en) | Improved soft-NMS target detection method based on unbalanced data | |
CN116977859A (en) | Weak supervision target detection method based on multi-scale image cutting and instance difficulty | |
CN116206146A (en) | Real-time detection method for lawn object, electronic device and storage medium | |
CN115272741A (en) | Detection method of slender flexible object, terminal equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant |