CN117437465A - Improved soft-NMS target detection method based on unbalanced data - Google Patents

Improved soft-NMS target detection method based on unbalanced data Download PDF

Info

Publication number
CN117437465A
CN117437465A CN202311375631.7A CN202311375631A CN117437465A CN 117437465 A CN117437465 A CN 117437465A CN 202311375631 A CN202311375631 A CN 202311375631A CN 117437465 A CN117437465 A CN 117437465A
Authority
CN
China
Prior art keywords
target
target detection
detection
frames
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311375631.7A
Other languages
Chinese (zh)
Other versions
CN117437465B (en
Inventor
林小泸
陈建明
杨建国
张小东
彭金龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CHANGXUN COMMUNICATION SERVICE CO LTD
Original Assignee
CHANGXUN COMMUNICATION SERVICE CO LTD
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CHANGXUN COMMUNICATION SERVICE CO LTD filed Critical CHANGXUN COMMUNICATION SERVICE CO LTD
Priority to CN202311375631.7A priority Critical patent/CN117437465B/en
Publication of CN117437465A publication Critical patent/CN117437465A/en
Application granted granted Critical
Publication of CN117437465B publication Critical patent/CN117437465B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an improved soft-NMS target detection method based on unbalanced data, which comprises the following steps: step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target; training a target detection model based on a Faster RCNN; step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames; step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence; step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame; and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results. The invention redefines the updated formula of the confidence coefficient of the target detection frame, and effectively reduces the omission ratio of the tail class targets.

Description

Improved soft-NMS target detection method based on unbalanced data
Technical Field
The invention relates to the technical field of target detection of image processing, in particular to an improved soft-NMS target detection method based on unbalanced data.
Background
Object detection is an important research direction in the field of computer vision, and the object is to determine whether a specific object exists in an image or video and determine its category and position. In practical systems, unbalanced distribution data is widely used in various application fields of target detection, such as security monitoring, automatic driving, remote sensing, image segmentation, target tracking, and the like. An unbalanced distribution is a particular problem of unbalance in a large data background, meaning that most classes in a data set occupy a large number of samples, while few classes have only a small number of samples.
The common target detection model is based on balanced sample types in the training data set, so that most samples are trained excessively in the training process, and few samples are not trained sufficiently, so that the target detection model trained by unbalanced data is often biased to most samples with a large number of training samples, and the target detection model has poor performance on few samples with limited sample size, thereby reducing the overall performance of the target detection model, and particularly has poorer detection precision on small-size targets of few types. In practical application, even if a higher detection rate can be obtained on the whole, false detection of a few types of examples may cause serious consequences or high cost, such as false industrial process fault diagnosis, no hacker intrusion detection, remote sensing detection omission detection or false judgment of obstacles in automatic driving, etc., all belong to the false detection of few types, and great harm is caused to industrial production and network security.
The non-maximum suppression NMS algorithm and soft-NMS algorithm are technical methods for selecting detection frames frequently used in target detection algorithms, and when unbalanced data characteristics are not considered, the technical methods can cause missed detection and false detection of target detection due to excessive suppression of few target detection frames, so that the detection accuracy is reduced. Therefore, it is important to study non-maximum suppression NMS-related algorithms under unbalanced data.
Disclosure of Invention
(one) solving the technical problems
Aiming at the defects of the prior art, the invention provides an improved soft-NMS target detection method based on unbalanced data, which considers the long-tail data phenomenon widely existing in the actual application of target detection and redefines an updating formula of the confidence coefficient of a target detection frame, thereby protecting the tail type target detection frame from being excessively restrained and effectively reducing the omission ratio of the tail type target.
(II) technical scheme
In order to achieve the above purpose, the present invention provides the following technical solutions: an improved soft-NMS target detection method based on imbalance data, comprising the steps of:
step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target;
training a target detection model based on a Faster RCNN;
step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames;
step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence;
step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame;
and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results.
Preferably, in the first step, calculating the balance coefficient of each type of object includes the following operations:
according to the characteristics of unbalanced data, calculating a balance coefficient k (i) of each type of target, wherein the formula is as follows;
where i=1, 2, …, C is the number of target categories in the imbalance data; n (i) is the total number of class i targets.
Preferably, the training of the target detection model based on the fast RCNN in the second step includes the following operations:
s1, inputting an unbalanced data training set picture into a Regnet backbone network, and extracting feature graphs C2, C3, C4 and C5 of 4 different stages;
s2, inputting the characteristic graphs C2, C3, C4 and C5 of the 4 different stages into an FPN characteristic pyramid for fusion to obtain new characteristic graphs P2, P3, P4 and P5 with 4 different resolutions;
s3, extracting suggestion frames of the feature graphs P2, P3, P4 and P5 through an RPN (remote procedure network) area suggestion network;
s4, inputting the feature graphs P2, P3, P4 and P5 and the suggestion boxes into the RoI align for pooling synthesis, and then classifying and regressing through a fully connected network to obtain a detection result and corresponding loss;
and S5, repeating the steps from S1 to S4 after the training of one round is completed until the set iteration round is reached, and outputting and storing parameters in a network to obtain a target detection model.
Preferably, in the fifth step, the improved soft-NMS method based on unbalance data comprises the following operations:
(a) Selecting a target detection frame with highest confidence in each type of targets, marking as M, and adding the M into the set D;
(b) Calculating the cross ratio between the rest target detection frames and M, and processing the confidence s of the target detection frames according to the following formula j The threshold is set in a segmented threshold mode, and the formula is as follows:
where j=1, 2, …, B (i) -1, represents the j-th detection frame of the i-th detection target, and B (i) is the number of detection frames of the i-th detection target; u (U) Iou (M,b j ) Is the intersection ratio of the detection frame M and the jth detection frame; k (i) is the balance coefficient of the unbalanced data;and alpha < beta;
(c) And (c) if all the types of target detection frames obtained in the step (b) are empty, finishing the screening of the detection frames.
(III) beneficial effects
Compared with the prior art, the invention provides an improved soft-NMS target detection method based on unbalanced data, which has the following beneficial effects:
1. the invention provides an improved method for target detection aiming at unbalanced data characteristics. Compared with the prior art, the invention considers the phenomenon of unbalanced data widely existing in the practical application of target detection, the unbalanced data of most types of targets and few types of targets causes the reduction of the accuracy of the class identification of the target detection, and the omission ratio of the tail class target frame in the non-maximum value inhibition algorithm is high. The improved method provided by the invention can effectively reduce the omission ratio of tail class targets;
2. the invention provides an improved non-maximum suppression algorithm, wherein the algorithm introduces the balance coefficient of unbalanced data, redefines an updating formula of the confidence coefficient of a target detection frame, sets a threshold value in a sectional threshold value mode, reduces the influence of the unbalance of the data on the performance of the algorithm, and improves the application flexibility and the performance of the non-maximum suppression algorithm;
3. the improved soft-NMS target detection method based on unbalanced data has universality, can be popularized to the target detection field of other computer vision, and has better popularization.
Drawings
Fig. 1 is a training flow chart of the method of the present invention.
Detailed Description
For a better understanding of the objects, structures and functions of the present invention, the improved soft-NMS object detection method based on unbalance data of the present invention will be described in further detail with reference to the specific embodiments and the accompanying drawings.
Experiments the balanced CIFAR10, CIFAR100 and ImageNet2012 datasets may be downsampled using exponential decay to generate an imbalance dataset for the experiment, referred to as CIFAR10-LT, CIFAR100-LT and ImageNet-LT. CIFAR10-LT and CIFAR100-LT generate three different types of training sets according to imbalance rates {10,50,100 }. The imbalance ratio of ImageNet-LT was 256. The category of the maximum sample includes 1280 pictures, and only contains 5 pictures at least. The validation sets for all data sets are balanced.
As shown in fig. 1, the improved soft-NMS target detection method based on unbalanced data provided by the present invention comprises the following steps:
step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target;
training a target detection model based on a Faster RCNN;
step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames;
step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence;
step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame;
and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results.
Further, in step one, calculating the balance coefficient of each type of object includes the following operations:
according to the characteristics of unbalanced data, calculating a balance coefficient k (i) of each type of target, wherein the formula is as follows;
where i=1, 2, …, C is the number of target categories in the imbalance data; n (i) is the total number of class i targets.
Further, training the target detection model based on the Faster RCNN in the second step comprises the following operations:
s1, inputting an unbalanced data training set picture into a Regnet backbone network, and extracting feature graphs C2, C3, C4 and C5 of 4 different stages;
s2, inputting the characteristic graphs C2, C3, C4 and C5 of the 4 different stages into an FPN characteristic pyramid for fusion to obtain new characteristic graphs P2, P3, P4 and P5 with 4 different resolutions;
s3, extracting suggestion frames of the feature graphs P2, P3, P4 and P5 through an RPN (remote procedure network) area suggestion network;
s4, inputting the feature graphs P2, P3, P4 and P5 and the suggestion boxes into the RoI align for pooling synthesis, and then classifying and regressing through a fully connected network to obtain a detection result and corresponding loss;
and S5, repeating the steps from S1 to S4 after the training of one round is completed until the set iteration round is reached, and outputting and storing parameters in a network to obtain a target detection model.
Specifically, in the second step, a fast RCNN model is used to train a target detection model, which includes the following operations:
defining a bottleneck layer, wherein one branch sequentially passes through a group convolution layer with the convolution kernel size of 1x1, a regularized BN layer, a group convolution layer with the convolution kernel size of 3x3, a regularized BN layer, a convolution layer with the convolution kernel size of 1x1, a regularized BN layer and a relu activation function, the other branch is an identity mapping, and the outputs of the two branches are added to obtain the output of the bottleneck; the pictures in the training set are normalized and input into a regnet backbone network, the pictures are sequentially subjected to convolution layers with the convolution kernel size of 3x3, the step length of 2, the filling of 1 and the output channel of 48, a regularized BN layer, an activation function relu, a characteristic diagram C2 is obtained through 2 bottleck with the output channel of 96, a characteristic diagram C3 is obtained through 6 bottleck with the output channel of 192, a characteristic diagram C4 is obtained through 15 bottleck with the output channel of 432, and a characteristic diagram C5 is obtained through 2 bottleck with the output channel of 1008.
C5 is subjected to a convolution layer with a convolution kernel size of 1x1 and an output channel of 256 to obtain a feature map M5, and then is subjected to a convolution layer with a convolution kernel size of 3x3 to output as a feature map P5;
c4 is added with a characteristic diagram obtained by up-sampling of M5 with a sampling multiplying power of 2 through a convolution layer with a convolution kernel size of 1x1 and an output channel of 256 to obtain a characteristic diagram M4, and then the characteristic diagram P4 is output through a convolution layer with a convolution kernel size of 3x 3;
c3 is added with a characteristic diagram obtained by up-sampling M4 with a sampling multiplying power of 2 through a convolution layer with a convolution kernel size of 1x1 and an output channel of 256 to obtain a characteristic diagram M3, and then the characteristic diagram P3 is output through the convolution layer with a convolution kernel size of 3x 3;
c2 is added with a characteristic map obtained by up-sampling M3 with a sampling multiplying power of 2 through a convolution layer output channel with a convolution kernel size of 1x1 and an output channel of 256 to obtain a characteristic map M2, and then the characteristic map P2 is output through a convolution layer with a convolution kernel size of 3x 3.
Determining the proportion and the size of the anchors, and extracting suggestion boxes from the feature maps P2, P3, P4 and P5.
The RPN firstly carries out 3X3 convolution on the feature map obtained by the feature extraction network to fuse the feature map information; then, according to anchors; setting two groups of parallel 1X1 convolutions, and respectively classifying and regressing the feature images; the classification layer carries out two classifications on each anchor box, judges whether the anchor box belongs to the foreground or the background, and outputs a result for each anchor on the feature map; the regression layer predicts the offset between each anchor box and the real labeling frame, and comprises the offsets deltax and deltay of the center coordinates and the offsets deltaw and deltah of the width and the height, so that the regression layer outputs a result for each anchor on the feature map; and finally, integrating the results of the two branches, selecting an anchor box with a classification layer prediction result as a foreground, and adjusting the central point and the length and the width of the anchor box by using the offset obtained by the regression layer calculation.
Inputting the suggestion boxes output by each feature map and the RPN area suggestion network into the RoI align for pooling synthesis to obtain suggestion boxes of each feature map with the size of 7x7, merging and inputting the suggestion boxes into the full-connection network to obtain 6-channel and 24-channel outputs, representing classification and regression results, and calculating target detection loss.
After the training of one round is completed, repeating the steps until the set iteration round is reached, outputting and storing parameters in the network, and obtaining the FaterRCNN target detection model.
In step five, an improved soft-NMS method based on imbalance data, comprising the operations of:
(a) Selecting a target detection frame with highest confidence in each type of targets, marking as M, and adding the M into the set D;
(b) Calculating the cross ratio between the rest target detection frames and M, and processing the confidence s of the target detection frames according to the following formula j The threshold is set in a segmented threshold mode, and the formula is as follows:
where j=1, 2, …, B (i) -1, represents the j-th detection frame of the i-th detection target, and B (i) is the number of detection frames of the i-th detection target; u (U) Iou (M,b j ) Is the intersection ratio of the detection frame M and the jth detection frame; k (i) is the balance coefficient of the unbalanced data;and alpha < beta;
(c) And (c) if all the types of target detection frames obtained in the step (b) are empty, finishing the screening of the detection frames.
According to the improved soft-NMS target detection method based on unbalanced data, which is provided by the invention, the phenomenon of long-tail data widely existing in the actual application of target detection is considered, the balance coefficient of the unbalanced data is introduced, the updating formula of the confidence coefficient of the target detection frame is redefined, the threshold is set in a sectional threshold mode, the influence of the unbalance of the data on the algorithm performance is reduced, the target detection frame of the tail type is protected from being excessively restrained, and the problem of target omission of the tail type is reduced.
It will be understood that the invention has been described in terms of several embodiments, and that various changes and equivalents may be made to these features and embodiments by those skilled in the art without departing from the spirit and scope of the invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to the particular embodiment disclosed, but that the invention will include all embodiments falling within the scope of the appended claims.

Claims (4)

1. An improved soft-NMS target detection method based on imbalance data, comprising the steps of:
step one, acquiring unbalanced training data of target detection, and calculating a balance coefficient of each type of target;
training a target detection model based on a Faster RCNN;
step three, inputting unbalanced target detection data to be detected into a Faster RCNN target detection model to obtain a plurality of target candidate frames;
step four, grouping all target candidate frames according to different category labels, and sorting all target candidate frames in each category according to confidence;
step five, for each category, adopting an improved soft-NMS method based on unbalanced data to update the confidence coefficient of a target detection frame in the category and screen the detection frame;
and step six, finally, screening target detection frames of all types of frames to be detected to obtain target detection results.
2. The improved soft-NMS object detection method based on unbalanced data of claim 1, wherein in the first step, the balance coefficient of each class of objects is calculated, comprising the following operations:
according to the characteristics of unbalanced data, calculating a balance coefficient k (i) of each type of target, wherein the formula is as follows;
where i=1, 2, …, C is the number of target categories in the imbalance data; n (i) is the total number of class i targets.
3. The improved soft-NMS object detection method based on unbalanced data of claim 1, wherein training the object detection model based on Faster RCNN in the second step comprises the following operations:
s1, inputting an unbalanced data training set picture into a Regnet backbone network, and extracting feature graphs C2, C3, C4 and C5 of 4 different stages;
s2, inputting the characteristic graphs C2, C3, C4 and C5 of the 4 different stages into an FPN characteristic pyramid for fusion to obtain new characteristic graphs P2, P3, P4 and P5 with 4 different resolutions;
s3, extracting suggestion frames of the feature graphs P2, P3, P4 and P5 through an RPN (remote procedure network) area suggestion network;
s4, inputting the feature graphs P2, P3, P4 and P5 and the suggestion boxes into the RoI align for pooling synthesis, and then classifying and regressing through a fully connected network to obtain a detection result and corresponding loss;
and S5, repeating the steps from S1 to S4 after the training of one round is completed until the set iteration round is reached, and outputting and storing parameters in a network to obtain a target detection model.
4. The improved soft-NMS object detection method based on unbalanced data according to claim 1, wherein in the fifth step, the improved soft-NMS method based on unbalanced data comprises the following operations:
(a) Selecting a target detection frame with highest confidence in each type of targets, marking as M, and adding the M into the set D;
(b) Calculating the cross ratio between the rest target detection frames and M, and processing the confidence s of the target detection frames according to the following formula j The threshold is set in a segmented threshold mode, and the formula is as follows:
where j=1, 2, …, B (i) -1, represents the j-th detection frame of the i-th detection target, and B (i) is the number of detection frames of the i-th detection target; u (U) Iou (M,b j ) Is the intersection ratio of the detection frame M and the jth detection frame; k (i) is the balance coefficient of the unbalanced data;and alpha < beta;
(c) And (c) if all the types of target detection frames obtained in the step (b) are empty, finishing the screening of the detection frames.
CN202311375631.7A 2023-10-23 2023-10-23 Improved soft-NMS target detection method based on unbalanced data Active CN117437465B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311375631.7A CN117437465B (en) 2023-10-23 2023-10-23 Improved soft-NMS target detection method based on unbalanced data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311375631.7A CN117437465B (en) 2023-10-23 2023-10-23 Improved soft-NMS target detection method based on unbalanced data

Publications (2)

Publication Number Publication Date
CN117437465A true CN117437465A (en) 2024-01-23
CN117437465B CN117437465B (en) 2024-06-07

Family

ID=89556256

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311375631.7A Active CN117437465B (en) 2023-10-23 2023-10-23 Improved soft-NMS target detection method based on unbalanced data

Country Status (1)

Country Link
CN (1) CN117437465B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104134080A (en) * 2014-08-01 2014-11-05 重庆大学 Method and system for automatically detecting roadbed collapse and side slope collapse of road
CN110348329A (en) * 2019-06-24 2019-10-18 电子科技大学 Pedestrian detection method based on video sequence interframe information
CN111882546A (en) * 2020-07-30 2020-11-03 中原工学院 Weak supervised learning-based three-branch convolutional network fabric defect detection method
CN114078106A (en) * 2020-08-06 2022-02-22 沈阳中科数控技术股份有限公司 Defect detection method based on improved Faster R-CNN
CN114359199A (en) * 2021-12-28 2022-04-15 集美大学 Fish counting method, device, equipment and medium based on deep learning
CN114821271A (en) * 2022-05-19 2022-07-29 平安科技(深圳)有限公司 Model training method, image description generation device and storage medium
CN115346135A (en) * 2022-08-19 2022-11-15 陕西航天技术应用研究院有限公司 Optical remote sensing image ship target identification method based on convolutional neural network
CN116561622A (en) * 2023-04-25 2023-08-08 广西师范大学 Federal learning method for class unbalanced data distribution

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104134080A (en) * 2014-08-01 2014-11-05 重庆大学 Method and system for automatically detecting roadbed collapse and side slope collapse of road
CN110348329A (en) * 2019-06-24 2019-10-18 电子科技大学 Pedestrian detection method based on video sequence interframe information
CN111882546A (en) * 2020-07-30 2020-11-03 中原工学院 Weak supervised learning-based three-branch convolutional network fabric defect detection method
CN114078106A (en) * 2020-08-06 2022-02-22 沈阳中科数控技术股份有限公司 Defect detection method based on improved Faster R-CNN
CN114359199A (en) * 2021-12-28 2022-04-15 集美大学 Fish counting method, device, equipment and medium based on deep learning
CN114821271A (en) * 2022-05-19 2022-07-29 平安科技(深圳)有限公司 Model training method, image description generation device and storage medium
CN115346135A (en) * 2022-08-19 2022-11-15 陕西航天技术应用研究院有限公司 Optical remote sensing image ship target identification method based on convolutional neural network
CN116561622A (en) * 2023-04-25 2023-08-08 广西师范大学 Federal learning method for class unbalanced data distribution

Also Published As

Publication number Publication date
CN117437465B (en) 2024-06-07

Similar Documents

Publication Publication Date Title
CN110084292B (en) Target detection method based on DenseNet and multi-scale feature fusion
CN108776772B (en) Cross-time building change detection modeling method, detection device, method and storage medium
CN110991311B (en) Target detection method based on dense connection deep network
CN110033481B (en) Method and apparatus for image processing
CN110826379B (en) Target detection method based on feature multiplexing and YOLOv3
KR102328734B1 (en) Method for automatically evaluating labeling reliability of training images for use in deep learning network to analyze images, and reliability-evaluating device using the same
CN113269073A (en) Ship multi-target tracking method based on YOLO V5 algorithm
US20120134586A1 (en) Device with datastream pipeline architecture for recognizing and locating objects in an image by detection window scanning
CN110147707B (en) High-precision vehicle identification method and system
CN109902576B (en) Training method and application of head and shoulder image classifier
CN115410100A (en) Small target detection method and system based on unmanned aerial vehicle image
CN114078197A (en) Small sample target detection method and device based on support sample characteristic enhancement
CN111414882B (en) Infrared target detection method based on multi-level resolution balance network
CN116363532A (en) Unmanned aerial vehicle image traffic target detection method based on attention mechanism and re-parameterization
CN114120138A (en) Method, device, equipment and medium for detecting and identifying remote sensing image target
CN113343989A (en) Target detection method and system based on self-adaption of foreground selection domain
CN113887455B (en) Face mask detection system and method based on improved FCOS
CN111553184A (en) Small target detection method and device based on electronic purse net and electronic equipment
CN113454649B (en) Target detection method, apparatus, electronic device, and computer-readable storage medium
CN111222575A (en) KLXS multi-model fusion method and system based on HRRP target recognition
CN110399868B (en) Coastal wetland bird detection method
CN117437465B (en) Improved soft-NMS target detection method based on unbalanced data
CN116977859A (en) Weak supervision target detection method based on multi-scale image cutting and instance difficulty
CN116206146A (en) Real-time detection method for lawn object, electronic device and storage medium
CN115272741A (en) Detection method of slender flexible object, terminal equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant