CN111553212A - Remote sensing image target detection method based on smooth frame regression function - Google Patents

Remote sensing image target detection method based on smooth frame regression function Download PDF

Info

Publication number
CN111553212A
CN111553212A CN202010302996.7A CN202010302996A CN111553212A CN 111553212 A CN111553212 A CN 111553212A CN 202010302996 A CN202010302996 A CN 202010302996A CN 111553212 A CN111553212 A CN 111553212A
Authority
CN
China
Prior art keywords
regression
frame
target detection
image
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010302996.7A
Other languages
Chinese (zh)
Other versions
CN111553212B (en
Inventor
申原
刘军
李洪忠
郭善昕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Institute of Advanced Technology of CAS
Original Assignee
Shenzhen Institute of Advanced Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Institute of Advanced Technology of CAS filed Critical Shenzhen Institute of Advanced Technology of CAS
Priority to CN202010302996.7A priority Critical patent/CN111553212B/en
Publication of CN111553212A publication Critical patent/CN111553212A/en
Priority to PCT/CN2020/140022 priority patent/WO2021208502A1/en
Application granted granted Critical
Publication of CN111553212B publication Critical patent/CN111553212B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Biology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Image Analysis (AREA)

Abstract

A remote sensing image target detection method based on a smooth frame regression function comprises the steps of carrying out necessary preprocessing on a training image and setting a hyper-parameter of network training; inputting the picture into a target detection type convolutional neural network to obtain a characteristic diagram; then inputting the feature map into a regional suggestion network to obtain a candidate frame; sending the candidate frame and the feature map into an interested region pooling layer to obtain interested region features, and classifying the interested region features in a classifier; sending the obtained region of interest characteristics into a full-connection layer to obtain a predicted offset, sending the predicted offset into a smooth frame regression function to obtain an actual offset, and correcting a candidate frame to a new position; repeating the steps until the training process is finished; and inputting the image to be detected into a trained network after preprocessing to obtain a target detection result. The regression of the high-precision frame can be effectively realized, and the target detection with higher precision can be realized under the condition of high IoU threshold value.

Description

Remote sensing image target detection method based on smooth frame regression function
Technical Field
The invention belongs to the field of image processing and machine learning, and relates to a remote sensing image target detection method based on a smooth frame regression function.
Background
With the rapid development of remote sensing technology, the remote sensing data volume is rapidly increased, and how to rapidly and efficiently process an original remote sensing image in the face of increasingly large and complex remote sensing information becomes an important research topic which can be understood and used by a user. The remote sensing image target detection is one of core tasks in remote sensing image understanding, the main purpose of the remote sensing image target detection is to quickly find and accurately locate an interested target in a remote sensing image, and the target detection is an important task and is also the basis of a plurality of tasks, such as example segmentation, image understanding and the like. However, the detection accuracy of the remote sensing image is low, and only the intersection ratio of the position of the prediction frame and the target real reference frame is larger than 0.5, the detection can be regarded as correct detection, but with the improvement of algorithm performance, people need to detect the target under the condition of more accurate positioning, so that high-quality detection is realized.
When the traditional target detection technology faces huge and complex remote sensing information, the detection precision, robustness and mobility are not good enough, the problems mentioned above cannot be solved, the requirements of human beings are difficult to meet, and a more efficient and accurate method is urgently needed. The deep learning is the most popular artificial intelligence basic technology at present, the strong representation learning ability of the deep learning can automatically learn the characteristics from big data, and the deep learning has strong robustness.
However, the current depth algorithm has the problems of inaccurate regression positioning, poor detection precision performance and the like. In general, in a network, a deep neural network is used for extracting features of a picture, and then a detector is used for detecting the features, but the detector is sensitive to feature fluctuation, so that robustness is not good enough, and regression effect is poor.
In the network, the regression process is mainly realized by a frame regression function. Bounding box regression is to make the candidate box go through regression to a position closer to the true reference box. Bounding box regression achieves regression by minimizing the difference from the true candidate box. The loss function of L2 used in RCNN was improved for smooth L1 in Fast RCNN. Along with the progress of the training process, the quality of the candidate frame is gradually improved and is closer to the real frame, the difference is smaller, the fluctuation is higher, the regression is more difficult to be stable, and particularly, the regression is ineffective due to continuous oscillation near the zero point, so that the regression precision is not high.
Disclosure of Invention
Based on the above, the invention provides a remote sensing image target detection method based on a smooth frame regression function, and aims to solve the technical problem of providing a smooth frame regression function, so that the regression process of fluctuation caused by too small difference between a candidate frame and a real reference frame becomes more stable, and higher regression precision and detection precision are obtained.
The invention provides a target detection method based on a smooth frame regression function, which comprises the following steps:
step one, image preprocessing: carrying out necessary preprocessing on a training image, including image rotation, mirror image and other enhancement operations, image normalization operation, image size adjustment operation and setting a hyper-parameter of network training;
step two, feature extraction: inputting the picture into a target detection type convolutional neural network to obtain a characteristic diagram; then inputting the feature map into a regional suggestion network to obtain a candidate frame; then the candidate frame and the feature map are sent into the region of interest pooling layer together to obtain the region of interest features;
step three, classification: sending the region of interest characteristics obtained in the step two into a softmax classifier for classification;
step four, regression: sending the region of interest characteristics obtained in the step three into a full-connection layer to obtain a predicted offset, sending the predicted offset into a smooth frame regression function to obtain an actual offset, and correcting the candidate frame to a new position according to the offset;
step five, correction: taking the bounding box after the candidate frame is subjected to regression correction as a new candidate frame, sending the new candidate frame and the feature map into an interested area layer together to obtain the features of the interested area, and repeating the third step, the fourth step and the fifth step until the training process is finished to obtain a trained network;
and step six, inputting the image to be detected into a trained network after preprocessing to obtain a target detection result.
Further, the smooth bounding box regression function for regression is:
(sgn((tx/cx))×|(tx/cx)|)4/3×pw+px=Gx
(sgn((ty/cy))×|(ty/cy)|)4/3×ph+py=Gy
exp(sgn((tw/cw))×|(tw/cw)|)4/3×pw=Gw
exp(sgn((th/ch))×|(th/ch)|)4/3×ph=Gh
wherein sgn represents a coincidence function, so as to ensure no error in negative number operation, exp is an exponential function, cx,cy,cw,chFor the weight adjustment value of the regression, tx,ty,th,twIs the offset, p, of the convolutional neural network predictionx,pyIs the position coordinate of the center point of the candidate frame, pw,phIs the width and height of the candidate box, Gx,GyIs the position coordinate of the center point of the bounding box after the regression correction, Gw,GhIs the width and height of the bounding box after the regression correction.
Further, the target detection convolutional neural network includes, but is not limited to, fast RCNN, YOLO v1, YOLO v2, YOLO v3, SSD, FPN, RetinaNet, Cascade RCNN.
The method has the advantages that the stability of the regression process can be enhanced by constructing the smooth frame regression function, the regression process which generates fluctuation due to too small difference between the candidate frame and the real reference frame becomes more stable, the problem of regression failure caused by continuous oscillation near the zero point is solved, the detection precision is higher under the high IoU threshold value, and the higher target detection precision is obtained.
Drawings
FIG. 1 is a graph comparing the tuning ranges of a smooth bounding box regression function and an original bounding box regression function;
FIG. 2 is an enlarged comparison plot of the smoothed bounding box regression function and the original bounding box regression function near zero;
FIG. 3 is a feature map visualization of the convolutional neural network output;
FIG. 4 is a schematic diagram of a regional proposal network architecture;
FIG. 5 is a schematic diagram of a cascade detector;
FIG. 6 is a graph of the results of Cascade RCNN method using the regression function of the original frame;
FIG. 7 is a diagram of the regression function detection results of the Cascade RCNN method using the smooth bounding box provided by the present invention;
fig. 8 is a schematic view of a work flow of a hyperspectral image retrieval method according to an embodiment of the invention.
Detailed Description
In order to make the technical problems, technical solutions and advantageous effects solved by the present invention more clearly apparent, the present invention is further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
With reference to fig. 1 and 2, the smooth bounding box regression function is described:
(sgn((tx/cx))×|(tx/cx)|)4/3×pw+px=Gx
(sgn((ty/cy))×|(ty/cy)|)4/3×ph+py=Gy
exp(sgn((tw/cw))×|(tw/cw)|)4/3×pw=Gw
exp(sgn((th/ch))×|(th/ch)|)4/3×ph=Gh
wherein sgn represents a coincidence function, and no error is ensured during negative number operation. exp is an exponential function, cx,cy,cw,chThe weight adjustment value is a regression whose value is usually (10, 10, 5, 5), t by defaultx,ty,th,twIs the offset, p, of the convolutional neural network predictionx,pyIs the position coordinate of the center point of the candidate frame, pw,phIs the width and height of the candidate box, Gx,GyIs the position coordinate of the center point of the bounding box after the regression correction, Gw,GhIs the width and height of the bounding box after the regression correction.
As shown in fig. 1 and 2, where the straight line represents the original regression function and the curve represents the modified bounding box function, fig. 2 is an enlarged view of fig. 1 as it approaches zero. It can be seen that the improved regression function is smoother near the true value, so that the bounding box tends to move to the true box after regression, and does not easily cross the true box, thereby enhancing the property of convergence.
The steps of the embodiment of the present invention will be described in detail below with Cascade RCNN as the basic network structure, as shown in FIG. 8. Cascade RCNN uses three cascaded detectors to achieve target detection.
1. Introduction to data set
The method is characterized in that the method is selected from a DOTA data set, the original DOTA data set is large in size and contains a large number of objects, pictures containing a large number of densely arranged small targets such as airplanes, ships, automobiles and the like are selected according to the requirements of the method, then the pictures are cut to a certain degree, the pictures are cut to be between 600 and 800, the needed data set is obtained, the training set comprises 15070 pictures, and the testing set comprises 2700 pictures.
2. Training process
1) Data pre-processing
Sending the picture into a network, and firstly carrying out operations such as horizontal mirroring, rotation and the like on the picture to enhance a data set; then, performing gray value normalization on the picture, and then scaling according to the size set by training, wherein the minimum edge size is usually set to be 600, and the maximum edge size is set to be 1000; the pictures are then screened and excluded if no object is present on the picture.
2) Training parameter settings
Training is carried out by using 4 GPUs, a frame used is cafe 2, a backbone network is resnet101, the minimum edge of a picture in the training is set to be 600, the maximum edge is limited to be 1000, the training mode uses SGD with momentum, the momentum is set to be 0.9, the initial learning rate is set to be 0.01, and the penalty term coefficient is 0.0001. Using segmented training, 360000 iterations are used herein, with a learning rate decay to 0.001,0.0001 at 240000,320000 times, respectively.
3) Feature extraction
And sequentially sending the preprocessed pictures into a convolutional neural network layer, carrying out operation calculation such as convolution and pooling on picture data through convolutional neural network, and extracting the features of the pictures for subsequent detection of a Cascade RCNN detector. Shown in fig. 3 is a visual display of a feature map of the convolutional neural network layer output.
4) Selection of candidate boxes
Features extracted from the convolutional neural network are input into an area suggestion network, and in the area suggestion network, a series of anchor points are preset for all areas on a picture in a sliding window mode, as shown in fig. 4. And screening all preset anchor points according to a foreground confidence degree sequencing mode to finally obtain the anchor point with the highest confidence degree as a candidate frame.
5) Iterative regression of cascaded detectors
After the feature map is obtained from the convolutional neural network, the feature map is sent to a detector to detect a target object. As shown in fig. 5, a Cascade detector structure of Cascade RCNN is shown, where B0 is a candidate region selected from a region-suggested network, conv represents a convolutional neural network, the candidate region and a feature map obtained from the convolutional neural network are sent to a ropooling layer together to obtain features of a region of interest, then the features are sent to a fully-connected layer (H1), and the features output by the fully-connected layer are respectively sent to a classifier (C1) for classification and fine-tuning positioning in a smooth bounding box regression function (B1) provided by the present invention. The network has three probes. And (3) taking the candidate box B1 of the previous layer which is subjected to fine tuning by the smooth border regression function provided by the invention as a new input, and sending the new input into the detector of the next layer until a candidate box B3 is obtained. And calculating the error between the B3 and the real frame as loss, carrying out backward propagation, and adjusting the parameters of the convolutional neural network. The above process is repeated until the training process is finished.
3. Test procedure
Firstly, preprocessing a test picture, zooming the test picture to the size set by a network, and carrying out gray value normalization. And then, the pictures are sequentially sent into a convolutional neural network to extract features to obtain a feature map. And inputting the extracted feature map into a regional suggestion network to obtain a candidate frame, and then inputting the candidate frame and the feature map into a region of interest pooling layer together to obtain the region of interest features. Sending the interesting region features into a first layer of a cascade detector to obtain regression offset, then calculating according to a smooth frame regression function to obtain the corrected position of a boundary frame, inputting the boundary frame serving as a new candidate frame and a feature map into an interesting region pooling layer to obtain new interesting region features, then inputting the new interesting region features into a detector of a second layer, and repeating the operation until the last layer of detectors. And a boundary frame obtained by the last layer of detectors after frame regression correction is the final boundary frame of the network. And simultaneously inputting the characteristics of the region of interest of the last layer into the classifier of each layer of detector to obtain a classification result, and then synthesizing the classification results of all the classifiers to obtain a final classification result of the network.
In order to verify the effectiveness of the remote sensing image target detection method, the DOTA data set is adopted for testing. In order to verify the effect of the smooth frame regression function provided by the invention, a YOLO v2, SSD, FasterRCNN, YOLO v3, RetinaNet, FPN and Cascade RCNN method is adopted for testing, the frame regression function of the original method is firstly adopted for calculation, then the frame regression function of the original method is replaced by the smooth frame regression function provided by the invention, and then the calculation is carried out again.
Evaluation indexes are as follows:
in the field of machine learning, evaluating the performance of a classifier is generally measured by two quantities, Precision (Precision) and Recall (Recall). To calculate these two indexes, samples can be classified into four categories according to the situation between the real value and the predicted value of the sample: true cases (True Positives, TP): predicting a positive sample as a positive example; false positives (FalsePositives, FP): predicting the positive sample as a negative example; true negative examples (True negotives, TN): predicting negative examples as counter examples; false negative (False Negatives, FN): predicting negative examples as positive examples; these four types of relationships can be clearly represented by a confusion matrix (consisionmatrix).
Precision P and recall R resolution are expressed as:
Figure BDA0002452959510000091
the higher the precision ratio and the recall ratio, the better, but the contradiction exists between the precision ratio and the recall ratio under the general condition, and the precision ratio is lower when the recall ratio is high; and when the precision ratio is high, the recall ratio is low. In general, we sort according to the scores of the classification, calculate the classification samples in sequence from high to low according to the scores, and count the current precision and recall. By taking the recall ratio as the horizontal axis and the precision ratio as the vertical axis, a curve can be made, which is called as a P-R curve. The performance can be reflected to a certain extent by calculating the area enclosed by the P-R curve and the transverse and longitudinal axes, and the performance is better when the area is higher.
The performance of the detector in target detection is measured by AP and mAP. For single-class target detection, Average Precision (AP) is generally sampled as an evaluation index. The AP for detecting the single type of target calculates the area enclosed by the P-R curve and the horizontal and vertical axes formed by the type. In target detection, determining four types of samples, namely TP, FP, TN and FN, requires calculation of IoU between each prediction frame and a real reference frame, and the sample can be determined as a positive sample only when the threshold is larger than a set threshold.
For a single class of target detection, the AP is expressed as:
Figure BDA0002452959510000092
in the present embodiment, the target detection accuracy is evaluated by using the respective matching methods using the indices AP, AP50, AP75, AP60, AP70, AP80, AP90, and the like. AP50 refers to the AP value when IoU was set to 0.5, the meaning of the remaining indicators is similar to AP 50. It can be seen that the difficulty is greater when IoU is higher, indicating that the accuracy of target detection is higher.
Fig. 6 and 7 show the target detection results of the original frame regression function and the smooth frame regression function provided by the present invention when using Cascade RCNN as the basic network architecture. FIG. 6 shows the result of the regression function of the original frame, and FIG. 7 shows the result of the regression function of the smoothed frame according to the present invention. It can be clearly seen that the positioning accuracy of the method of the present invention is higher than that of the regression function of the original frame.
Table 1 shows the comparison of the detection results of the method of the present invention and the original method under other network architectures. Where √ denotes the accuracy of using the smooth bounding box regression function provided by the present invention under a given network architecture, and absence of √ denotes that the original bounding box regression function is used.
TABLE 1 comparison of target detection accuracy under various network architectures
Figure BDA0002452959510000101
Figure BDA0002452959510000111
AP denotes the overall average accuracy indicator, AP50 denotes the average accuracy at a threshold of IoU greater than 0.5, and AP750 denotes the average accuracy at a threshold of IoU greater than 0.75, the same applies. It can be seen that after the smooth frame regression function provided by the present invention is adopted, the precision of each network architecture at each IoU level is obviously improved, and especially the detection improvement under the high IoU threshold value is higher, which indicates that the smooth frame regression function provided by the present invention has an obvious effect on improving the positioning precision, especially under the high IoU threshold value, the effect is more obvious.
By utilizing the technical scheme of the invention, the frame regression process in target detection can be better realized based on the smooth frame regression function, the detected target frame has better precision under the condition of high IoU threshold value, and compared with the original frame regression function, the smooth frame regression function provided by the invention can realize target detection with higher precision. The smooth frame regression function provided by the invention can be used in any target detection network framework.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.

Claims (3)

1. A remote sensing image target detection method based on a smooth frame regression function is characterized by comprising the following steps:
step one, image preprocessing: carrying out necessary preprocessing on a training image, including image rotation, mirror image and other enhancement operations, image normalization operation, image size adjustment operation and setting a hyper-parameter of network training;
step two, feature extraction: inputting the picture into a target detection type convolutional neural network to obtain a characteristic diagram; then inputting the feature map into a regional suggestion network to obtain a candidate frame; then the candidate frame and the feature map are sent into the region of interest pooling layer together to obtain the region of interest features;
step three, classification: sending the region of interest characteristics obtained in the step two into a softmax classifier for classification;
step four, regression: sending the region of interest characteristics obtained in the step three into a full-connection layer to obtain a predicted offset, sending the predicted offset into a smooth frame regression function to obtain an actual offset, and correcting the candidate frame to a new position according to the offset;
step five, correction: taking the bounding box after the candidate frame is subjected to regression correction as a new candidate frame, sending the new candidate frame and the feature map into an interested area layer together to obtain the features of the interested area, and repeating the third step, the fourth step and the fifth step until the training process is finished to obtain a trained network;
and step six, inputting the image to be detected into a trained network after preprocessing to obtain a target detection result.
2. The method for detecting the target of the remote sensing image based on the regression function of the smooth border of claim 1, wherein the regression function of the smooth border for regression is as follows:
(sgn((tx/cx))×|(tx/cx)|)4/3×pw+px=Gx
(sgn((ty/cy))×|(ty/cy)|)4/3×ph+py=Gy
exp(sgn((tw/cw))×|(tw/cw)|)4/3×pw=Gw
exp(sgn((th/ch))×|(th/ch)|)4/3×ph=Gh
wherein sgn represents a coincidence function, so as to ensure no error in negative number operation, exp is an exponential function, cx,cy,cw,chFor the weight adjustment value of the regression, tx,ty,th,twIs the offset, p, of the convolutional neural network predictionx,pyIs the position coordinate of the center point of the candidate frame, pw,phIs the width and height of the candidate box, Gx,GyIs the edge after the regression correctionPosition coordinates of the center point of the bounding box, Gw,GhIs the width and height of the bounding box after the regression correction.
3. The method as claimed in claim 1, wherein the object detection convolutional neural network includes but is not limited to fast RCNN, YOLO v1, YOLO v2, YOLO v3, SSD, FPN, RetinaNet, Cascade RCNN.
CN202010302996.7A 2020-04-16 2020-04-16 Remote sensing image target detection method based on smooth frame regression function Active CN111553212B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202010302996.7A CN111553212B (en) 2020-04-16 2020-04-16 Remote sensing image target detection method based on smooth frame regression function
PCT/CN2020/140022 WO2021208502A1 (en) 2020-04-16 2020-12-28 Remote-sensing image target detection method based on smooth bounding box regression function

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010302996.7A CN111553212B (en) 2020-04-16 2020-04-16 Remote sensing image target detection method based on smooth frame regression function

Publications (2)

Publication Number Publication Date
CN111553212A true CN111553212A (en) 2020-08-18
CN111553212B CN111553212B (en) 2022-02-22

Family

ID=72005720

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010302996.7A Active CN111553212B (en) 2020-04-16 2020-04-16 Remote sensing image target detection method based on smooth frame regression function

Country Status (2)

Country Link
CN (1) CN111553212B (en)
WO (1) WO2021208502A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112132033A (en) * 2020-09-23 2020-12-25 平安国际智慧城市科技股份有限公司 Vehicle type recognition method and device, electronic equipment and storage medium
CN112232180A (en) * 2020-10-14 2021-01-15 上海海洋大学 Night underwater fish target detection method
CN112464769A (en) * 2020-11-18 2021-03-09 西北工业大学 High-resolution remote sensing image target detection method based on consistent multi-stage detection
CN112560682A (en) * 2020-12-16 2021-03-26 重庆守愚科技有限公司 Valve automatic detection method based on deep learning
WO2021208502A1 (en) * 2020-04-16 2021-10-21 中国科学院深圳先进技术研究院 Remote-sensing image target detection method based on smooth bounding box regression function
CN115035552A (en) * 2022-08-11 2022-09-09 深圳市爱深盈通信息技术有限公司 Fall detection method and device, equipment terminal and readable storage medium
CN118196401A (en) * 2024-05-17 2024-06-14 南昌大学 Target detection method, target detection system, storage medium and electronic equipment

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113920375B (en) * 2021-11-01 2024-06-25 国网新疆电力有限公司营销服务中心(资金集约中心、计量中心) Fusion characteristic typical load identification method and system based on combination of Faster R-CNN and SVM
CN114707532B (en) * 2022-01-11 2023-05-19 中铁隧道局集团有限公司 Improved Cascade R-CNN-based ground penetrating radar tunnel disease target detection method
CN114792300B (en) * 2022-01-27 2024-02-20 河南大学 X-ray broken needle detection method based on multi-scale attention
CN114529552A (en) * 2022-03-03 2022-05-24 北京航空航天大学 Remote sensing image building segmentation method based on geometric contour vertex prediction
CN114925387B (en) * 2022-04-02 2024-06-07 北方工业大学 Sorting system, method and readable storage medium based on end-edge cloud architecture
CN114757970B (en) * 2022-04-15 2024-03-08 合肥工业大学 Sample balance-based multi-level regression target tracking method and tracking system
CN115170883B (en) * 2022-07-19 2023-03-14 哈尔滨市科佳通用机电股份有限公司 Brake cylinder piston push rod opening pin loss fault detection method
CN116645523B (en) * 2023-07-24 2023-12-01 江西蓝瑞存储科技有限公司 Rapid target detection method based on improved RetinaNet

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020012451A1 (en) * 2000-06-13 2002-01-31 Ching-Fang Lin Method for target detection and identification by using proximity pixel information
CN108052940A (en) * 2017-12-17 2018-05-18 南京理工大学 SAR remote sensing images waterborne target detection methods based on deep learning
CN108564109A (en) * 2018-03-21 2018-09-21 天津大学 A kind of Remote Sensing Target detection method based on deep learning
CN109800755A (en) * 2018-12-14 2019-05-24 中国科学院深圳先进技术研究院 A kind of remote sensing image small target detecting method based on Analysis On Multi-scale Features
CN110211097A (en) * 2019-05-14 2019-09-06 河海大学 Crack image detection method based on fast R-CNN parameter migration
CN110288017A (en) * 2019-06-21 2019-09-27 河北数云堂智能科技有限公司 High-precision cascade object detection method and device based on dynamic structure optimization
CN110956157A (en) * 2019-12-14 2020-04-03 深圳先进技术研究院 Deep learning remote sensing image target detection method and device based on candidate frame selection

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111553212B (en) * 2020-04-16 2022-02-22 中国科学院深圳先进技术研究院 Remote sensing image target detection method based on smooth frame regression function

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020012451A1 (en) * 2000-06-13 2002-01-31 Ching-Fang Lin Method for target detection and identification by using proximity pixel information
CN108052940A (en) * 2017-12-17 2018-05-18 南京理工大学 SAR remote sensing images waterborne target detection methods based on deep learning
CN108564109A (en) * 2018-03-21 2018-09-21 天津大学 A kind of Remote Sensing Target detection method based on deep learning
CN109800755A (en) * 2018-12-14 2019-05-24 中国科学院深圳先进技术研究院 A kind of remote sensing image small target detecting method based on Analysis On Multi-scale Features
CN110211097A (en) * 2019-05-14 2019-09-06 河海大学 Crack image detection method based on fast R-CNN parameter migration
CN110288017A (en) * 2019-06-21 2019-09-27 河北数云堂智能科技有限公司 High-precision cascade object detection method and device based on dynamic structure optimization
CN110956157A (en) * 2019-12-14 2020-04-03 深圳先进技术研究院 Deep learning remote sensing image target detection method and device based on candidate frame selection

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
CAI ZHAOWEI 等: "Cascade R-CNN: Delving into High Quality Object Detection", 《2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021208502A1 (en) * 2020-04-16 2021-10-21 中国科学院深圳先进技术研究院 Remote-sensing image target detection method based on smooth bounding box regression function
CN112132033A (en) * 2020-09-23 2020-12-25 平安国际智慧城市科技股份有限公司 Vehicle type recognition method and device, electronic equipment and storage medium
CN112132033B (en) * 2020-09-23 2023-10-10 平安国际智慧城市科技股份有限公司 Vehicle type recognition method and device, electronic equipment and storage medium
CN112232180A (en) * 2020-10-14 2021-01-15 上海海洋大学 Night underwater fish target detection method
CN112464769A (en) * 2020-11-18 2021-03-09 西北工业大学 High-resolution remote sensing image target detection method based on consistent multi-stage detection
CN112560682A (en) * 2020-12-16 2021-03-26 重庆守愚科技有限公司 Valve automatic detection method based on deep learning
CN115035552A (en) * 2022-08-11 2022-09-09 深圳市爱深盈通信息技术有限公司 Fall detection method and device, equipment terminal and readable storage medium
CN118196401A (en) * 2024-05-17 2024-06-14 南昌大学 Target detection method, target detection system, storage medium and electronic equipment
CN118196401B (en) * 2024-05-17 2024-07-19 南昌大学 Target detection method, target detection system, storage medium and electronic equipment

Also Published As

Publication number Publication date
CN111553212B (en) 2022-02-22
WO2021208502A1 (en) 2021-10-21

Similar Documents

Publication Publication Date Title
CN111553212B (en) Remote sensing image target detection method based on smooth frame regression function
CN110070074B (en) Method for constructing pedestrian detection model
CN112101430B (en) Anchor frame generation method for image target detection processing and lightweight target detection method
CN110796186A (en) Dry and wet garbage identification and classification method based on improved YOLOv3 network
CN104680542B (en) Remote sensing image variation detection method based on on-line study
CN111723798B (en) Multi-instance natural scene text detection method based on relevance hierarchy residual errors
CN111598156B (en) PM based on multi-source heterogeneous data fusion2.5Prediction method
CN111091095B (en) Method for detecting ship target in remote sensing image
CN111738258A (en) Pointer instrument reading identification method based on robot inspection
CN108446616B (en) Road extraction method based on full convolution neural network ensemble learning
CN109284779A (en) Object detection method based on deep full convolution network
CN114694178A (en) Method and system for monitoring safety helmet in power operation based on fast-RCNN algorithm
CN111860587A (en) Method for detecting small target of picture
CN115953666B (en) Substation site progress identification method based on improved Mask-RCNN
CN114998566A (en) Interpretable multi-scale infrared small and weak target detection network design method
CN115761534A (en) Method for detecting and tracking small target of infrared unmanned aerial vehicle under air background
CN115661569A (en) High-precision fine-grained SAR target detection method
CN114677602A (en) Front-view sonar image target detection method and system based on YOLOv5
CN114529552A (en) Remote sensing image building segmentation method based on geometric contour vertex prediction
CN117333845A (en) Real-time detection method for small target traffic sign based on improved YOLOv5s
CN115861204A (en) Smartphone panel surface defect detection method based on YOLO V5 model
CN110276358A (en) High similarity wooden unit cross section detection method under intensive stacking
CN114155411A (en) Intelligent detection and identification method for small and weak targets
CN114972153A (en) Bridge vibration displacement visual measurement method and system based on deep learning
Liu et al. Infrared dim small target detection based on regional refinement network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant