CN113158943A - Cross-domain infrared target detection method - Google Patents

Cross-domain infrared target detection method Download PDF

Info

Publication number
CN113158943A
CN113158943A CN202110474134.7A CN202110474134A CN113158943A CN 113158943 A CN113158943 A CN 113158943A CN 202110474134 A CN202110474134 A CN 202110474134A CN 113158943 A CN113158943 A CN 113158943A
Authority
CN
China
Prior art keywords
network
domain data
cnn
target
mask
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202110474134.7A
Other languages
Chinese (zh)
Inventor
颜成钢
路统宇
戴振宇
孙垚棋
张继勇
李宗鹏
张勇东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Dianzi University
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN202110474134.7A priority Critical patent/CN113158943A/en
Publication of CN113158943A publication Critical patent/CN113158943A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a cross-domain infrared target detection method, which comprises the following steps of firstly obtaining source domain data with labels and target domain data with a small number of labels and without labels: and then training a Mask R-CNN-1 network by using source domain data containing tags: carrying out domain adaptive training on a new Mask R-CNN-2 by utilizing the source domain data and the target domain data: and finally, inputting the infrared image to be subjected to target detection into the trained Mask R-CNN-2 network, and finally realizing the target detection function. The method of the invention utilizes the domain adaptation technology, solves the problem of poor network training effect under the condition of insufficient target domain data labels, and improves the accuracy of target detection; the defect that the domain adaptation task is only carried out on the characteristic level is overcome, a Mask R-CNN network is used, and pixel-level target detection is further realized on the basis of target detection.

Description

Cross-domain infrared target detection method
Technical Field
The invention belongs to the technical field of computer vision and pattern recognition, relates to a cross-domain infrared target detection method, and particularly relates to a semi-supervised cross-domain infrared target detection method based on a feature map and a CAM (computer-aided manufacturing) map layer, which can align a source domain and a target domain on the feature map layer and the CAM map layer to solve the problem of few labels of a target domain training set, namely, the network training task which takes the target domain as a target is assisted by source domain data, and the network training and the realization of infrared image target detection are realized under the condition of few labels of the target domain data.
Background
Since the human eye can perceive light in the wavelength range of 0.43 μm to 0.79 μm, light outside this range needs to be detected by a specialized detector. The infrared ray is an electromagnetic wave with the wavelength between 750nm and 1mm, the infrared imaging is realized by receiving the infrared radiation generated by an object and converting the infrared radiation into an infrared image through a sensor, and the infrared imaging has the outstanding characteristics of long detection distance, high concealment, capability of working day and night and the like, not only extends the human visual system, but also can greatly make up the defect that the visible light imaging is limited by the environmental influence, has very important status in the field of target detection, and is widely applied to civil and military scenes. For example, in public gathering places such as railway stations, airports, markets and the like and battlefield environments, targets are generally required to be identified and detected accurately in time, and compared with visible light imaging, infrared imaging can provide a good supplementary effect for target detection in a complicated field of view.
The traditional infrared target detection algorithm is to manually design a target template, perform template matching on a target, and determine the position of the target according to the compared similarity, and the main detection methods include a gradient histogram algorithm, a scale invariant feature detection algorithm, a haar feature algorithm and the like. With the development of deep neural networks in recent years, many scholars propose different target detection algorithms, the basic principle of which is to abstract and extract the features of an image through convolution operation, such as RCNN algorithm, Fast R-CNN algorithm, Mask R-CNN algorithm and the like.
However, the existing target detection methods such as Fast R-CNN and Mask R-CNN are all fully supervised target detection algorithms, a large amount of labeled training data is needed, and under the condition of less labeled training data, an overfitting phenomenon often occurs, so that the final performance of the network is reduced; meanwhile, compared with the visible light image, the labeling cost of the infrared image is higher, the infrared image data set with the label is less, the contrast of the infrared image is lower, the detail information is lacked, the difference between the infrared image data set and the visible light image is larger, and if the infrared image is directly subjected to target detection by a network trained by the visible light image, a good effect cannot be achieved. The domain adaptation technology is used as a method for reducing inter-domain differences, source domain data with rich labels can be used as assistance to constrain target domain features extracted by a network, so that the feature distribution of the target domain features is close to that of the source domain data, and the purpose of improving the detection precision is achieved.
Disclosure of Invention
Aiming at the defects in the prior art, the invention provides a cross-domain infrared target detection method.
The method is an infrared target detection method based on a characteristic diagram and a CAM (class Activation map) diagram level semi-supervised cross-domain.
A cross-domain infrared target detection method comprises the following steps:
step 1, acquiring source domain data with labels and target domain data with a small number of labels and without labels:
the source domain data and the target domain data are respectively visible light images and infrared images, the related scenes of the source domain data and the target domain data are similar, and the targets are the same. The labeled source domain is denoted as S ═ { Xs,YsIn which xs∈XsAnd ys∈YsRespectively representing source domain data and corresponding tags; the target domain is composed of two parts, the target domain T containing a label1={Xt1,Yt1In which xt1∈Xt1And yt1∈Yt1Respectively representing target domain data and corresponding labels, and a target domain T without labels2={Xt2In which xt2∈Xt2Representing the target domain data.
Step 2, training a Mask R-CNN-1 network by using source domain data containing labels:
inputting labeled source domain data into a Mask R-CNN-1 network for full supervision training, wherein a loss function of the network is divided into two parts, the first part is RPN foreground/background classification loss and RPN target frame regression loss, and the two parts can be expressed as LRPNThe second part is Mask R-CNN head loss, including classification loss, regression loss and pixel segmentation loss, which can be expressed as LHEADI.e. the total loss function of the Mask R-CNN network is expressed as Lsupervised=LRPN+LHEADThe network minimizes the total loss function by adopting a random gradient descent algorithm, and returns the parameters to the convolution network, the RPN network and the head structure of the Mask R-CNN-1, and the trained Mask R-CNN-1 network can be obtained after the training is finished.
Step 3, performing domain adaptive training on a new Mask R-CNN-2 by using the source domain data and the target domain data:
fixing the convolutional network backbone-1 in the Mask R-CNN-1 network trained in the step 2, and respectively inputting the source domain data and the target domain data into the backbone-1 and the Mask R-CNN-2 to obtain a characteristic diagram F of the source domain datasAnd feature map F of target domain datatAnd respectively carrying out CAM (class Activation mapping) operation on the feature graphs of the two-domain data to obtain an attention graph A of the two-domain datasAnd AtWherein
Figure BDA0003046734980000041
FjIs the jth channel, ω, of the feature map FjIs FjWeight value obtained after GAP (Global Average Pooling) operation. In the obtained FsAnd FtAnd AsAnd AtRespectively adding a feature discriminator DFAnd attention discriminator DAThe structure is used for countertraining, so that the feature distribution and the attention distribution of the target domain data are close to the feature distribution and the attention distribution of the source domain data to reduce the inter-domain gap, wherein the discriminator is a multilayer convolution structure, and the loss functions are respectively as follows:
Figure BDA0003046734980000042
Figure BDA0003046734980000043
and 4, carrying out target detection on the infrared image:
inputting the infrared image to be subjected to target detection into the Mask R-CNN-2 network trained in the step 3, obtaining a characteristic diagram of the infrared image through forward propagation, and sending the characteristic diagram into an RPN network to obtain the position, the category and the Mask information of the target, thereby finally realizing the target detection function.
The invention has the following beneficial effects:
1. by using a domain adaptation technology, the problem of poor network training effect under the condition of insufficient target domain data labels is solved, and the target detection accuracy is improved to a certain extent;
2. the method has the advantages that the region which has decisive significance on the target detection task is focused by using the attention mechanism while the region adaptation is carried out on the characteristic layer, the region adaptation task is carried out on the CAM layer surface, and the defect that the region adaptation task is carried out only on the characteristic layer is overcome.
3. The Mask R-CNN network is used, Mask (Mask) generation branches are added on the basis of the fast R-CNN network, pixel-level target detection is further realized on the basis of target detection, namely, each pixel point of a target can be identified to perform more accurate image target segmentation, and realization possibility is provided for downstream tasks such as human body key point detection, unmanned driving and other target identification tasks needing more accuracy.
Drawings
FIG. 1 is a flow chart of a target detection algorithm proposed by the present invention;
FIG. 2 is a schematic diagram of a pre-training Mask R-CNN-1 network structure;
FIG. 3 is a schematic diagram of Mask R-CNN-2 network domain adaptive training structure;
FIG. 4 is a schematic diagram of target detection based on a trained Mask R-CNN-2 network.
Detailed Description
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
At present, most target detection tasks can detect the position of a target, but foreground pixels and background pixels at the position of the target cannot be distinguished, and good effects cannot be achieved in some tasks needing more accurate target detection, a Mask R-CNN Network is adopted to realize pixel-level target detection, the target detection is evolved from fast R-CNN, a Mask generation branch is added on the basis of detecting the position and the type of the target, the structure of the target detection branch mainly comprises a Convolutional Network (Convolutional Back), a Region recommendation Network (Region Proposal Network), namely an RPN Network, an FPN Network (Feature Pyramid Network), a RoI Align structure and a multi-task head structure, wherein the Convolutional Network is a series of Convolutional layers for extracting image Feature maps, and the common structures comprise VGG19, Res 50 and the like; the area recommendation network is used to help the network recommend areas of interest, which consist of two convolutional layers. The process of detecting the target by the Mask R-CNN network is mainly divided into two stages, wherein the first stage is to divide an image alternative area, and the second stage is to detect the target in the alternative area; meanwhile, the cost for acquiring the infrared image is high, the infrared image data set with the label is less, and a large amount of manpower and financial resources are consumed for labeling the infrared image, so that an over-fitting phenomenon is often caused if a traditional full-supervision type network training method is performed, a good training effect cannot be achieved, and the method is difficult to be applied in practice; in addition, although the domain adaptation is well applied to the small-sample semi-supervised target detection task, the most domain adaptation is to set constraints between the feature maps of the source domain and the target domain, so that the feature distribution of the target domain data with few labels is close to the feature distribution of the source domain data with a large number of labels, and the region which has decisive significance on the target detection task in the feature maps is not considered.
Aiming at the problems, the invention provides a cross-domain indoor infrared human body detection algorithm, a flow chart of which is shown in figure 1, and the steps are as follows:
step (1), acquiring source domain data with labels and target domain data with a small number of labels and without labels:
the source domain data and the target domain data are respectively visible light images and infrared images, the scenes related to the source domain data and the target domain data are both indoor key areas, and the target is a human body. The labeled source domain is denoted as S ═ { Xs,YsIn which xs∈XsAnd ys∈YsRespectively representing a source domain image and a corresponding human body target label; the target domain is composed of two parts, the target domain T containing a label1={Xt1,Yt1In which xt1∈Xt1And yt1∈Yt1Respectively representing an image of the target field and a corresponding label of the target of the human body, and a target field T without label2={Xt2In which xt2∈Xt2Representing the target domain image. Wherein the number of images in the source domain data is more than 2 times the number of images in the target domain.
Step (2), training a Mask R-CNN-1 network by using source domain data containing labels: referring to fig. 2, the labeled source domain data is input into a Mask R-CNN-1 network for full supervision training, wherein a convolutional network in the Mask R-CNN-1 network is composed of a ResNet50 network and a FPN network and generates a feature map, and then the convolutional network is obtainedThe feature map input RPN network of (1) including two convolutional layers to generate a region Proposal (RoI), and then inputting the feature map and the region Proposal into a RoI Align structure, the feature obtained by each RoI is better aligned to the RoI region on the original map by bilinear interpolation. Parameters of the ResNet50 network are initialized using the pre-trained model, and parameters of FPN, PRN, and RoI Align are initialized randomly. The loss function of the network can be divided into two parts, the first part is RPN foreground/background classification loss and RPN target frame regression loss, which can be expressed as LRPNThe second part is Mask R-CNN head loss, including classification loss, regression loss and pixel segmentation loss, which can be expressed as LHEAD, i.e. the total loss function of the Mask R-CNN network is expressed as Lsupervised=LRPN+LHEADThe network adopts a random gradient descent algorithm to minimize a total loss function, and transmits parameters back to a Mask R-CNN convolution network, an RPN network and a RoI Align structure, the learning rate of the random gradient descent algorithm is set to be 0.0001, and the iteration number is set to be 30. And obtaining the trained Mask R-CNN-1 network after the training is finished.
And (3) performing domain adaptive training on a new Mask R-CNN network Mask R-CNN-2 by using the source domain data and the target domain data: as shown in FIG. 3, the structure and initialization parameter setting of each part in the Mask R-CNN-2 network are the same as those of the Mask R-CNN-1 network. Fixing the convolutional network backbone-1 in the Mask R-CNN-1 network trained in the step 2, and respectively inputting the source domain data and the target domain data into the backbone-1 and the Mask R-CNN-2 to obtain a characteristic diagram F of the source domain datasAnd feature map F of target domain datatAnd respectively carrying out CAM (class Activation mapping) operation on the feature graphs of the two-domain data to obtain an attention graph A of the two-domain datasAnd AtWherein
Figure BDA0003046734980000081
FjIs the jth channel, ω, of the feature map FjIs FjWeight value obtained after GAP (Global Average Pooling) operation. In the obtained FsAnd FtAnd AsAnd AtRespectively adding a feature discriminator DFAnd attention discriminator DAThe structure is used for countertraining, so that the feature distribution and the attention distribution of the target domain data are close to the feature distribution and the attention distribution of the source domain data to reduce the inter-domain gap, wherein the discriminator is a multilayer convolution structure, and the loss functions are respectively as follows:
Figure BDA0003046734980000082
and:
Figure BDA0003046734980000083
the arbiter network adopts a random gradient descent algorithm to minimize a total loss function, and transmits parameters back to a Mask R-CNN-2 convolution network, an RPN network and a RoI Align structure, the learning rate of the random gradient descent algorithm is set to be 0.0001, and the iteration number is set to be 30. The overall loss function of the domain adaptation part is:
Figure BDA0003046734980000084
where α and β are the feature domain adaptation loss weight and the attention domain adaptation weight, 0.4 and 0.6, respectively.
And (4) carrying out target detection on the infrared image: as shown in fig. 4, the infrared image to be subjected to human body detection is input to the Mask R-CNN-2 network trained in step 3, the characteristic diagram of the infrared image is obtained by forward propagation to the convolutional network, and is sent to the RPN network to obtain the position, category and Mask information of the human body, so that the function of detecting the human body in the indoor key area scene is finally realized.

Claims (4)

1. A cross-domain infrared target detection method is characterized by comprising the following steps:
step 1, acquiring source domain data with labels and target domain data with a small number of labels and without labels:
source domain data and target domain numberAccording to the visible light image and the infrared image, the related scenes of the visible light image and the infrared image are similar, and the targets are the same; the labeled source domain is denoted as S ═ { Xs,YsIn which xs∈XsAnd ys∈YsRespectively representing source domain data and corresponding tags; the target domain is composed of two parts, the target domain T containing a label1={Xt1,Yt1In which xt1∈Xt1And yt1∈Yt1Respectively representing target domain data and corresponding labels, and a target domain T without labels2={Xt2In which xt2∈Xt2Representing target domain data;
step 2, training a MaskR-CNN-1 network by using source domain data containing labels:
inputting the labeled source domain data into a MaskR-CNN-1 network for full supervision training, wherein a loss function of the network is divided into two parts, the first part is RPN foreground/background classification loss and RPN target frame regression loss, and the two parts can be expressed as LRPNThe second part is Mask R-CNN head loss, including classification loss, regression loss and pixel segmentation loss, which can be expressed as LHEADI.e. the overall loss function of the MaskR-CNN network is expressed as Lsupervised=LRPN+LHEADThe network minimizes the total loss function by adopting a random gradient descent algorithm, and returns parameters to the convolution network, the RPN network and the head structure of the MaskR-CNN-1, and the trained MaskR-CNN-1 network can be obtained after the training is finished;
step 3, performing domain adaptive training on a new MaskR-CNN-2 by using the source domain data and the target domain data:
fixing the convolutional network backbone-1 in the MaskR-CNN-1 network trained in the step 2, respectively inputting the source domain data and the target domain data to the backbone-1 and the Mask R-CNN-2 to obtain a characteristic diagram F of the source domain datasAnd feature map F of target domain datatAnd respectively carrying out CAM (class Activation mapping) operation on the feature graphs of the two-domain data to obtain an attention graph A of the two-domain datasAnd AtWherein
Figure FDA0003046734970000021
FjIs the jth channel, ω, of the feature map FjIs FjWeight value obtained after GAP (Global Average Power) operation; in the obtained FsAnd FtAnd AsAnd AtRespectively adding a feature discriminator DFAnd attention discriminator DAThe structure is used for countertraining, so that the feature distribution and the attention distribution of the target domain data are close to the feature distribution and the attention distribution of the source domain data to reduce the inter-domain gap, wherein the discriminator is a multilayer convolution structure, and the loss functions are respectively as follows:
Figure FDA0003046734970000022
Figure FDA0003046734970000023
and 4, carrying out target detection on the infrared image:
inputting the infrared image to be subjected to target detection into the Mask R-CNN-2 network trained in the step 3, obtaining a characteristic diagram of the infrared image through forward propagation, and sending the characteristic diagram into an RPN network to obtain the position, the category and the Mask information of the target, thereby finally realizing the target detection function.
2. The method of claim 1, wherein the number of images in the source domain data is more than 2 times the number of images in the target domain.
3. The method for detecting the cross-domain infrared target according to claim 1 or 2, characterized in that the specific method in the step (2) is as follows:
inputting the labeled source domain data into a Mask R-CNN-1 network for full supervision training, wherein a convolution network in the Mask R-CNN-1 network consists of a ResNet50 network and an FPN network and generates a feature map, and then convolvingInputting a feature map obtained by a network into an RPN network, wherein the feature map comprises two convolution layers to generate a region Proposal (RoI Proposal), then inputting the feature map and the region Proposal into a RoI Align structure, and enabling the feature obtained by each RoI to be better aligned with the RoI region on an original map through bilinear interpolation; initializing parameters of a ResNet50 network by using a pre-training model, and randomly initializing parameters of FPN, PRN and RoI Align; the loss function of the network can be divided into two parts, the first part is RPN foreground/background classification loss and RPN target frame regression loss, which can be expressed as LRPNThe second part is Mask R-CNN head loss, including classification loss, regression loss and pixel segmentation loss, which can be expressed as LHEADI.e. the total loss function of the Mask R-CNN network is expressed as Lsupervised=LRPN+LHEADThe network adopts a random gradient descent algorithm to minimize a total loss function, and transmits parameters back to a Mask R-CNN convolution network, an RPN network and a RoI Align structure, the learning rate of the random gradient descent algorithm is set to be 0.0001, and the iteration times are set to be 30; and obtaining the trained Mask R-CNN-1 network after the training is finished.
4. The method for detecting the cross-domain infrared target according to claim 3, wherein the specific method in the step (3) is as follows:
the structure and the initialization parameter setting of each part in the Mask R-CNN-2 network are the same as those of the Mask R-CNN-1 network; fixing the convolutional network backbone-1 in the Mask R-CNN-1 network trained in the step 2, and respectively inputting the source domain data and the target domain data into the backbone-1 and the Mask R-CNN-2 to obtain a characteristic diagram F of the source domain datasAnd feature map F of target domain datatAnd respectively carrying out CAM (class Activation mapping) operation on the feature graphs of the two-domain data to obtain an attention graph A of the two-domain datasAnd AtWherein
Figure FDA0003046734970000031
FjIs the jth channel, ω, of the feature map FjIs FjWeight value obtained after GAP (Global Average Power) operation; in the obtained FsAnd FtAnd AsAnd AtRespectively adding a feature discriminator DFAnd attention discriminator DAThe structure is used for countertraining, so that the feature distribution and the attention distribution of the target domain data are close to the feature distribution and the attention distribution of the source domain data to reduce the inter-domain gap, wherein the discriminator is a multilayer convolution structure, and the loss functions are respectively as follows:
Figure FDA0003046734970000041
and:
Figure FDA0003046734970000042
the arbiter network adopts a random gradient descent algorithm to minimize a total loss function, and transmits parameters back to a Mask R-CNN-2 convolution network, an RPN network and a RoI Align structure, the learning rate of the random gradient descent algorithm is set to be 0.0001, and the iteration times are set to be 30; the overall loss function of the domain adaptation part is:
Figure FDA0003046734970000043
where α and β are the feature domain adaptation loss weight and the attention domain adaptation weight, 0.4 and 0.6, respectively.
CN202110474134.7A 2021-04-29 2021-04-29 Cross-domain infrared target detection method Withdrawn CN113158943A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110474134.7A CN113158943A (en) 2021-04-29 2021-04-29 Cross-domain infrared target detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110474134.7A CN113158943A (en) 2021-04-29 2021-04-29 Cross-domain infrared target detection method

Publications (1)

Publication Number Publication Date
CN113158943A true CN113158943A (en) 2021-07-23

Family

ID=76872403

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110474134.7A Withdrawn CN113158943A (en) 2021-04-29 2021-04-29 Cross-domain infrared target detection method

Country Status (1)

Country Link
CN (1) CN113158943A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113554013A (en) * 2021-09-22 2021-10-26 华南理工大学 Cross-scene recognition model training method, cross-scene road recognition method and device
CN113591773A (en) * 2021-08-10 2021-11-02 武汉中电智慧科技有限公司 Power distribution room object detection method, device and equipment based on convolutional neural network
CN113780135A (en) * 2021-08-31 2021-12-10 中国科学技术大学先进技术研究院 Cross-scene VOCs gas leakage detection method and system and storage medium
CN113807420A (en) * 2021-09-06 2021-12-17 湖南大学 Domain self-adaptive target detection method and system considering category semantic matching
CN114693983A (en) * 2022-05-30 2022-07-01 中国科学技术大学 Training method and cross-domain target detection method based on image-instance alignment network
WO2023092582A1 (en) * 2021-11-25 2023-06-01 Hangzhou Innovation Institute, Beihang University A scene adaptive target detection method based on motion foreground
CN117036918A (en) * 2023-08-09 2023-11-10 北京航空航天大学 Infrared target detection method based on domain adaptation

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113591773A (en) * 2021-08-10 2021-11-02 武汉中电智慧科技有限公司 Power distribution room object detection method, device and equipment based on convolutional neural network
CN113591773B (en) * 2021-08-10 2024-03-08 武汉中电智慧科技有限公司 Distribution room object detection method, device and equipment based on convolutional neural network
CN113780135A (en) * 2021-08-31 2021-12-10 中国科学技术大学先进技术研究院 Cross-scene VOCs gas leakage detection method and system and storage medium
CN113780135B (en) * 2021-08-31 2023-08-04 中国科学技术大学先进技术研究院 Cross-scene VOCs gas leakage detection method, system and storage medium
CN113807420A (en) * 2021-09-06 2021-12-17 湖南大学 Domain self-adaptive target detection method and system considering category semantic matching
CN113807420B (en) * 2021-09-06 2024-03-19 湖南大学 Domain self-adaptive target detection method and system considering category semantic matching
CN113554013A (en) * 2021-09-22 2021-10-26 华南理工大学 Cross-scene recognition model training method, cross-scene road recognition method and device
WO2023092582A1 (en) * 2021-11-25 2023-06-01 Hangzhou Innovation Institute, Beihang University A scene adaptive target detection method based on motion foreground
CN114693983A (en) * 2022-05-30 2022-07-01 中国科学技术大学 Training method and cross-domain target detection method based on image-instance alignment network
CN117036918A (en) * 2023-08-09 2023-11-10 北京航空航天大学 Infrared target detection method based on domain adaptation
CN117036918B (en) * 2023-08-09 2024-01-30 北京航空航天大学 Infrared target detection method based on domain adaptation

Similar Documents

Publication Publication Date Title
CN113158943A (en) Cross-domain infrared target detection method
CN110210463B (en) Precise ROI-fast R-CNN-based radar target image detection method
Yin et al. Hot region selection based on selective search and modified fuzzy C-means in remote sensing images
CN111899172A (en) Vehicle target detection method oriented to remote sensing application scene
CN111611874B (en) Face mask wearing detection method based on ResNet and Canny
CN111931684A (en) Weak and small target detection method based on video satellite data identification features
CN109543632A (en) A kind of deep layer network pedestrian detection method based on the guidance of shallow-layer Fusion Features
Liu et al. Deep multi-level fusion network for multi-source image pixel-wise classification
Chen et al. Corse-to-fine road extraction based on local Dirichlet mixture models and multiscale-high-order deep learning
CN109635726B (en) Landslide identification method based on combination of symmetric deep network and multi-scale pooling
CN109492700A (en) A kind of Target under Complicated Background recognition methods based on multidimensional information fusion
CN113762009B (en) Crowd counting method based on multi-scale feature fusion and double-attention mechanism
Sahu et al. A novel parameter adaptive dual channel MSPCNN based single image dehazing for intelligent transportation systems
Ren et al. Infrared small target detection via region super resolution generative adversarial network
Li et al. An aerial image segmentation approach based on enhanced multi-scale convolutional neural network
CN112507893A (en) Distributed unsupervised pedestrian re-identification method based on edge calculation
Zhou et al. A pedestrian extraction algorithm based on single infrared image
Li et al. Real-time tracking algorithm for aerial vehicles using improved convolutional neural network and transfer learning
Sun et al. IRDCLNet: Instance segmentation of ship images based on interference reduction and dynamic contour learning in foggy scenes
Xing et al. The Improved Framework for Traffic Sign Recognition Using Guided Image Filtering
Li et al. DSPCANet: Dual-channel scale-aware segmentation network with position and channel attentions for high-resolution aerial images
CN117115555A (en) Semi-supervised three-dimensional target detection method based on noise data
Kajabad et al. YOLOv4 for urban object detection: Case of electronic inventory in St. Petersburg
CN111950476A (en) Deep learning-based automatic river channel ship identification method in complex environment
Li et al. Monocular 3-D Object Detection Based on Depth-Guided Local Convolution for Smart Payment in D2D Systems

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication

Application publication date: 20210723

WW01 Invention patent application withdrawn after publication