CN111768415A - Image instance segmentation method without quantization pooling - Google Patents

Image instance segmentation method without quantization pooling Download PDF

Info

Publication number
CN111768415A
CN111768415A CN202010542619.0A CN202010542619A CN111768415A CN 111768415 A CN111768415 A CN 111768415A CN 202010542619 A CN202010542619 A CN 202010542619A CN 111768415 A CN111768415 A CN 111768415A
Authority
CN
China
Prior art keywords
pooling
feature
quantization
image
instance segmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010542619.0A
Other languages
Chinese (zh)
Inventor
苏丽
孙雨鑫
苑守正
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Engineering University
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering University filed Critical Harbin Engineering University
Priority to CN202010542619.0A priority Critical patent/CN111768415A/en
Publication of CN111768415A publication Critical patent/CN111768415A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20016Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

The invention provides a quantification-free pooling image instance segmentation method, which comprises the following steps: s1: inputting any two-dimensional image into a deep layer feature extraction network to obtain a multilayer feature map, and extracting a candidate region through a region recommendation network; s2, pooling the feature maps with different size candidate regions to the same size by using an unquantized pooling layer; and S3, inputting two detection branches to predict the category and position of each candidate area, and simultaneously connecting mask branches in parallel to perform foreground and background mask segmentation on each candidate area to restore the size of the original image. The invention solves the problem of pixel space information loss when pooling the feature maps of the candidate regions with different sizes in the prior art, uses the non-quantization pooling layer, and enables the feature image pixels to correspond to the original image pixels one by one under the condition of not introducing any parameter, thereby ensuring the accuracy of the target position and further improving the accuracy of image instance segmentation.

Description

Image instance segmentation method without quantization pooling
Technical Field
The invention relates to an image instance segmentation method, in particular to an image instance segmentation method without quantization pooling, and belongs to the field of computer vision two-dimensional image processing.
Background
In recent years, the development of the field of deep learning rapidly and directly pushes the field of image processing in computer vision to enter a new technical era, image instance segmentation is taken as a basic research problem, and the technology can be widely applied to the fields of automobile automatic driving, robot control, video monitoring and the like.
Image instance segmentation is considered a combination of object detection and semantic segmentation because it requires the correct separation of all instances in the image, while semantic segmentation at the pixel level assigns different class labels to each instance. Such as: all dogs are present in a picture, and different dogs are also specifically indicated, and are labeled as dog 1 and dog 2.
The image example segmentation method mainly comprises two stages and one stage, wherein the Mask R-CNN is the classic work of the two-stage method, is used as a baseline algorithm of many derivative applications, inherits the idea of two-stage target detection Faster R-CNN, and expands a full convolution network to predict target Mask information. Example segmentation is challenging in that pixel spatial position information is well utilized in processing images to distinguish objects in the same class of different examples in the graph. Different candidate region positions generated by a region recommendation network in Mask R-CNN are obtained by model regression, usually floating point numbers, when pooling is carried out, a bilinear interpolation method is used for obtaining pixel values of the floating point position, fixed 4 sampling points are required to be selected for solving the pixel value of a floating point coordinate, namely four integer pixel points of the upper, lower, left and right sides around the floating point coordinate, value fixing has no adaptability to different images, each pixel point in a feature map corresponding to a candidate region is not fully considered, the problem of quantization rounding exists indirectly, pixel space information is lost, pixel deviation exists to enable a target position to be inaccurate, a target Mask cannot be accurately segmented according to the position, and further the segmentation accuracy of a subsequent example is influenced.
Similar patents are not found through the search of the existing method. Therefore, aiming at the problem of pixel space information loss in the pooling process in the prior art, the image instance segmentation without quantization pooling is provided, which has a certain meaning for improving the segmentation accuracy.
Disclosure of Invention
The invention aims to provide a quantification-free pooling image instance segmentation method which is characterized in that complete pixel space information is reserved when candidate region feature maps with different sizes are pooled, the accuracy of target positions is ensured, instance segmentation accuracy is further improved, and particularly small image targets are obtained.
The purpose of the invention is realized as follows:
an image instance segmentation method without quantization pooling comprises the following steps:
s1: inputting any two-dimensional image into a deep layer feature extraction network to obtain a multilayer feature map, and extracting a candidate region through a region recommendation network; constructing a deep feature extraction network, fusing different feature layers by using a residual convolution network and a feature pyramid, and fully utilizing shallow position information and deep semantic information of the image; extracting a candidate region and distinguishing a foreground and a background by the regional recommendation network;
s2, pooling the feature maps with different size candidate regions to the same size by using an unquantized pooling layer;
s3, inputting two detection branches to predict the category and position of each candidate area, and simultaneously parallelly connecting mask branches to perform foreground and background mask segmentation on each candidate area to restore the size of the original image;
the invention also includes such features:
the construction of the deep feature extraction network is to use feature maps activated in 4 stages in 101 layers of residual convolutional networks to be fused with 4 layers of feature pyramids, and input two-dimensional images with any size into the deep feature extraction network to obtain 4 layers of feature maps with the same channel and different sizes;
the regional recommendation network is a convolution layer with a convolution kernel of 3 x 3 and a channel number of 256;
the foreground is a position with a target, and the background is a position without a target;
step S2 is to interpolate the candidate region discrete feature map to a continuous space using the quantization-free pooling layer, and directly calculate a double integral on the continuous feature map and divide the double integral by the area to obtain a mean value;
the two detection branches are multilayer full-connection classification and regression networks;
the mask branches into a full convolution network comprising 4 layers of convolution and 1 layer of deconvolution.
Compared with the prior art, the invention has the beneficial effects that:
1. the invention solves the problem of pixel space information loss when the feature maps of the candidate areas with different sizes are pooled, and better ensures the position accuracy of the target; by using the quantization-free pooling layer, the characteristic image pixels and the original image pixels are in one-to-one correspondence under the condition of not introducing any parameter, so that the accuracy of the target position is ensured, and the accuracy of image instance segmentation is further improved;
2. the method comprehensively improves the accuracy of two-dimensional image instance segmentation in a common scene, particularly the improvement of small targets.
Drawings
FIG. 1 is a flow chart of an example segmentation method for an image without quantization pooling of the present invention;
FIG. 2 is a schematic diagram of a deep feature extraction network of an image segmentation method without quantization pooling of the present invention;
FIG. 3 is a ROIAlign diagram of an example segmentation method for an image without quantization pooling of the present invention;
FIG. 4 is a diagram illustrating quantization-free pooling of an example segmentation method for an image according to the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and specific embodiments.
The invention comprises the following steps:
and S1, inputting the two-dimensional image with any size into a deep layer feature extraction network to obtain a multilayer feature map, and extracting a candidate region through a region recommendation network.
And constructing a deep feature extraction network, fusing different feature layers by using a residual convolution network and a feature pyramid, and fully utilizing shallow position information and deep semantic information of the image. Extracting candidate regions by the regional recommendation network, and carrying out secondary classification (without distinguishing specific belonged categories) of foreground and background and fine adjustment of positions of the candidate regions by classification and regression branches;
and S2, pooling the candidate region feature maps with different sizes to the same size by using an unquantized pooling layer. Compared with the original Mask R-CNN pooling method, the method has the advantages that the double integral based on the continuous characteristic diagram is directly calculated by the non-quantization pooling layer, then the average value is obtained, and the pixel space information is fully reserved. It does not require the number of sampling points to be predefined, avoids any quantization of the coordinates, and has a continuous gradient over the candidate region coordinates.
And mapping the candidate area from the original image to the feature map according to the coordinate position, and mapping the discrete feature map to a continuous space by using an interpolation method. And performing double integration on the pixels of the continuous feature map to obtain an average value as a pooled pixel value.
Figure BDA0002539436770000031
And S3, inputting two detection branches to predict the category and the position of each candidate area, and simultaneously connecting the parallel mask branches with a full convolution network to perform foreground and background mask segmentation on each candidate area.
The method uses 135000 common scene pictures and corresponding real annotation files in an open source COCO data set as a training set, and 5000 pictures and corresponding real annotation files as a verification set and COCO segmentation indexes to evaluate the example segmentation accuracy.
Referring to the first embodiment, an embodiment of the image example segmentation method without quantization pooling in this patent is described, which specifically includes
S1: inputting any two-dimensional image into a deep layer feature extraction network to obtain a multilayer feature map, and extracting a candidate region through a region recommendation network.
S11: the two-dimensional image may be a three-channel RGB color image.
S12: construction of deep layerAnd (4) characterizing and extracting a network, referring to the second step, fusing different feature layers of high and low layers by using a 101-layer residual convolutional network and a 4-layer feature pyramid, and fully utilizing shallow position information and deep semantic information of the image. Residual convolutional network 4-stage activation feature map is { C2,C3,C4,C5(wherein C)1Large-layer parameter amount occupied memory is not selected) is fused into the feature pyramid, and then the feature graph is { P }2,P3,P4,P5And the number of channels of all the characteristic maps is 256. In particular as C5Dimension reduction to P by 1 x 1 convolution operation5To P52 times up-sampling and C after 1 x 1 convolution operation4Performing addition operation to generate P4Other feature layers operate the same.
S13: and refining the characteristic through the regional recommendation network to generate a candidate region. The region recommendation network is a common method for generating candidate regions by preprocessing images in a target detection task, and is usually connected with convolution layers with convolution kernels of 3 x 3 and channel number of 256 after a feature extraction network. And connecting two 1-by-1 convolutions in parallel to form classification and regression branches, and performing two classifications (without distinguishing specific classes) of the foreground and background of the candidate region and fine adjustment of the position of the first candidate region.
S2: referring to fig. three, in Mask RCNN, in order to pool the feature maps corresponding to the candidate regions to the same scale, and send the feature maps into a subsequent full-connected classification and regression network, a roiign algorithm is used, and the brief method includes:
s21: and (3) obtaining a pixel point corresponding to the image pixel value with the floating point coordinate by using a bilinear interpolation method to supplement a missing coordinate point on the boundary of the candidate region, so that the pooling process of the candidate region is converted into a continuous operation, and the process can refer to the second step.
The bilinear interpolation method is related to, four surrounding pixel points are fixedly sampled for solving the pixel of one point, the value is fixed, the adaptability is not available, the pixel of each point in a characteristic diagram corresponding to a candidate region is not fully considered, the space position information of a target is lost due to the coordinate quantization problem, and the positioning and the segmentation of the subsequent target are seriously influenced.
S22: the non-quantization pooling layer firstly calculates double integral of corresponding characteristic regions and then takes an average value, thereby avoiding any quantization of coordinates, and having continuous gradual change on the boundary frame coordinates, so that the characteristic image pixels correspond to the original image pixels one by one, and the accuracy of the target position is ensured. With reference to fig. four, the specific algorithm includes:
firstly, mapping a candidate area from an original image to a feature map according to coordinate positions, cutting the feature map according to the area, namely a discrete feature map, carrying out interpolation processing on the feature map f of the candidate area in order to retain more detailed information, and mapping the discrete feature map to a continuous space to obtain a continuous feature map
Figure BDA0002539436770000043
The formula is as follows:
Figure BDA0002539436770000041
C(x,y,i,j)=max(0,1-|x-i|)*max(0,1-|y-j|)
where (i, j) is the position coordinate on the discrete feature map, f (i, j) represents the pixel value of the corresponding coordinate, and (x, y) is the position coordinate on the continuous feature map,
Figure BDA0002539436770000044
and C (x, y, i, j) is an interpolation coefficient.
Assuming that the feature map size corresponding to the candidate region is (W, H), and the dimensionless pooled size is S × S, one pixel value needs to be selected from each cell g ═ W/S, H/S, to form a pixel map with S × S. If get (x)1,y1,x2,y2) A certain cell coordinate for the candidate region, wherein (x)1,y1) Is the coordinate of the upper left corner of the small grid, (x)2,y2) The coordinates of the upper right corner of the small grid are (x, y) are continuous coordinates after interpolation, and in order to fully utilize all pixel information on the characteristic diagram, the unquantized pooling layer performs double integration on the pixel value of each small grid g and then divides the pixel value by the area of the small grid to obtain an average valueBy the pixel value representing this cell, all cells are finally combined to obtain a continuous feature map, in the same way as above. The formula is as follows:
Figure BDA0002539436770000042
analytical formula (I)
Figure BDA0002539436770000045
Are continuously differentiable, so that the candidate regions can participate in the back propagation of the neural network, where x is the pair1The partial derivative of (c) is calculated as follows:
Figure BDA0002539436770000051
Figure BDA0002539436770000053
for continuous functions, other coordinate partial derivatives can be calculated in the same way.
Through the improvement of the method, the continuity of the candidate area coordinates can be realized under the condition of introducing any parameter, the pixel space information is better reserved by considering all the surrounding pixels, the characteristic image pixels correspond to the pixels of the original image one by one, and the accuracy of the target position is ensured, particularly the small target in the image (for example, the deviation of 0.5 pixel points is negligible for the larger target, but the influence of the error is much higher for the small target).
S3: after quantization-free pooling, the candidate region feature map with the same size is input into two classifier branches, namely two multilayer fully-connected classification networks to predict the category and the position of each candidate region, the mask branch and the candidate region share the feature layer, the feature layer passes through the quantization-free pooling layer, and then four layers of fully-convolution networks are connected to realize foreground and background mask segmentation on each candidate region with a possible target, and finally the category type, the position coordinate and the segmentation mask of each example in the image are obtained.
According to the method, a comparison experiment is performed on the open source COCO data set, so that the image instance segmentation accuracy can be effectively improved. The training set is 135000 common scene pictures, and the verification set is 5000 pictures.
Accuracy evaluation used the common evaluation index AP (average of all IOU thresholds) of COCO data set50,AP75,APs,APM,APLThe results of the experiments performed on the validation set are as follows.
Figure BDA0002539436770000052
The example segmentation accuracy of the invention is higher than MaskRCNN under the same setting, especially in the image small target.
In summary, the following steps: the patent discloses a quantization-pooling-free image instance segmentation method which is mainly used for the problem of target instance segmentation of two-dimensional images in common scenes. The method mainly comprises the following steps: (1) inputting two-dimensional images with any size into a deep layer feature extraction network to obtain a multilayer feature map, extracting a candidate region through a region recommendation network, (2) pooling the feature map of the candidate region by using a non-quantization pooling layer, (3) inputting two detection branches to predict the category and the position of each candidate region, and simultaneously connecting mask branches in parallel to perform foreground and background mask segmentation on each candidate region. The invention solves the problem of pixel space information loss when pooling the feature maps of the candidate regions with different sizes in the prior art, uses the non-quantization pooling layer, and enables the feature image pixels to correspond to the original image pixels one by one under the condition of not introducing any parameter, thereby ensuring the accuracy of the target position and further improving the accuracy of image instance segmentation.

Claims (7)

1. A quantification-free pooling image instance segmentation method is characterized by comprising the following steps:
s1: inputting any two-dimensional image into a deep layer feature extraction network to obtain a multilayer feature map, and extracting a candidate region through a region recommendation network; constructing a deep feature extraction network, fusing different feature layers by using a residual convolution network and a feature pyramid, and fully utilizing shallow position information and deep semantic information of the image; extracting a candidate region and distinguishing a foreground and a background by the regional recommendation network;
s2, pooling the feature maps with different size candidate regions to the same size by using an unquantized pooling layer;
and S3, inputting two detection branches to predict the category and position of each candidate area, and simultaneously connecting mask branches in parallel to perform foreground and background mask segmentation on each candidate area to restore the size of the original image.
2. The quantization-free pooling image instance segmentation method as claimed in claim 1, wherein the constructing of the deep feature extraction network is implemented by fusing 4 stages of activated feature maps in 101 layers of residual convolutional networks with 4 layers of feature pyramids, and inputting two-dimensional images with any size into the deep feature extraction network to obtain 4 layers of feature maps with same channels and different sizes.
3. The quantization-free pooled image instance segmentation method of claim 1 wherein the region recommendation network is a convolution layer with a convolution kernel of 3 x 3 and a channel number of 256.
4. The quantization-free pooled image instance segmentation method of claim 1 wherein the foreground is a location with an object and the background is a location without an object.
5. The method for image instance segmentation without quantization pooling of claim 1, wherein said step S2 is to use the quantization-free pooling layer to interpolate the candidate region discrete feature map into a continuous space, and directly calculate the double integral over the continuous feature map and divide the double integral by the area to obtain the mean value.
6. The quantization-free pooled image instance segmentation method of claim 1 wherein the two detection branches are multi-tiered fully-connected classification and regression networks.
7. The quantization-free pooled image instance segmentation method of claim 1 wherein the mask branches are full convolution networks comprising 4 layers of convolution and 1 layer of deconvolution.
CN202010542619.0A 2020-06-15 2020-06-15 Image instance segmentation method without quantization pooling Pending CN111768415A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010542619.0A CN111768415A (en) 2020-06-15 2020-06-15 Image instance segmentation method without quantization pooling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010542619.0A CN111768415A (en) 2020-06-15 2020-06-15 Image instance segmentation method without quantization pooling

Publications (1)

Publication Number Publication Date
CN111768415A true CN111768415A (en) 2020-10-13

Family

ID=72721737

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010542619.0A Pending CN111768415A (en) 2020-06-15 2020-06-15 Image instance segmentation method without quantization pooling

Country Status (1)

Country Link
CN (1) CN111768415A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112381835A (en) * 2020-10-29 2021-02-19 中国农业大学 Crop leaf segmentation method and device based on convolutional neural network
CN112613483A (en) * 2021-01-05 2021-04-06 中国科学技术大学 Outdoor fire early warning method based on semantic segmentation and recognition
CN113011332A (en) * 2021-03-19 2021-06-22 中国科学技术大学 Face counterfeiting detection method based on multi-region attention mechanism
CN113269792A (en) * 2021-05-07 2021-08-17 上海交通大学 Image post-harmony processing method, system and terminal
CN113421269A (en) * 2021-06-09 2021-09-21 南京瑞易智能科技有限公司 Real-time semantic segmentation method based on double-branch deep convolutional neural network
CN114066788A (en) * 2021-10-26 2022-02-18 华南理工大学 Balanced instance segmentation data synthesis method

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108734123A (en) * 2018-05-18 2018-11-02 武昌理工学院 Highway signs recognition methods, electronic equipment, storage medium and system
CN110059589A (en) * 2019-03-21 2019-07-26 昆山杜克大学 The dividing method of iris region in a kind of iris image based on Mask R-CNN neural network
CN110276765A (en) * 2019-06-21 2019-09-24 北京交通大学 Image panorama dividing method based on multi-task learning deep neural network
CN110659664A (en) * 2019-08-02 2020-01-07 杭州电子科技大学 SSD-based method for high-precision identification of small objects
KR102097120B1 (en) * 2018-12-31 2020-04-09 주식회사 애자일소다 System and method for automatically determining the degree of breakdown by vehicle section based on deep running
US20200356718A1 (en) * 2019-05-10 2020-11-12 Sandisk Technologies Llc Implementation of deep neural networks for testing and quality control in the production of memory devices

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108734123A (en) * 2018-05-18 2018-11-02 武昌理工学院 Highway signs recognition methods, electronic equipment, storage medium and system
KR102097120B1 (en) * 2018-12-31 2020-04-09 주식회사 애자일소다 System and method for automatically determining the degree of breakdown by vehicle section based on deep running
CN110059589A (en) * 2019-03-21 2019-07-26 昆山杜克大学 The dividing method of iris region in a kind of iris image based on Mask R-CNN neural network
US20200356718A1 (en) * 2019-05-10 2020-11-12 Sandisk Technologies Llc Implementation of deep neural networks for testing and quality control in the production of memory devices
CN110276765A (en) * 2019-06-21 2019-09-24 北京交通大学 Image panorama dividing method based on multi-task learning deep neural network
CN110659664A (en) * 2019-08-02 2020-01-07 杭州电子科技大学 SSD-based method for high-precision identification of small objects

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SHU LIU: "Path Aggregation Network for Instance Segmentation", 《IEEE》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112381835A (en) * 2020-10-29 2021-02-19 中国农业大学 Crop leaf segmentation method and device based on convolutional neural network
CN112613483A (en) * 2021-01-05 2021-04-06 中国科学技术大学 Outdoor fire early warning method based on semantic segmentation and recognition
CN113011332A (en) * 2021-03-19 2021-06-22 中国科学技术大学 Face counterfeiting detection method based on multi-region attention mechanism
CN113269792A (en) * 2021-05-07 2021-08-17 上海交通大学 Image post-harmony processing method, system and terminal
CN113269792B (en) * 2021-05-07 2023-07-21 上海交通大学 Image later-stage harmony processing method, system and terminal
CN113421269A (en) * 2021-06-09 2021-09-21 南京瑞易智能科技有限公司 Real-time semantic segmentation method based on double-branch deep convolutional neural network
CN114066788A (en) * 2021-10-26 2022-02-18 华南理工大学 Balanced instance segmentation data synthesis method
CN114066788B (en) * 2021-10-26 2024-03-29 华南理工大学 Balanced instance segmentation data synthesis method

Similar Documents

Publication Publication Date Title
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN109859190B (en) Target area detection method based on deep learning
CN109344701B (en) Kinect-based dynamic gesture recognition method
CN108961235B (en) Defective insulator identification method based on YOLOv3 network and particle filter algorithm
CN111768415A (en) Image instance segmentation method without quantization pooling
CN109583483B (en) Target detection method and system based on convolutional neural network
CN114202672A (en) Small target detection method based on attention mechanism
CN112396607A (en) Streetscape image semantic segmentation method for deformable convolution fusion enhancement
CN110659664B (en) SSD-based high-precision small object identification method
CN110176024B (en) Method, device, equipment and storage medium for detecting target in video
CN113609896A (en) Object-level remote sensing change detection method and system based on dual-correlation attention
CN110969171A (en) Image classification model, method and application based on improved convolutional neural network
CN112651423A (en) Intelligent vision system
CN111401380A (en) RGB-D image semantic segmentation method based on depth feature enhancement and edge optimization
CN113159043A (en) Feature point matching method and system based on semantic information
CN108345835B (en) Target identification method based on compound eye imitation perception
CN114140623A (en) Image feature point extraction method and system
CN112861970A (en) Fine-grained image classification method based on feature fusion
CN114519819B (en) Remote sensing image target detection method based on global context awareness
CN114299101A (en) Method, apparatus, device, medium, and program product for acquiring target region of image
CN111368637B (en) Transfer robot target identification method based on multi-mask convolutional neural network
CN109583584B (en) Method and system for enabling CNN with full connection layer to accept indefinite shape input
CN116596966A (en) Segmentation and tracking method based on attention and feature fusion
CN116091823A (en) Single-feature anchor-frame-free target detection method based on fast grouping residual error module
CN112699898B (en) Image direction identification method based on multi-layer feature fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20201013