CN113077484A - Image instance segmentation method - Google Patents

Image instance segmentation method Download PDF

Info

Publication number
CN113077484A
CN113077484A CN202110342899.5A CN202110342899A CN113077484A CN 113077484 A CN113077484 A CN 113077484A CN 202110342899 A CN202110342899 A CN 202110342899A CN 113077484 A CN113077484 A CN 113077484A
Authority
CN
China
Prior art keywords
contour
deformation
point
points
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110342899.5A
Other languages
Chinese (zh)
Other versions
CN113077484B (en
Inventor
张永生
吕可枫
于英
汪汉云
宋亮
李力
李磊
闵杰
张磊
王自全
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Information Engineering University of PLA Strategic Support Force
Original Assignee
Information Engineering University of PLA Strategic Support Force
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Information Engineering University of PLA Strategic Support Force filed Critical Information Engineering University of PLA Strategic Support Force
Priority to CN202110342899.5A priority Critical patent/CN113077484B/en
Publication of CN113077484A publication Critical patent/CN113077484A/en
Application granted granted Critical
Publication of CN113077484B publication Critical patent/CN113077484B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/44Morphing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the technical field of image instance segmentation, and particularly relates to an image instance segmentation method. Firstly, inputting a target image into a constructed initial contour extraction model to obtain an initial contour of a target in the image; then sampling a plurality of contour points on the initial contour, and extracting a plurality of assumed deformation points around each contour point; then, by utilizing the constructed feature extraction model, the features of all contour points and the features of assumed deformation points around each contour point are obtained; and finally, inputting the obtained characteristics of each contour point and the characteristics of the assumed deformation points around each contour point into the constructed deformation contour extraction model to obtain the deformation result of each contour point, and further obtain the deformation contour of the target in the image. The method can obtain more precise segmentation results, so that the segmentation results are more precise, the edge contour precision is improved, and the method is suitable for tasks with higher requirements on precision.

Description

Image instance segmentation method
Technical Field
The invention belongs to the technical field of image instance segmentation, and particularly relates to an image instance segmentation method.
Background
Segmentation tasks are the basis of many image processing tasks, such as change detection, dynamic monitoring, classification tasks, etc., and the accuracy and efficiency of segmentation can affect subsequent work. Whereas example segmentation integrates three tasks, classification, localization and segmentation, which is one of the most difficult visual tasks compared to similar image processing tasks. From a procedural point of view, instance segmentation can be divided into two categories, detection-based methods and contour-based methods.
Firstly detecting an object based on a detection method, and then classifying the object pixel by a certain method to obtain a final example segmentation result. However, such pixel-based methods do not utilize edge information, which results in poor segmentation at the edges.
Another method is to achieve segmentation by obtaining the contour of the object. The contour-based method needs to solve the problem of how to measure the distance between an approximate contour formed by a point set and an actual contour, and accordingly, a reasonable learning network loss function is provided. The problems of initial position sampling of the point set, feature extraction of the points and the like also need to be solved. Quantitatively, the contour-based method can reduce the amount of determination required compared with the region-by-region pixel segmentation method, and a smaller detection amount means a smaller error rate, so the example contour-based segmentation method is also very worthy of study. Compared with the pixel-by-pixel classification method, the contour method has higher efficiency and speed, but has higher requirements on a training network and an initial contour.
For some specific image processing tasks (e.g., change detection using remote sensing images), higher image resolution and larger image range both place higher precision requirements on the segmentation effect. The existing example segmentation method has a good overall example segmentation effect on a public data set, but the precision of the existing example segmentation method on an object outline is not enough, and the existing example segmentation method is not applicable to some tasks with high requirements on fineness and is not enough to be used as a fine segmentation result to guide subsequent works such as change detection, example positioning and the like.
Disclosure of Invention
The invention provides an image instance segmentation method, which is used for solving the problem that the contour extracted by adopting the prior art is low in precision.
In order to solve the technical problems, the technical scheme and the corresponding beneficial effects of the technical scheme are as follows:
the invention provides an image instance segmentation method, which comprises the following steps:
1) inputting a target image into the constructed initial contour extraction model to obtain an initial contour of a target in the image; and samples a number of contour points on the initial contour,
2) extracting a plurality of assumed deformation points around each contour point;
3) obtaining the characteristics of each contour point and the characteristics of assumed deformation points around each contour point by using the constructed characteristic extraction model;
4) and inputting the obtained features of the contour points and the features of the assumed deformation points around each contour point into the constructed deformation contour extraction model to obtain the deformation result of each contour point, and further obtain the deformation contour of the target in the image so as to perform optimized segmentation edge processing on the initial contour.
The beneficial effects of the above technical scheme are: after the initial contour of the target is obtained, a plurality of assumed deformation points are extracted around the contour points, the characteristics of each contour point and the assumed deformation point network around each contour point are determined and input into the constructed deformed contour extraction model, and the deformation result of each contour point can be obtained.
Further, in order to adaptively select the scale of the assumed deformation point network to achieve the effect of better adapting to the initial contour, a contour point and a plurality of assumed deformation points extracted around the contour point form an assumed deformation point network, and the size of the assumed deformation point network is as follows:
S=[S0+log2(wh/WH)]
wherein, the [ alpha ], [ beta ]]Representing rounding processing; s is the size of the assumed deformation point network; s0The hierarchy where the object with the area of W × H is located; w and h are the width and height of the detection frame respectively; w and H are the width and height of the target image, respectively.
Further, the feature extraction model is a ResNet-FPN model.
Further, the means for obtaining the features of the contour points and the features of the assumed deformation points around each contour point by using the ResNet-FPN model is as follows: firstly, inputting a target image into a ResNet-FPN model to obtain a plurality of characteristic graphs; and obtaining the characteristics of each contour point and each assumed deformation point from each characteristic diagram by utilizing an interpolation method.
Further, in order to adaptively select feature maps to adapt to target example segmentation tasks with different sizes, the number of the obtained feature maps PX is 3 and 3 adjacent feature maps PX are respectively a first feature map P (K-1), a second feature map PK and a third feature map P (K + 1); wherein, the smaller the value of X, the lower the convolution layer is used for processing to obtain the corresponding characteristic diagram; the value of K is:
K=[K0+log2(wh/WH)]
wherein, the [ alpha ], [ beta ]]Representing rounding processing; w and h are the width and height of the detection frame respectively; w and H are mesh respectivelyWidth and height of the target image; k0The level of the object with the area of W × H.
Further, in the step 1), the initial contour extraction model is a Mask R-CNN model.
Further, in step 4), the deformation contour extraction model is a graph convolution neural network model.
Further, when the constructed initial contour extraction model, the constructed feature extraction model and the constructed deformed contour extraction model are trained, the loss function L is as follows:
Figure BDA0003000050940000031
Figure BDA0003000050940000032
Figure BDA0003000050940000033
wherein, SmoothL1Express SmoothL1A loss function; l ispoi_nA loss function for each contour point is represented,
Figure BDA0003000050940000034
in order to predict the coordinates of the point,
Figure BDA0003000050940000035
coordinates of real points; l iscon_nRepresenting the loss function for the entire contour, N is the total number of contour points,
Figure BDA0003000050940000036
as coordinates of predicted contour points, xiCoordinates of real contour points; n is the number of iterations; l isMask_nIs the average binary cross entropy loss of all pixels within the detection box.
Further, in order to ensure that the efficiency and the effect reach the optimal balance, the number of repeated iterations is required to be 3.
Drawings
FIG. 1 is a flow chart of an example segmentation method of an image of the present invention;
FIG. 2 is a schematic view of a hypothetical network of inflection points of different sizes of the present invention;
FIG. 3 is a schematic diagram of a graph convolution warping process of the present invention.
Detailed Description
The basic concept of the invention is as follows: aiming at the problem that the method in the prior art is not suitable for tasks with higher requirements on fineness, the invention carries out deformation processing on the outline on the basis of the segmentation result obtained by Mask R-CNN to obtain a more accurate segmentation result. The specific deformation treatment process comprises the following steps: firstly, representing a contour by using a sufficient number of contour points, and designing N assumed deformation points for each contour point, wherein the N assumed deformation points and the corresponding contour points form a plane network; then, after the characteristics of the points are obtained, the characteristics are input into a graph convolution neural network (GCN) as a local graph to obtain the weight of each point; and then, carrying out weighted summation operation on the weights of all the points to obtain a deformation result of each contour point, thereby obtaining a final deformation contour.
The following describes the image example segmentation method of the present invention in detail with reference to the flowchart of fig. 1.
Step one, inputting a target image (Images) into a constructed Initial Contour extraction model (Mask R-CNN model) to obtain an Initial Contour (Initial Contour) of a target in the image.
And step two, sampling a plurality of contour points enough to represent the contour on the initial contour, and extracting N assumed deformation points around each contour point.
The graph refers to a topological graph in mathematics, wherein corresponding relations are established by using vertexes and edges, and the essential purpose of the GCN is to extract spatial features of the topological graph and use the spatial features for learning. In this step, the present invention designs a hypothetical deformed point network for each point sampled from the contour, and treats the hypothetical deformed point network as a graph. When a hypothetical deformation point network is selected, the number of vector directions needs to meet the deformation requirement on the premise of ensuring the local point density.
As shown in fig. 2, it is assumed that a deformed Point network (Local Point-Net) is composed of 21 points, and the points form a network, and together form 48 edges, and have a certain topological relationship. The assumed deformation point network has 16 different vector directions, ensures the omnibearing deformation requirement, has layering sense and can meet different deformation sizes. It is assumed that a network of deformed points is used to predict the direction and magnitude of the deformation of the sampled contour points and that the relative orientation of each point involved is determined. In this respect, the present invention designs an adaptive scaling strategy to adapt to contour deformation of objects of different sizes, assuming that the selected dimensions of the deformed point network are:
S=[S0+log2(wh/WH)] (1)
wherein, the [ alpha ], [ beta ]]Representing rounding processing; s is the size of the assumed deformation point network; s0The hierarchy level where the object with the area of W × H is located is set to 20 pixels in this embodiment; w and h are the width and height of the detection frame respectively; w and H are the width and height of the target image, respectively.
And step three, after the contour points and the assumed deformation points of each contour point are obtained, the features of each contour point and the features of the assumed deformation points around each contour point are obtained by utilizing the constructed feature extraction model (ResNet-FPN model). Specifically, the method comprises the following steps:
1. and (3) putting the target Images (Images) into a ResNet-FPN model to obtain 3 adjacent feature maps.
Mask R-CNN uses a ResNet-FPN model for feature extraction, and the model is also used for feature extraction in the embodiment. FPN (feature pyramid network) is a well-designed multi-scale detection method, which comprises three parts of bottom-up, top-down and transverse connection, and the structure can fuse the features of all levels to make the features have strong semantic information and strong spatial information simultaneously. As shown in fig. 1, in the bottom-up process, ResNet is used as a skeleton network, and is divided into 5 layers according to the size of a feature map, and the output results of each layer are ConV1, ConV2, ConV3, ConV4 and ConV 5; the top-down process is to sample the ConV feature map by using a nearest neighbor method to obtain M2, M3, M4 and M5, wherein M6 is obtained by M5 downsampling; the transverse connection process is to fuse the M layers and the ConV layers, eliminate aliasing effect of up-sampling through 3-by-3 convolution kernel processing, and finally obtain P2, P3, P4, P5 and P6. As can be seen from the figure, P2 is obtained by the fusion process of ConV2 and M3 and the 3 × 3 convolution kernel process, and P3 is obtained by the fusion process of ConV3 and M4 and the 3 × 3 convolution kernel process, that is, the smaller X in the feature map PX indicates that PX is obtained by the lower convolution layer.
Assuming that (H, W) is the height and width of the target image, the feature maps P2, P3, P4, P5 and P6 correspond to the sizes of (H/4, W/4), (H/8, W/8), (H/16, W/16), (H/32, W/32) and (H/64, W/64) in sequence, and the number of channels is 256. And taking [ P2, P3, P4 and P5] as candidate feature maps, and selecting the feature maps according to the sizes of the segmentation objects in different tasks. For example, in the task of segmenting the small target instances, a feature map with higher resolution should be selected, for example, the feature map with higher resolution is selected to improve the segmentation accuracy [ P2, P3, P4], whereas for the task of segmenting the large target instances, a feature map with relatively lower resolution may be selected to improve the segmentation efficiency [ P3, P4, P5 ]. In this embodiment, the K value calculated by the following formula is used to adaptively select a feature map [ P (K-1), PK, P (K +1) ]:
K=[K0+log2(wh/WH)] (2)
wherein, the [ alpha ], [ beta ]]Representing rounding processing; w and h are the width and height of the detection frame respectively; w and H are the width and height of the target image respectively; k0The hierarchy level where the object with the area W × H should be located is set to 5 in this embodiment.
2. Because the feature map cannot contain all points (including contour points and corresponding assumed deformation points), after the feature map is selected, the feature corresponding to each point is obtained from each feature map by using a bilinear interpolation method according to the coordinate information of the contour points and the assumed deformation points, and the feature map is connected.
In the neural network, as the network deepens, each layer loses some information, including the feature pyramid idea in the FPN, and the multi-scale multi-layer feature fusion can retain more information. Because of its simplicity and ease of operation, feature linking has been widely used as a loss reduction method. Since the number of channels of the previously selected feature maps P2-P5 is 256, the features of each point are summed, and the number of channels is kept unchanged at 256. In summary, the feature dimension of each contour point is 21 × 256 by feature extraction and feature connection.
And step four, inputting the obtained characteristics of each contour point and the characteristics of the assumed deformation points around each contour point into a constructed deformation contour extraction model (GCN model) to obtain a deformation result (Refine Mask) of each contour point, and further obtaining a deformation contour of the target in the image so as to perform optimized segmentation edge processing on the initial contour. Specifically, the method comprises the following steps:
after obtaining the features corresponding to the contour points and the assumed deformation points, it is necessary to use these features to infer the deformation information of the contour points from the respective assumed deformation points. As illustrated in fig. 3, 128 contour points are chosen for each initial mask in order to be able to represent the shape of most objects. The coordinate point dimension as input is therefore 128 × 21 × 2. After the features are obtained, they are input into the GCN to calculate the weights of the various assumed deformed points.
In this embodiment, a scoring graph neural network consisting of 5 graph convolution layers + ReLU is used to predict the weight of each hypothesis deformation point. And then obtaining the deformation vector information of the contour point through a Softmax layer and weighted summation operation, and further obtaining a deformation result. The GCN is adopted to realize the information transmission of the multi-order neighborhood by superposing a plurality of convolution layers, and the topological information of the local deformation point network is fully utilized.
In addition, considering that a good effect is difficult to achieve only by one-time deformation and the quality is poor, in the embodiment, a final deformation result is obtained by multiple times of iterative deformation in an iterative optimization manner. That is, the deformed contour of the target in the image obtained as the result of the first iteration is used as the initial contour when the second iteration is performed, the above-described steps two to four are repeated to obtain the deformed result of the second iteration, and the result obtained by the second iteration is used as the final deformed result. It should be noted that the quality here refers to the refinement degree of the segmentation, and what the method of the present invention improves is the refinement degree of the edge, which is expressed as that the segmentation result is closer to the target at the edge degree. The used quality evaluation index is a standard Mask AP (average of IoU thresholds) value, wherein AP refers to the average value of AP50, AP55, AP60, … and AP95, and the figure represents an IOU threshold value and is the evaluation index of COCO data set standard.
It should be noted that the initial contour extraction model, the feature extraction model, and the deformed contour extraction model are trained together during training. In the training process, two loss functions are selected, one is SmoothL1And the other is to inherit the Mask loss function in Mask R-CNN.
For each contour point, a loss function is defined as follows:
Figure BDA0003000050940000061
wherein n is the number of iterations;
Figure BDA0003000050940000062
in order to predict the coordinates of the point,
Figure BDA0003000050940000063
is the coordinate of the real point.
Summarizing in an iterative process, for the entire contour, a loss function is defined as follows:
Figure BDA0003000050940000064
wherein N is the total number of contour points,
Figure BDA0003000050940000065
to be driven from
Figure BDA0003000050940000066
The resulting coordinates are resampled.
L inherited from Mask R-CNNMaskIs the average binary cross-entropy loss (the average binary cross-entropy) of all pixels within the detection box.
Finally defining a multitask loss function as:
Figure BDA0003000050940000067
in this embodiment, the initial contour extraction model is Mask R-CNN model, and the feature extraction model is ResNet-FPN model. As other embodiments, other models in the prior art may be selected for performing corresponding initial contour extraction and feature extraction operations, for example, the initial contour extraction model may be a PointRend model, a Cascade Mask R-CNN model, or the like, and the feature extraction model may be a resenext model, or the like.
In this embodiment, two iterations are performed and the final deformation result is used as the deformation contour of the target in the image. As another embodiment, if one iteration has a better effect, the second iteration may not be performed to improve the calculation efficiency. Of course, if the two iterations are still poor, a third iteration may be performed, but some computational efficiency must be sacrificed.
The invention has the following characteristics:
1. according to the method, N assumed deformation points are extracted around the contour point, one contour point and the characteristics of the N assumed deformation points around the contour point are extracted and input into the constructed deformation contour extraction model, so that the deformation result of each contour point can be obtained, and further the deformation contour of the target in the image can be obtained, and the contour is deformed on the basis of the segmentation result obtained by using the initial contour extraction model, so that a more precise segmentation result can be obtained, the edge contour precision is improved, and the method is suitable for tasks with higher requirements on the precision degree.
2. The initial contour extraction model adopts a Mask R-CNN model, the feature extraction model adopts a ResNet-FPN model, and each point inherits the features of the corresponding position of the feature map in the Mask R-CNN model network architecture, so that the deformation result can be better adapted to the initial contour.
3. Because the deformation degrees between the initial contour and the real contour of the target object with different sizes are different, when the assumed deformation point network is designed, the assumed deformation point network with the self-adaptive scale is designed, specifically see formula (1), so that the size of the assumed deformation point network can be adaptively adjusted according to the size of the target image and the size of the target object, and the effect of better adapting to the initial contour is achieved.
4. When the feature map is selected, a method for automatically selecting the feature layer combination is adopted, specifically, see formula (2), so that the feature layer combination can be adaptively selected according to the size of a target image and the size of a target object, the segmentation precision is improved by adaptively selecting a shallow network for a small target instance segmentation task, and the segmentation efficiency is improved by adaptively selecting a deep network for a large target instance segmentation task.
5. An iterative method is used, and a more refined segmentation result is obtained through multiple times of deformation.

Claims (9)

1. An image instance segmentation method is characterized by comprising the following steps:
1) inputting a target image into the constructed initial contour extraction model to obtain an initial contour of a target in the image; sampling a plurality of contour points on the initial contour;
2) extracting a plurality of assumed deformation points around each contour point;
3) obtaining the characteristics of each contour point and the characteristics of assumed deformation points around each contour point by using the constructed characteristic extraction model;
4) inputting the obtained characteristics of each contour point and the characteristics of assumed deformation points around each contour point into a constructed deformation contour extraction model to obtain a deformation result of each contour point, further obtaining a deformation contour of a target in an image, judging whether the quality of the deformation contour meets requirements or not, if not, sampling a plurality of contour points on the deformation contour, and repeating the steps 2) to 4) until the quality of the finally obtained deformation contour meets the requirements or the number of repeated iterations meets the requirements.
2. The image instance segmentation method according to claim 1, wherein a contour point and a plurality of assumed deformation points extracted around the contour point form an assumed deformation point network, and the size of the assumed deformation point network is:
S=[S0+log2(wh/WH)]
wherein, the [ alpha ], [ beta ]]Representing rounding processing; s is the size of the assumed deformation point network; s0The hierarchy where the object with the area of W × H is located; w and h are the width and height of the detection frame respectively; w and H are the width and height of the target image, respectively.
3. An image instance segmentation method according to claim 1, characterized in that the feature extraction model is the ResNet-FPN model.
4. An image instance segmentation method as claimed in claim 3, characterized in that the means for deriving the features of the respective contour points and the features of the assumed deformation points around each contour point by means of the ResNet-FPN model are: firstly, inputting a target image into a ResNet-FPN model to obtain a plurality of characteristic graphs; and obtaining the characteristics of each contour point and each assumed deformation point from each characteristic diagram by utilizing an interpolation method.
5. The image example division method according to claim 4, wherein the number of the obtained feature maps PX is 3 and 3 adjacent feature maps P (K-1), PK, and P (K + 1); wherein, the smaller the value of X, the lower the convolution layer is used for processing to obtain the corresponding characteristic diagram; the value of K is:
K=[K0+log2(wh/WH)]
wherein, the [ alpha ], [ beta ]]Representing rounding processing; w and h are the width and height of the detection frame respectively; w and H are the width and height of the target image respectively; k0The level of the object with the area of W × H.
6. The image instance segmentation method according to claim 1, wherein in step 1), the initial contour extraction model is a Mask R-CNN model.
7. The image instance segmentation method according to claim 1, wherein in step 4), the deformation contour extraction model is a atlas neural network model.
8. The image instance segmentation method according to any one of claims 1 to 7, wherein a loss function L adopted in training the constructed initial contour extraction model, the constructed feature extraction model and the constructed deformed contour extraction model is as follows:
Figure FDA0003000050930000021
Figure FDA0003000050930000022
Figure FDA0003000050930000023
wherein, SmoothL1Express SmoothL1A loss function; l ispoi_nA loss function for each contour point is represented,
Figure FDA0003000050930000024
in order to predict the coordinates of the point,
Figure FDA0003000050930000025
coordinates of real points; l iscon_nRepresenting the loss function for the entire contour, N is the total number of contour points,
Figure FDA0003000050930000026
as coordinates of predicted contour points, xiCoordinates of real contour points; n is the number of iterations; l isMask_nIs the average binary cross entropy loss of all pixels within the detection box.
9. The image instance segmentation method according to claim 1, wherein the requirement for the number of repeated iterations is 3.
CN202110342899.5A 2021-03-30 2021-03-30 Image instance segmentation method Active CN113077484B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110342899.5A CN113077484B (en) 2021-03-30 2021-03-30 Image instance segmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110342899.5A CN113077484B (en) 2021-03-30 2021-03-30 Image instance segmentation method

Publications (2)

Publication Number Publication Date
CN113077484A true CN113077484A (en) 2021-07-06
CN113077484B CN113077484B (en) 2023-05-23

Family

ID=76612002

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110342899.5A Active CN113077484B (en) 2021-03-30 2021-03-30 Image instance segmentation method

Country Status (1)

Country Link
CN (1) CN113077484B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113822905A (en) * 2021-09-09 2021-12-21 辽宁工程技术大学 Improved example segmentation method based on target contour

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140270540A1 (en) * 2013-03-13 2014-09-18 Mecommerce, Inc. Determining dimension of target object in an image using reference object
CN106875365A (en) * 2017-03-01 2017-06-20 淮安信息职业技术学院 Spill image partition method based on GSA
CN110148148A (en) * 2019-03-01 2019-08-20 北京纵目安驰智能科技有限公司 A kind of training method, model and the storage medium of the lower edge detection model based on target detection
WO2019180745A1 (en) * 2018-03-21 2019-09-26 Karade Vikas Systems and methods for obtaining 3-d images from x-ray information for deformed elongate bones
CN110569857A (en) * 2019-07-28 2019-12-13 景德镇陶瓷大学 image contour corner detection method based on centroid distance calculation
CN111126389A (en) * 2019-12-20 2020-05-08 腾讯科技(深圳)有限公司 Text detection method and device, electronic equipment and storage medium
CN111340784A (en) * 2020-02-25 2020-06-26 安徽大学 Image tampering detection method based on Mask R-CNN
CN111723845A (en) * 2020-05-19 2020-09-29 浙江工业大学 Cell image segmentation method based on Mask contour
CN112001403A (en) * 2020-08-11 2020-11-27 北京化工大学 Image contour detection method and system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140270540A1 (en) * 2013-03-13 2014-09-18 Mecommerce, Inc. Determining dimension of target object in an image using reference object
CN106875365A (en) * 2017-03-01 2017-06-20 淮安信息职业技术学院 Spill image partition method based on GSA
WO2019180745A1 (en) * 2018-03-21 2019-09-26 Karade Vikas Systems and methods for obtaining 3-d images from x-ray information for deformed elongate bones
CN110148148A (en) * 2019-03-01 2019-08-20 北京纵目安驰智能科技有限公司 A kind of training method, model and the storage medium of the lower edge detection model based on target detection
CN110569857A (en) * 2019-07-28 2019-12-13 景德镇陶瓷大学 image contour corner detection method based on centroid distance calculation
CN111126389A (en) * 2019-12-20 2020-05-08 腾讯科技(深圳)有限公司 Text detection method and device, electronic equipment and storage medium
CN111340784A (en) * 2020-02-25 2020-06-26 安徽大学 Image tampering detection method based on Mask R-CNN
CN111723845A (en) * 2020-05-19 2020-09-29 浙江工业大学 Cell image segmentation method based on Mask contour
CN112001403A (en) * 2020-08-11 2020-11-27 北京化工大学 Image contour detection method and system

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
HUAN LING等: "Fast Interactive Object Annotation with Curve-GCN", 《ARXIV》 *
SIDA PENG等: "Deep Snake for Real-Time Instance Segmentation", 《ARXIV》 *
张继凯等: "深度学习的图像实例分割方法综述", 《小型微型计算机系统》 *
李学龙等: "像素级语义理解: 从分类到回归", 《中国科学:信息科学》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113822905A (en) * 2021-09-09 2021-12-21 辽宁工程技术大学 Improved example segmentation method based on target contour
CN113822905B (en) * 2021-09-09 2024-03-29 辽宁工程技术大学 Improved object contour-based instance segmentation method

Also Published As

Publication number Publication date
CN113077484B (en) 2023-05-23

Similar Documents

Publication Publication Date Title
CN111598174B (en) Model training method based on semi-supervised antagonistic learning and image change analysis method
Li et al. Multiscale features supported DeepLabV3+ optimization scheme for accurate water semantic segmentation
CN105701508B (en) Global local optimum model and conspicuousness detection algorithm based on multistage convolutional neural networks
Zhou et al. BOMSC-Net: Boundary optimization and multi-scale context awareness based building extraction from high-resolution remote sensing imagery
CN111310861A (en) License plate recognition and positioning method based on deep neural network
CN103049763B (en) Context-constraint-based target identification method
CN111079739B (en) Multi-scale attention feature detection method
CN112488025B (en) Double-temporal remote sensing image semantic change detection method based on multi-modal feature fusion
CN110751195B (en) Fine-grained image classification method based on improved YOLOv3
CN113177456B (en) Remote sensing target detection method based on single-stage full convolution network and multi-feature fusion
CN114627052A (en) Infrared image air leakage and liquid leakage detection method and system based on deep learning
CN111738055A (en) Multi-class text detection system and bill form detection method based on same
CN111461213A (en) Training method of target detection model and target rapid detection method
CN114782734A (en) Visual detection method for pipeline leakage of valve cooling system, computer and storage medium
CN105405138A (en) Water surface target tracking method based on saliency detection
CN111915628A (en) Single-stage instance segmentation method based on prediction target dense boundary points
Liu et al. Image edge recognition of virtual reality scene based on multi-operator dynamic weight detection
CN113920468A (en) Multi-branch pedestrian detection method based on cross-scale feature enhancement
Zuo et al. A remote sensing image semantic segmentation method by combining deformable convolution with conditional random fields
Bobyr et al. A method for creating a depth map based on a three-level fuzzy model
CN113888505B (en) Natural scene text detection method based on semantic segmentation
CN113077484A (en) Image instance segmentation method
CN111160372B (en) Large target identification method based on high-speed convolutional neural network
CN115984712A (en) Multi-scale feature-based remote sensing image small target detection method and system
CN114494999B (en) Double-branch combined target intensive prediction method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant