CN111899169A - Network segmentation method of face image based on semantic segmentation - Google Patents
Network segmentation method of face image based on semantic segmentation Download PDFInfo
- Publication number
- CN111899169A CN111899169A CN202010628571.5A CN202010628571A CN111899169A CN 111899169 A CN111899169 A CN 111899169A CN 202010628571 A CN202010628571 A CN 202010628571A CN 111899169 A CN111899169 A CN 111899169A
- Authority
- CN
- China
- Prior art keywords
- module
- convolution
- network
- model
- face image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 31
- 238000000034 method Methods 0.000 title claims abstract description 23
- 230000003993 interaction Effects 0.000 claims abstract description 19
- 238000012360 testing method Methods 0.000 claims abstract description 15
- 238000012795 verification Methods 0.000 claims abstract description 13
- 238000012549 training Methods 0.000 claims abstract description 12
- 230000004927 fusion Effects 0.000 claims abstract description 9
- 238000005070 sampling Methods 0.000 claims description 17
- 238000012805 post-processing Methods 0.000 claims description 6
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 5
- 238000004364 calculation method Methods 0.000 claims description 5
- 238000012935 Averaging Methods 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 4
- 238000010586 diagram Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000003709 image segmentation Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000010339 dilation Effects 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000004043 dyeing Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformation in the plane of the image
- G06T3/40—Scaling the whole image or part thereof
- G06T3/4007—Interpolation-based scaling, e.g. bilinear interpolation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformation in the plane of the image
- G06T3/40—Scaling the whole image or part thereof
- G06T3/4053—Super resolution, i.e. output image resolution higher than sensor resolution
- G06T3/4076—Super resolution, i.e. output image resolution higher than sensor resolution by iteratively correcting the provisional high resolution image using the original low-resolution image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
-
- G06T5/70—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/40—Analysis of texture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20132—Image cropping
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20192—Edge enhancement; Edge preservation
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Abstract
The invention discloses a network segmentation method of a face image based on semantic segmentation, which comprises the following steps: obtaining an image dataset; constructing a segmented deep convolutional network structure; using a network structure to train data to obtain a training model; carrying out verification and parameter adjustment by using the verification set, and selecting an optimal model; and testing the selected optimal model by using a test set. The invention adopts a lightweight model, combines a space channel and a context information channel, gradually increases subnets from high resolution to low resolution on the basis of the original space network structure to form more stages, and connects the subnets with multiple resolutions in parallel to obtain the information interaction module. And then, multi-scale fusion is carried out for multiple times, so that each high-resolution to low-resolution representation repeatedly receives information from other parallel representations, and abundant high-resolution representations are obtained. Because parallel connections are used, a high resolution representation can be maintained, and thus the prediction is spatially more accurate.
Description
Technical Field
The invention relates to the technical field of image processing, in particular to a network segmentation method of a face image based on semantic segmentation.
Background
Due to the development and application of convolutional neural networks (CNN for short), many tasks in the field of computer vision have been greatly developed, wherein image segmentation is a task in computer vision, and the purpose of the task is to divide and label images according to areas where different targets exist. Furthermore, semantic segmentation is to label the image at a pixel level, and label each pixel of the image with its corresponding category, because each pixel is considered, semantic segmentation is a kind of density-type prediction. There are many ideas of semantic segmentation, such as patch classification, full convolution method, encoder-decoder architecture, etc., and the currently popular is the encoder-decoder architecture, and this document also adopts this architecture to design a deep convolutional network.
However, in designing a semantic segmentation model, people excessively pursue the accuracy of the model, and a complicated backbone is introduced, which brings heavy computational burden and memory occupation. And due to the complexity of the backbone network, the model is difficult to deploy in practical application. Therefore, solving this problem is an important task in the current field of semantic segmentation, which is originally aimed at balancing the relationship between efficiency and speed of the segmentation network, and providing a simpler solution for the segmentation of multitasks.
At present, various beautifying and makeup software exist in the market. If a certain part of the face is to be processed, each part of the face must be segmented, and then different parts are beautified and made up. The invention processes the segmentation task of the face image, mainly aims at segmenting the face part from the hair, reserves the edge between the face part and the hair and properly denoises, so that the processed image is more natural and softer.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a method for segmenting a network of a face image based on semantic segmentation.
The purpose of the invention is realized by the following technical scheme:
a method for segmenting a network based on a face image segmented by semantics mainly comprises the following specific steps:
step S1: after a series of operations, a corresponding image data set for face segmentation is obtained.
Further, the step S1 further includes the following steps:
and step S11, adopting a face recognition data set laboratory Faces in the world (LFW), and dividing the training set, the verification set and the test set according to the proportion respectively.
Further, the series of operations in step S1 includes averaging, defogging, and clipping operations.
Step S2: and constructing a segmented deep convolutional network structure.
Further, the step S2 further includes the following steps:
step S21: the deep convolutional network adopts an encoder-decoder architecture as a network structure, and the network structure comprises an encoder module and a decoder module.
Step S22: the encoder module comprises three parts, namely a fast down-sampling module, an information interaction module and an expansion group.
Further, the step S22 further includes the following steps:
step S221: the fast down-sampling module consists of three convolution layers, and the adopted convolution layers have two forms, one is a standard convolution layer, and the other is a depth-separable convolution layer; the convolution layer with separable depth can effectively reduce the parameter quantity of the model, thereby reducing the calculation burden; after each convolutional layer there is a BN layer and a RELU activation function is used.
Step S222: the structure of the information interaction module adopts an Inverted residual bottleneck block (Inverted bottle residual block) of MobileNet V2, and beautiful output is obtained through information interaction and combination of feature graphs with different dimensions.
Step S223: the expansion group is used for performing spatial expansion convolution on the module subjected to feature fusion through the information interaction module, and the expansion convolution can increase the receptive field of a convolution kernel and capture more levels of context information.
Step S23: the decoder module mainly comprises a bilinear up-sampling layer and a convolution layer, wherein the convolution layer is followed by a softmax layer to classify the pixel level.
Step S24: post-processing the output of the decoder module, preserving details of face and hair edges by employing a guided filter, and reducing noise.
Step S3: using the network structure of the step S2 to train data to obtain a corresponding training model;
step S4: carrying out verification and parameter adjustment by using the verification set, and selecting an optimal model;
step S5: and testing the selected optimal model by using a test set to evaluate the performance of the model.
The working process and principle of the invention are as follows: the invention provides a method for segmenting a network based on a semantic segmentation face image, which adopts a lightweight model, adopts a combination of a space channel and a context information channel, gradually increases subnets from high resolution to low resolution on the basis of an original space network structure to form more stages, and connects the subnets with multiple resolutions in parallel to obtain the information interaction module. And then, multi-scale fusion is carried out for multiple times, so that each high-resolution to low-resolution representation repeatedly receives information from other parallel representations, and abundant high-resolution representations are obtained. Because parallel connections are used, a high resolution representation can be maintained, and thus the prediction is spatially more accurate.
Compared with the prior art, the invention also has the following advantages:
(1) the method for segmenting the network based on the semantic segmentation human face image improves the speed, does not reduce the performance of the network, and obviously improves the efficiency compared with the prior art.
(2) The method for segmenting the network based on the semantic segmentation human face image can obtain the hair area of the human face part by utilizing the image processed by the network, and then can perform corresponding dyeing operation and the like according to the requirements of a user, and is simple, convenient and quick to operate.
Drawings
FIG. 1 is a flow chart of an image segmentation method provided by the present invention.
Fig. 2 is a schematic structural diagram of the whole face image segmentation network provided by the present invention.
Fig. 3 is a schematic structural diagram of an inversion residual block of the MobileNet V2 network structure provided by the present invention.
Fig. 4 is a schematic diagram of a network structure provided by the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention clearer and clearer, the present invention is further described below with reference to the accompanying drawings and examples.
Example 1:
as shown in fig. 1 to 4, the present embodiment discloses a method for segmenting a network based on a face image of semantic segmentation, which mainly includes the following specific steps:
step S1: after a series of operations, a corresponding image data set for face segmentation is obtained.
Further, the step S1 further includes the following steps:
and step S11, adopting a face recognition data set laboratory Faces in the world (LFW), and dividing the training set, the verification set and the test set according to the proportion respectively.
Further, the series of operations in step S1 includes averaging, defogging, and clipping operations.
Step S2: and constructing a segmented deep convolutional network structure.
Further, the step S2 further includes the following steps:
step S21: the deep convolutional network adopts an encoder-decoder architecture as a network structure, and the network structure comprises an encoder module and a decoder module.
Step S22: the encoder module comprises three parts, namely a fast down-sampling module, an information interaction module and an expansion group.
Further, the step S22 further includes the following steps:
step S221: the fast down-sampling module consists of three convolution layers, and the adopted convolution layers have two forms, one is a standard convolution layer, and the other is a depth-separable convolution layer; the convolution layer with separable depth can effectively reduce the parameter quantity of the model, thereby reducing the calculation burden; after each convolutional layer there is a BN layer and a RELU activation function is used.
Step S222: the structure of the information interaction module adopts an Inverted residual bottleneck block (Inverted bottle residual block) of MobileNet V2, and beautiful output is obtained through information interaction and combination of feature graphs with different dimensions.
Step S223: the expansion group is used for performing spatial expansion convolution on the module subjected to feature fusion through the information interaction module, and the expansion convolution can increase the receptive field of a convolution kernel and capture more levels of context information.
Step S23: the decoder module mainly comprises a bilinear up-sampling layer and a convolution layer, wherein the convolution layer is followed by a softmax layer to classify the pixel level.
Step S24: post-processing the output of the decoder module, preserving details of face and hair edges by employing a guided filter, and reducing noise.
Step S3: using the network structure of the step S2 to train data to obtain a corresponding training model;
step S4: carrying out verification and parameter adjustment by using the verification set, and selecting an optimal model;
step S5: and testing the selected optimal model by using a test set to evaluate the performance of the model.
The working process and principle of the invention are as follows: the invention provides a method for segmenting a network based on a semantic segmentation face image, which adopts a lightweight model, adopts a combination of a space channel and a context information channel, gradually increases subnets from high resolution to low resolution on the basis of an original space network structure to form more stages, and connects the subnets with multiple resolutions in parallel to obtain the information interaction module. And then, multi-scale fusion is carried out for multiple times, so that each high-resolution to low-resolution representation repeatedly receives information from other parallel representations, and abundant high-resolution representations are obtained. Because parallel connections are used, a high resolution representation can be maintained, and thus the prediction is spatially more accurate.
Example 2:
referring to fig. 1 to 4, the present embodiment discloses a method for segmenting a network based on a semantic segmentation face image, which includes the following steps:
in step S1, a corresponding image dataset for face segmentation is obtained.
Further, the step S1 includes:
step S11, adopting a known face recognition data set Labeled Faces in the world (LFW) for training LFW is a face recognition data set popular on the network, which contains more than 13000 face pictures, and using an extended version Part Labels of the LFW for training, wherein Labels of the data set are Labeled by a super-pixel segmentation algorithm contained in the data set, so that the obtained data set is Labeled. The data set then divides the training set, validation set, and test set by a number of 1500, 500, 1000, respectively.
And then a series of operations including averaging, defogging and cutting are carried out. The resulting picture is a 224 x 224 RGB input picture, awaiting training of the model.
Step S2, the structure of the segmented deep convolutional network is constructed.
Further, the step S2 includes:
in step S21, the network structure adopted by the invention is an encoder-decoder architecture which is popular in the current semantic segmentation field, so that the network structure can be roughly divided into two parts.
For the encoder portion, this portion is the subject of this network, step S22. The system comprises three small parts, namely a fast down-sampling module, an information interaction module and an expansion group. The network architecture mainly referred to comprises a down-sampling learning part of Fast-SCNN, an invoked bottleckresidual block module of MobileNet V2, and an FFM (feature Fusion module) attribute mechanism of lightweight BiSeNet which is popular in the semantic segmentation field at present, wherein a spatial channel and a context channel are adopted in a parallel mode, a high-resolution feature diagram of the spatial channel, which retains semantic information, and a low-resolution feature diagram of the context channel, which is obtained by rapidly down-sampling to increase a receptive field, are fused together, and the performance of the network can be well increased.
In step S221, for each fraction, the fast down-sampling module is composed of three convolutional layers, and the convolutional layers used have two forms, one is a standard convolutional layer, and the other is a depth-separable convolutional layer. The depth separable convolutional layer can effectively reduce the parameter quantity of the model, thereby reducing the calculation load. The convolution kernels used by all three convolution layers are (3 x 3), the step size is 2, each convolution layer is followed by a BN layer and a RELU activation function is used. The 224 × 3 pictures pass through the first convolution layer conv2D (3,3), after stride 2, a 112 × 32 signature is obtained, and then input to the second convolution layer Dwconv2D (3,3), after stride 2, a 56 × 64 signature is obtained, and then input to the third convolution layer Dwconv2D (3,3), after stride 2, a 28 × 64 signature is obtained.
In step S222, the structure of the information interaction module refers to an Inverted residual bottleneck block (Inverted bottle residual block) of MobileNet V2, and the structure of fig. 3 is adopted for design. Through the information interaction and combination of the feature maps with different dimensions, more beautiful output is obtained. Furthermore, because the employed up-sampling layer is a bilinear interpolation method, no learning parameter is needed, because the huge calculation amount brought by the transposition convolution can be greatly reduced. The 28 × 64 feature map obtained after step S221 is sampled by three different multiples, and the convolution layers are selected as conv2D (3,3), and the subsequent downsampling layers are selected as the convolution kernel in the same way. The up-sampling layer selects a bilinear up-sampling module. The step size is selected according to the scale of the resolution to be reduced, 1,2 and 4 are respectively selected, and three feature maps (1) (2) (3) with different resolution sizes are obtained. And then, carrying out downsampling convolution on the graph (1) according to step sizes of different scales to obtain three feature graphs (4), (5) and (6) with different resolutions. Fig. 2 is fused to fig. 4 after up-sampling twice and high resolution fig. 1, and fig. 4 is fused to fig. 3 after up-sampling four times. The upsampling of FIG. 3 is then doubled and merged into FIG. 5. Fig. 3 is directly merged with fig. 6 by the volume block. And adding the graph (5) and the graph (6), adding the result to the graph (4), and obtaining the feature graph with the output of 28 by 64 after passing through a feature fusion module.
In step S223, the expansion group performs spatial expansion convolution on the module subjected to feature fusion by the information interaction module, and through the expansion convolution, the receptive field of the convolution kernel can be increased, and more levels of context information can be captured. The feature maps obtained after step S222 are subjected to dilation convolution with dilation coefficients of 2,4, and 8, respectively, so that the receptive fields of the feature maps are increased, feature maps of three different receptive fields are obtained, and then the feature maps are added, so that the feature map size is 28 × 32.
In step S23, for the decoder module, pixel-level classification is performed by bilinear upsampling layer and convolutional layer followed by softmax layer. The size of the feature map obtained by the bilinear upsampling module in the feature map obtained in the step 223 is 224 × 32, and then the output feature map is obtained by adding a softmax layer to the convolution layer of conv2D, and the size is 224 × 3.
In step S24, post-processing is performed on the output image. Post-processing mechanisms can generally improve image edge detail and texture fidelity while maintaining a high degree of consistency with global information. Post-processing the output of the decoder, preserving details of face and hair edges by employing a guided filter, and reducing noise. The guiding filter can effectively suppress distortion and can smooth the contour of the edge to generate an edge which is comfortable for people to look at.
And step S3, using the network of S2 to train data to obtain a corresponding training model.
And step S4, carrying out verification by using the verification set, adjusting parameters and selecting an optimal model.
And step S5, testing the selected model by using the test set to evaluate the performance of the model. In the testing phase, we first used the MTCNN model with high recall to extract the ROI curve of the face, and then enlarged the ROI region by 0.8 times in both horizontal and vertical directions due to some contribution of the excess environmental information to segmentation of the background. We used four indices of the full convolution neural network (mIoU, fwIoU, pixelAcc, mPixelAcc) to evaluate the performance of the model. The experimental result of the model is compared with certain models of SOTA, such as VGG and U-Net, and the network structure obtained by balancing can well balance speed and performance and has good effect. Although the accuracy is not as good as that of some networks of SOTA in some aspects, the network has a great advantage in speed, and is a lightweight network architecture which has both speed and performance.
The above embodiments are preferred embodiments of the present invention, but the present invention is not limited to the above embodiments, and any other changes, modifications, substitutions, combinations, and simplifications which do not depart from the spirit and principle of the present invention should be construed as equivalents thereof, and all such changes, modifications, substitutions, combinations, and simplifications are intended to be included in the scope of the present invention.
Claims (5)
1. A method for segmenting a network based on a human face image segmented by semantics is characterized by comprising the following steps:
step S1: obtaining a corresponding image data set for face segmentation after a series of operations;
step S2: constructing a segmented deep convolutional network structure;
step S3: using the network structure of the step S2 to train data to obtain a corresponding training model;
step S4: carrying out verification and parameter adjustment by using the verification set, and selecting an optimal model;
step S5: and testing the selected optimal model by using a test set to evaluate the performance of the model.
2. The method for segmenting network based on semantic segmented face image according to claim 1, wherein the step S1 further comprises the following steps:
and step S11, adopting a face recognition data set laboratory Faces in the world (LFW), and dividing the training set, the verification set and the test set according to the proportion respectively.
3. The method for segmenting network based on semantic segmented face image according to claim 1, wherein the series of operations in the step S1 includes averaging, defogging and clipping operations.
4. The method for segmenting network based on semantic segmented face image according to claim 1, wherein the step S2 further comprises the following steps:
step S21: the deep convolutional network adopts an encoder-decoder architecture as a network structure, and the network structure comprises an encoder module and a decoder module;
step S22: the encoder module comprises three parts, namely a rapid downsampling module, an information interaction module and an expansion group;
step S23: the decoder module mainly comprises a bilinear up-sampling layer and a convolution layer, wherein the convolution layer is followed by a softmax layer to classify the pixel level;
step S24: post-processing the output of the decoder module, preserving details of face and hair edges by employing a guided filter, and reducing noise.
5. The method for segmenting network based on semantic segmented face image according to claim 4, wherein the step S22 further comprises the following steps:
step S221: the fast down-sampling module consists of three convolution layers, and the adopted convolution layers have two forms, one is a standard convolution layer, and the other is a depth-separable convolution layer; the convolution layer with separable depth can effectively reduce the parameter quantity of the model, thereby reducing the calculation burden; after each convolution layer, a BN layer and a RELU activation function are used;
step S222: the structure of the information interaction module adopts an inverted residual bottleneck block (Invertedbottleneck residual block) of MobileNet V2, and attractive output is obtained through information interaction and combination of feature graphs with different dimensions;
step S223: the expansion group is used for performing spatial expansion convolution on the module subjected to feature fusion through the information interaction module, and the expansion convolution can increase the receptive field of a convolution kernel and capture more levels of context information.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010628571.5A CN111899169B (en) | 2020-07-02 | 2020-07-02 | Method for segmenting network of face image based on semantic segmentation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010628571.5A CN111899169B (en) | 2020-07-02 | 2020-07-02 | Method for segmenting network of face image based on semantic segmentation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111899169A true CN111899169A (en) | 2020-11-06 |
CN111899169B CN111899169B (en) | 2024-01-26 |
Family
ID=73191423
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010628571.5A Active CN111899169B (en) | 2020-07-02 | 2020-07-02 | Method for segmenting network of face image based on semantic segmentation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111899169B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112508032A (en) * | 2021-01-29 | 2021-03-16 | 成都东方天呈智能科技有限公司 | Face image segmentation method and segmentation network for context information of association |
CN113096133A (en) * | 2021-04-30 | 2021-07-09 | 佛山市南海区广工大数控装备协同创新研究院 | Method for constructing semantic segmentation network based on attention mechanism |
CN113269786A (en) * | 2021-05-19 | 2021-08-17 | 青岛理工大学 | Assembly image segmentation method and device based on deep learning and guided filtering |
CN114267062A (en) * | 2021-12-07 | 2022-04-01 | 北京的卢深视科技有限公司 | Model training method, electronic device, and computer-readable storage medium |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107480707A (en) * | 2017-07-26 | 2017-12-15 | 天津大学 | A kind of deep neural network method based on information lossless pond |
CN108268870A (en) * | 2018-01-29 | 2018-07-10 | 重庆理工大学 | Multi-scale feature fusion ultrasonoscopy semantic segmentation method based on confrontation study |
US20180260793A1 (en) * | 2016-04-06 | 2018-09-13 | American International Group, Inc. | Automatic assessment of damage and repair costs in vehicles |
CN109543838A (en) * | 2018-11-01 | 2019-03-29 | 浙江工业大学 | A kind of image Increment Learning Algorithm based on variation self-encoding encoder |
CN109543502A (en) * | 2018-09-27 | 2019-03-29 | 天津大学 | A kind of semantic segmentation method based on the multiple dimensioned neural network of depth |
CN109801232A (en) * | 2018-12-27 | 2019-05-24 | 北京交通大学 | A kind of single image to the fog method based on deep learning |
CN110188817A (en) * | 2019-05-28 | 2019-08-30 | 厦门大学 | A kind of real-time high-performance street view image semantic segmentation method based on deep learning |
CN110837683A (en) * | 2019-05-20 | 2020-02-25 | 全球能源互联网研究院有限公司 | Training and predicting method and device for prediction model of transient stability of power system |
CN110895814A (en) * | 2019-11-30 | 2020-03-20 | 南京工业大学 | Intelligent segmentation method for aero-engine hole detection image damage based on context coding network |
CN111080650A (en) * | 2019-12-12 | 2020-04-28 | 哈尔滨市科佳通用机电股份有限公司 | Method for detecting looseness and loss faults of small part bearing blocking key nut of railway wagon |
-
2020
- 2020-07-02 CN CN202010628571.5A patent/CN111899169B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180260793A1 (en) * | 2016-04-06 | 2018-09-13 | American International Group, Inc. | Automatic assessment of damage and repair costs in vehicles |
CN107480707A (en) * | 2017-07-26 | 2017-12-15 | 天津大学 | A kind of deep neural network method based on information lossless pond |
CN108268870A (en) * | 2018-01-29 | 2018-07-10 | 重庆理工大学 | Multi-scale feature fusion ultrasonoscopy semantic segmentation method based on confrontation study |
CN109543502A (en) * | 2018-09-27 | 2019-03-29 | 天津大学 | A kind of semantic segmentation method based on the multiple dimensioned neural network of depth |
CN109543838A (en) * | 2018-11-01 | 2019-03-29 | 浙江工业大学 | A kind of image Increment Learning Algorithm based on variation self-encoding encoder |
CN109801232A (en) * | 2018-12-27 | 2019-05-24 | 北京交通大学 | A kind of single image to the fog method based on deep learning |
CN110837683A (en) * | 2019-05-20 | 2020-02-25 | 全球能源互联网研究院有限公司 | Training and predicting method and device for prediction model of transient stability of power system |
CN110188817A (en) * | 2019-05-28 | 2019-08-30 | 厦门大学 | A kind of real-time high-performance street view image semantic segmentation method based on deep learning |
CN110895814A (en) * | 2019-11-30 | 2020-03-20 | 南京工业大学 | Intelligent segmentation method for aero-engine hole detection image damage based on context coding network |
CN111080650A (en) * | 2019-12-12 | 2020-04-28 | 哈尔滨市科佳通用机电股份有限公司 | Method for detecting looseness and loss faults of small part bearing blocking key nut of railway wagon |
Non-Patent Citations (2)
Title |
---|
AHMED H. SHAHIN,ET AL.,: ""Deep Convolutional Encoder-Decoders with Aggregated Multi-Resolution Skip Connections for Skin Lesion Segmentation"", 《2019 IEEE 16TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2019)》 * |
石维康: ""基于深度神经网络的车型识别设计与实现"", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》, no. 2 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112508032A (en) * | 2021-01-29 | 2021-03-16 | 成都东方天呈智能科技有限公司 | Face image segmentation method and segmentation network for context information of association |
CN113096133A (en) * | 2021-04-30 | 2021-07-09 | 佛山市南海区广工大数控装备协同创新研究院 | Method for constructing semantic segmentation network based on attention mechanism |
CN113269786A (en) * | 2021-05-19 | 2021-08-17 | 青岛理工大学 | Assembly image segmentation method and device based on deep learning and guided filtering |
CN114267062A (en) * | 2021-12-07 | 2022-04-01 | 北京的卢深视科技有限公司 | Model training method, electronic device, and computer-readable storage medium |
CN114267062B (en) * | 2021-12-07 | 2022-12-16 | 合肥的卢深视科技有限公司 | Training method of face analysis model, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN111899169B (en) | 2024-01-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102640237B1 (en) | Image processing methods, apparatus, electronic devices, and computer-readable storage media | |
CN111210443B (en) | Deformable convolution mixing task cascading semantic segmentation method based on embedding balance | |
Bashir et al. | A comprehensive review of deep learning-based single image super-resolution | |
CN111899169B (en) | Method for segmenting network of face image based on semantic segmentation | |
US20190220746A1 (en) | Image processing method, image processing device, and training method of neural network | |
CN110136062B (en) | Super-resolution reconstruction method combining semantic segmentation | |
CN110276354B (en) | High-resolution streetscape picture semantic segmentation training and real-time segmentation method | |
Daihong et al. | Multi-scale generative adversarial network for image super-resolution | |
CN110223304B (en) | Image segmentation method and device based on multipath aggregation and computer-readable storage medium | |
CN112508960A (en) | Low-precision image semantic segmentation method based on improved attention mechanism | |
Singla et al. | A review on Single Image Super Resolution techniques using generative adversarial network | |
CN112884668A (en) | Lightweight low-light image enhancement method based on multiple scales | |
CN111353544A (en) | Improved Mixed Pooling-Yolov 3-based target detection method | |
Wang et al. | 4k-nerf: High fidelity neural radiance fields at ultra high resolutions | |
CN115131797A (en) | Scene text detection method based on feature enhancement pyramid network | |
Zhang et al. | Dynamic selection of proper kernels for image deblurring: a multistrategy design | |
US20230153965A1 (en) | Image processing method and related device | |
CN115512100A (en) | Point cloud segmentation method, device and medium based on multi-scale feature extraction and fusion | |
CN115115860A (en) | Image feature point detection matching network based on deep learning | |
CN114445418A (en) | Skin mirror image segmentation method and system based on convolutional network of multitask learning | |
US20230053588A1 (en) | Generating synthesized digital images utilizing a multi-resolution generator neural network | |
CN113012045B (en) | Generation countermeasure network for synthesizing medical image | |
Zhang et al. | Perception-oriented single image super-resolution network with receptive field block | |
Lin et al. | Image super-resolution by estimating the enhancement weight of self example and external missing patches | |
CN113947530B (en) | Image redirection method based on relative saliency detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |