CN110008832A - Based on deep learning character image automatic division method, information data processing terminal - Google Patents

Based on deep learning character image automatic division method, information data processing terminal Download PDF

Info

Publication number
CN110008832A
CN110008832A CN201910144508.1A CN201910144508A CN110008832A CN 110008832 A CN110008832 A CN 110008832A CN 201910144508 A CN201910144508 A CN 201910144508A CN 110008832 A CN110008832 A CN 110008832A
Authority
CN
China
Prior art keywords
picture
personage
convolution
group
feature mapping
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910144508.1A
Other languages
Chinese (zh)
Inventor
杨刚
李肖
师夏珍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN201910144508.1A priority Critical patent/CN110008832A/en
Publication of CN110008832A publication Critical patent/CN110008832A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Abstract

The invention belongs to technical field of image processing, disclose a kind of based on deep learning character image automatic division method, information data processing terminal;Personage's picture is collected, training data set is formed;Construct the deep neural network model of first order image, semantic segmentation;The training dataset being collected into is inputted into first order deep neural network, generates trimap;Construct second level deep neural network model;The training dataset being collected into and obtained trimap are inputted into second level deep neural network together, generate the personage's mask picture divided;Personage's mask picture is synthesized with personage's original image, obtains the personage's picture divided.The present invention automatically separates the personage of character image and image background according to the feature of character image;Automatically personage in image is screened, in conjunction with character features, separates figure and ground's picture;It can be used for personage's automation and scratch figure aspect, it can also be used to the replacement of portrait photographs' background, processing of the blurred background to background.

Description

Based on deep learning character image automatic division method, information data processing terminal
Technical field
The invention belongs to technical field of image processing, more particularly to one kind to be based on the deep learning character image side of segmentation automatically Method, information data processing terminal.
Background technique
Currently, the immediate prior art: character image segmentation refers to the prospect and background separation of portrait photographs, target Be that each pixel of the picture of input is classified: foreground and background gets classification chart in pixel scale.Image scratches figure Existing more than 30 years history, current existing many ripe algorithms still need to participate in by hand by user, and inefficiency lacks Weary automation is realized.There is the algorithm of many image segmentations in early stage, for example Threshold Segmentation Algorithm, the segmentation based on edge are calculated Method, zone broadening algorithm, watershed algorithm etc., these algorithms are only applicable in the case where image is fairly simple, for color Complicated image is difficult to obtain ideal segmentation result.The interaction that researcher attempts to be added user is insufficient to make up this, uses Family, which only needs to hook in the prospect of picture, determines several threads item, can automatically come out foreground segmentation.Occur in recent years based on figure This defect is compensated for as semantic partitioning algorithm is then opposite, better segmentation result can be generated, become computer graphics Important algorithm.Although occurring method many based on deep learning in image, semantic segmentation in recent years, such as FCN and its spread out Generation method such as SegNet and DeepLab, but these methods aim at general image, semantic segmentation, and are not implemented high-precision Character image segmentation, is not implemented high-precision main reason is that these methods cannot refine segmentation object edge, does not eliminate Carry out the up-sampling part bring error in self-learning networks.
In conclusion problem of the existing technology is: the existing method based on the segmentation of deep learning image, semantic exists The low problem of character image segmentation precision;Rely on manual interaction input, it is difficult to accomplish automatically to scratch figure, need a large amount of processing figures Inefficiency when piece;Automation personage, which scratches figure, has obscurity boundary.
Solve the difficulty of above-mentioned technical problem: character image divides one of difficult point automatically and is that personage scratches the accuracy of figure, The context enriched multiplicity of personage's picture realizes that automatic figure of scratching is extremely difficult under background complicated and changeable, especially when background and Personage's color is difficult to be split by normal image processing method when close;Second difficult point is the edge processing portion of personage Point, especially it is difficult to realize Accurate Segmentation by existing general semantics dividing method in place of personage's hair and background aliasing, very Hardly possible is realized in character image automatic and accurate segmenting hair silk.
Solve the meaning of above-mentioned technical problem: at present in place of personage scratches figure field still and has and largely needs manpower intervention, example Designer needs to devote considerable time to pluck out personage from portrait photographs when such as needing to replace background, blurred background, is handling Efficiency is especially low when a large amount of pictures.It realizes that high-precision, high performance character image semantic segmentation are of great significance, saves big The amount time can have many further applications once obtaining human body prospect, such as background replaces, synthesizes multiple personages' Image/video carries out stylization to personage and for example makes charactersketch, in video compress, virtual reality, production of film and TV and general There is major application prospect in video editing.In today that embedded amusement equipment continue to bring out, fine automation personage point It cuts algorithm and is particularly suitable for the equipment such as mobile phone, digital camera, camera.
Summary of the invention
In view of the problems of the existing technology, the present invention provides one kind to be based on the deep learning character image side of segmentation automatically Method, information data processing terminal.
The invention is realized in this way it is a kind of based on deep learning character image automatic division method, it is described to be based on depth Learn character image automatic division method and collect personage's picture, forms training data set;Construct the segmentation of first order image, semantic Deep neural network model;The training dataset being collected into is inputted into first order deep neural network, generates trimap;Building Second level deep neural network model;The training dataset being collected into and obtained trimap are inputted into second level depth mind together Through network, the personage's mask picture divided is generated;Personage's mask picture is synthesized with personage's original image, obtains the personage divided Picture.
Further, described to be specifically included based on deep learning character image automatic division method:
The first step collects personage's picture, detects face position in picture and records, marks human body area in every picture Domain and background area generate corresponding mask picture, form training data set;
Second step, the deep neural network model of building first order image, semantic segmentation, inputs as personage's picture and face Position exports as the trimap comprising foreground area, background area, uncertain region;
The training dataset being collected into is inputted first order deep neural network by third step, and deployment training obtains the first order Neural network model generates trimap, can get rough personage's segmentation result at this time;
4th step constructs second level deep neural network model, inputs as trimap, exports personage's mask to have divided Picture;
The training dataset being collected into and obtained trimap are inputted second level deep neural network by the 5th step together, Deployment training obtains second level neural network model, generates the personage's mask picture divided, and can get accurately personage at this time Foreground segmentation result;
Personage's mask picture is synthesized with personage's original image, obtains the personage's picture divided by the 6th step.
Further, the forming process of the training data set of the first step specifically:
(1) to acquisition photo preliminary treatment: adjusting size to 600*800*3 to collected portrait photographs;
(2) picture is marked, for each training sample picture, for the personage head completely occurred in picture, mark Personage's head center position, and everyone body is marked out with contour line to be saved in a text information;
(3) the corresponding mask picture of picture is generated, an i.e. 600*800*1 size identical with samples pictures length and width is generated Pixel value is all 0 two-value mask picture, and personage's head center position of above-mentioned mark is read from the text file generated in (2) It sets and body contour line, and is mapped on the mask picture, and all pixels value of human body parts is set as 1, after processing entirely Mask picture be original picture label.
Further, the deep neural network structure description of the second step are as follows:
First group of convolution: process of convolution is done to input picture using the convolution of two continuous 3*3*64, extracts feature;So The operation of first time pondization is carried out afterwards;
Second group of convolution: the convolution operation of two continuous 3*3*128 is carried out to the Feature Mapping figure of first time Chi Huahou; Then second of pondization operation is carried out;
Third group convolution: the convolution operation of two continuous 3*3*256 is carried out to the Feature Mapping figure of second of Chi Huahou; Then the operation of third time pondization is carried out;
4th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of third time Chi Huahou; Then the 4th pondization operation is carried out;
Warp lamination: up-sampling operation is carried out to the Feature Mapping figure of the 4th Chi Huahou, uses the volume of three 3*3*512 Product carries out convolution operation to Feature Mapping figure;
Output is the i.e. trimap picture of 600*800*1 identical as input picture length and width.
Further, the second level deep neural network structure description of the 4th step are as follows:
First group of convolution: process of convolution is done to input picture using the convolution of two continuous 3*3*64, extracts feature;So After carry out batch standardization processing, and pass through network activation function;Finally carry out the operation of first time pondization;
Second group of convolution: the convolution operation of two continuous 3*3*128 is carried out to the Feature Mapping figure of first time Chi Huahou; Then batch standardization processing is carried out, and passes through network activation function;Finally carry out second of pondization operation;
Third group convolution: the convolution operation of two continuous 3*3*256 is carried out to the Feature Mapping figure of second of Chi Huahou; Then batch standardization processing is carried out, and passes through network activation function;Finally carry out the operation of third time pondization;
4th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of third time Chi Huahou; Then batch standardization processing is carried out, and passes through network activation function;Finally carry out the 4th pondization operation;
5th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of the 4th Chi Huahou; Then batch standardization processing is carried out, and passes through network activation function;Finally carry out the 5th pondization operation;
First group of deconvolution: carry out to the Feature Mapping figure of the 5th group of Chi Huahou: up-sampling operates for the first time, uses two The convolution of 3*3*512 carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation letter Number;
Second group of deconvolution: carry out to the Feature Mapping figure of first group of deconvolution: second of up-sampling operates, and uses two The convolution of 3*3*512 carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation letter Number;
Third group deconvolution: carry out to the Feature Mapping figure of second group of deconvolution: third time up-sampling operates, and uses two The convolution of 3*3*256 carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation letter Number;
4th group of deconvolution: carry out to the Feature Mapping figure of third group deconvolution: the 4th up-sampling operation uses two The convolution of 3*3*128 carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation letter Number;
5th group of deconvolution: carry out to the Feature Mapping figure of the 4th group of deconvolution: the 5th up-sampling operation uses two The convolution of 3*3*64 carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation function;
Deconvolution: convolution operation is carried out to the Feature Mapping figure of the 5th group of deconvolution using the convolution of 3*3*64.
Further, the 6th step synthesizes personage's mask picture with personage's original image, obtains the personage's picture divided tool Body includes: the picture for creating an i.e. pixel value of 600*800*3 identical with the wide height of personage's original image and being all 0, traverses mask picture It is constant to be set to 0 for being 0 pixel in mask picture by all pixels point for position pixel pixel value in new picture;For Pixel value is not 0 pixel in mask picture, and the position pixel pixel value in new picture is set to the position in personage's original image Set pixel point value.
Another object of the present invention is to provide deep learning character image automatic division method is based on described in a kind of application Personage automate scratch figure BACKGROUNDA processing system.After getting personage's foreground picture, figure and ground is separated, background is carried out It is further processed such as Gaussian transformation, finally merges background and personage.
Another object of the present invention is to provide deep learning character image automatic division method is based on described in a kind of application Portrait photographs' background replace processing system.After getting personage's foreground picture, figure and ground is separated, background is replaced with Other backgrounds finally merge background and personage.
Another object of the present invention is to provide deep learning character image automatic division method is based on described in a kind of application Blurred background processing system.After getting personage's foreground picture, figure and ground is separated, background is done into Fuzzy Processing, then Background and personage are merged, achieve the effect that highlight personage.
Another object of the present invention is to provide deep learning character image automatic division method is based on described in a kind of application Information data processing terminal.
In conclusion advantages of the present invention and good effect are as follows: selected character image and determined, according to personage The feature of image automatically separates the personage of character image and image background;Personage's picture is inputted, computer can be automatically to figure Personage as in screens, and in conjunction with character features, separates figure and ground's picture.The present invention can be used for personage's automation and scratch figure Aspect, it can also be used to the processing to background such as the replacement of portrait photographs' background, blurred background.It saves a large amount of scratch by hand to scheme the time, one Denier obtains human body prospect, can there is many further applications, for example, background replacement, synthesize multiple personages image/video, Stylization is carried out to personage and for example makes charactersketch, in video compress, virtual reality, production of film and TV and general video editing There is major application prospect.In today that embedded amusement equipment continue to bring out, fine automation personage's partitioning algorithm is special It is suitble to the equipment such as mobile phone, digital camera, camera.
Compared to the existing technology for scratching figure manually to personage, the present invention, which can automate, realizes that personage scratches figure;It can be in complexity Personage's picture in be accurately partitioned into personage's picture;Personage's marginal information can be accurately plucked out, such as can accurately pluck out head The region that hair and background overlap;Circumstance of occlusion also has good prediction result.It was done with existing method in same character data collection Comparison, cuts (Graph-Cut) method for figure, segmentation effect is handed over and than (Mean IoU) 79.02%, full convolutional neural networks Segmentation effect hand over and ratio 73.08%, the segmentation effect of method proposed by the present invention is handed over and ratio reaches 90.61%, in essence It is obviously improved in exactness.
Detailed description of the invention
Fig. 1 is provided in an embodiment of the present invention based on deep learning character image automatic division method flow chart.
Fig. 2 is provided in an embodiment of the present invention based on deep learning character image automatic division method implementation flow chart.
Fig. 3 is original image provided in an embodiment of the present invention and the picture example schematic diagram divided.
Fig. 4 is first order deep neural network structural schematic diagram provided in an embodiment of the present invention.
Fig. 5 is deep neural network structural schematic diagram in the second level provided in an embodiment of the present invention.
Fig. 6 is the result schematic diagram provided in an embodiment of the present invention to personage's picture.
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, with reference to embodiments, to the present invention It is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, it is not used to Limit the present invention.
There is a problem of that character image segmentation precision is low for the prior art;Rely on manual interaction input, it is difficult to accomplish complete It is automatic to scratch figure, the inefficiency when needing a large amount of processing pictures;Automation personage, which scratches figure, has obscurity boundary.The present invention It provides the character image automatic division method based on deep learning not needing manually to participate in scratching figure process, and provides more acurrate Personage's segmented image.
Application principle of the invention is explained in detail with reference to the accompanying drawing.
As shown in Figure 1, the deep learning character image automatic division method provided in an embodiment of the present invention that is based on includes following Step:
S101: collecting personage's picture, detects face position in picture and records, marks human region in every picture And background area, corresponding mask picture is generated, a training data set is formed;
S102: the deep neural network model of building first order image, semantic segmentation inputs as personage's picture and face institute In position, export as the trimap comprising foreground area, background area, uncertain region;
S103: the training dataset being collected into is inputted into first order deep neural network, deployment training obtains first order mind Through network model, trimap is generated;
S104: building second level deep neural network model inputs as trimap, exports personage's mask figure to have divided Piece;
S105: the training dataset being collected into and obtained trimap are inputted into second level deep neural network, portion together Administration's training obtains second level neural network model, generates the personage's mask picture divided;
S106: personage's mask picture is synthesized with personage's original image, obtains the personage's picture divided.
Application principle of the invention is further described with reference to the accompanying drawing.
As shown in Fig. 2, the deep learning character image automatic division method provided in an embodiment of the present invention that is based on includes following Step:
Step 1 collects personage's picture, detects face position in picture and records, marks human body area in every picture Domain and background area generate corresponding mask picture, form a training data set, specific steps are as follows:
(1) to acquisition photo preliminary treatment: adjusting size to suitable size, such as this example to collected portrait photographs and adopt The personage's picture size integrated is 768*1024*3 (3 is picture depth, and picture is rgb format in the present embodiment), directly in original The real-time of model prediction is very poor when training will cause too long training time and training completion in pattern sheet, it is therefore desirable to adjustment figure Its size adjusting can be 600*800*3 size to suitable size by chip size;
(2) picture is marked, for each training sample picture, for the personage head completely occurred in picture, mark Its head center position, and everyone body is marked out with contour line to be saved in a text information;
(3) the corresponding mask picture of picture is generated, it is big to firstly generate an i.e. 600*800*1 identical with samples pictures length and width Small pixel value is all 0 two-value mask picture, from the personage head for reading above-mentioned mark in the text file generated in (2) Heart position and body contour line, and be mapped on the mask picture, and all pixels value of human body parts is set as 1 entirely, place Mask picture after reason is the label of original picture, i.e. the normally understood groundtruth of those skilled in the art, in Fig. 6 Mask picture.
Step 2, the deep neural network model of building first order image, semantic segmentation, inputs as personage's picture and face Position, exports as the trimap comprising foreground area, background area, uncertain region, and network structure is as shown in Figure 4.It is complete Whole network structure is described as follows:
First group of convolution: process of convolution is done to input picture using the convolution of two continuous 3*3*64, extracts feature;So The operation of first time pondization is carried out afterwards;
Second group of convolution: the convolution operation of two continuous 3*3*128 is carried out to the Feature Mapping figure of first time Chi Huahou; Then second of pondization operation is carried out;
Third group convolution: the convolution operation of two continuous 3*3*256 is carried out to the Feature Mapping figure of second of Chi Huahou; Then the operation of third time pondization is carried out;
4th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of third time Chi Huahou; Then the 4th pondization operation is carried out;
Warp lamination: up-sampling operation is carried out to the Feature Mapping figure of the 4th Chi Huahou, uses the volume of three 3*3*512 Product carries out convolution operation to Feature Mapping figure;
Output is the i.e. trimap picture of 600*800*1 identical as input picture length and width.
The training dataset being collected into is inputted first order deep neural network by step 3, and deployment training obtains the first order Neural network model, obtained model are that trimap generates model, and the trimap of generation will be as the defeated of second level neural network Enter.
Step 4 constructs second level deep neural network model, inputs as trimap, exports personage's mask to have divided Picture, network structure are as shown in Figure 5.Complete network structure is described as follows:
First group of convolution: process of convolution is done to input picture using the convolution of two continuous 3*3*64, extracts feature;So After carry out batch standardization processing (Batch Normalization), and pass through network activation function;Finally carry out first time pond Operation;
Second group of convolution: the convolution operation of two continuous 3*3*128 is carried out to the Feature Mapping figure of first time Chi Huahou; Then it carries out batch standardization processing (Batch Normalization), and passes through network activation function;Finally carry out second of pond Change operation;
Third group convolution: the convolution operation of two continuous 3*3*256 is carried out to the Feature Mapping figure of second of Chi Huahou; Then it carries out batch standardization processing (Batch Normalization), and passes through network activation function;Finally carry out third time pond Change operation;
4th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of third time Chi Huahou; Then it carries out batch standardization processing (Batch Normalization), and passes through network activation function;Finally carry out the 4th pond Change operation;
5th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of the 4th Chi Huahou; Then it carries out batch standardization processing (Batch Normalization), and passes through network activation function;Finally carry out the 5th pond Change operation;
First group of deconvolution: carry out to the Feature Mapping figure of the 5th group of Chi Huahou: up-sampling operates for the first time, uses two The convolution of 3*3*512 carries out convolution operation to Feature Mapping figure, then carries out crowd standardization processing (Batch Normalization), and pass through network activation function;
Second group of deconvolution: carry out to the Feature Mapping figure of first group of deconvolution: second of up-sampling operates, and uses two The convolution of 3*3*512 carries out convolution operation to Feature Mapping figure, then carries out crowd standardization processing (Batch Normalization), and pass through network activation function;
Third group deconvolution: carry out to the Feature Mapping figure of second group of deconvolution: third time up-sampling operates, and uses two The convolution of 3*3*256 carries out convolution operation to Feature Mapping figure, then carries out crowd standardization processing (Batch Normalization), and pass through network activation function;
4th group of deconvolution: carry out to the Feature Mapping figure of third group deconvolution: the 4th up-sampling operation uses two The convolution of 3*3*128 carries out convolution operation to Feature Mapping figure, then carries out crowd standardization processing (Batch Normalization), and pass through network activation function;
5th group of deconvolution: carry out to the Feature Mapping figure of the 4th group of deconvolution: the 5th up-sampling operation uses two The convolution of 3*3*64 carries out convolution operation to Feature Mapping figure, then carries out crowd standardization processing (Batch Normalization), and pass through network activation function;
Deconvolution: convolution operation is carried out to the Feature Mapping figure of the 5th group of deconvolution using the convolution of 3*3*64.
The trimap that the training dataset being collected into and step 2 obtain is inputted second level depth nerve by step 5 together Network, deployment training obtain second level neural network model, and obtained model is personage's mask model, exports the people to have divided Object mask picture.
Step 6 obtains the personage's picture divided as shown in fig. 6, personage's mask picture is synthesized with personage's original image.Tool Gymnastics is made as follows:
The pixel value of the i.e. 600*800*3 identical with the wide height of personage's original image of creation one is all 0 picture, traverses mask picture All pixels point sets the position pixel pixel value in new picture for being 0 pixel (as background) in mask picture It is constant for 0;It is not 0 pixel (as personage's prospect) for pixel value in mask picture, by the position pixel in new picture Point pixel value is set to the position pixel point value in personage's original image.
As shown in Figure 3 be it is proposed by the present invention based on deep learning character image automatic segmentation algorithm input original image and Personage's picture of the removal background of output.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all in essence of the invention Made any modifications, equivalent replacements, and improvements etc., should all be included in the protection scope of the present invention within mind and principle.

Claims (10)

1. one kind is based on deep learning character image automatic division method, which is characterized in that described to be based on deep learning figure map As automatic division method collects personage's picture, formation training data set;Construct the depth nerve of first order image, semantic segmentation Network model;The training dataset being collected into is inputted into first order deep neural network, generates trimap;Construct second level depth Neural network model;The training dataset being collected into and obtained trimap are inputted into second level deep neural network together, it is raw At the personage's mask picture divided;Personage's mask picture is synthesized with personage's original image, obtains the personage's picture divided.
2. being based on deep learning character image automatic division method as described in claim 1, which is characterized in that described based on deep Degree study character image automatic division method specifically includes:
The first step collects personage's picture, detects and face position and records in picture, mark in every picture human region and Background area generates corresponding mask picture, forms training data set;
Second step, the deep neural network model of building first order image, semantic segmentation, inputs as where personage's picture and face Position exports as the trimap comprising foreground area, background area, uncertain region;
The training dataset being collected into is inputted first order deep neural network by third step, and deployment training obtains first order nerve Network model generates trimap;
4th step constructs second level deep neural network model, inputs as trimap, exports personage's mask figure to have divided Piece;
The training dataset being collected into and obtained trimap are inputted second level deep neural network by the 5th step together, deployment Training obtains second level neural network model, generates the personage's mask picture divided;
Personage's mask picture is synthesized with personage's original image, obtains the personage's picture divided by the 6th step.
3. being based on deep learning character image automatic division method as claimed in claim 2, which is characterized in that the first step Training data set forming process specifically:
(1) to acquisition photo preliminary treatment: adjusting size to 600*800*3 to collected portrait photographs;
(2) picture is marked, personage is marked for the personage head completely occurred in picture for each training sample picture Head center position, and everyone body is marked out with contour line to be saved in a text information;
(3) the corresponding mask picture of picture is generated, the pixel of an i.e. 600*800*1 size identical with samples pictures length and width is generated Value is all 0 two-value mask picture, read in the text file generated from (2) personage's head center position of above-mentioned mark with And body contour line, and be mapped on the mask picture, and all pixels value of human body parts is set as 1 entirely, treated covers Code picture is the label of original picture.
4. being based on deep learning character image automatic division method as claimed in claim 2, which is characterized in that the second step Deep neural network structure description are as follows:
First group of convolution: process of convolution is done to input picture using the convolution of two continuous 3*3*64, extracts feature;Then into The operation of row first time pondization;
Second group of convolution: the convolution operation of two continuous 3*3*128 is carried out to the Feature Mapping figure of first time Chi Huahou;Then Carry out second of pondization operation;
Third group convolution: the convolution operation of two continuous 3*3*256 is carried out to the Feature Mapping figure of second of Chi Huahou;Then Carry out the operation of third time pondization;
4th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of third time Chi Huahou;Then Carry out the 4th pondization operation;
Warp lamination: up-sampling operation is carried out to the Feature Mapping figure of the 4th Chi Huahou, uses the convolution pair of three 3*3*512 Feature Mapping figure carries out convolution operation;
Output is the i.e. trimap picture of 600*800*1 identical as input picture length and width.
5. being based on deep learning character image automatic division method as claimed in claim 2, which is characterized in that the 4th step The second level deep neural network structure description are as follows:
First group of convolution: process of convolution is done to input picture using the convolution of two continuous 3*3*64, extracts feature;Then into Row batch standardization processing, and pass through network activation function;Finally carry out the operation of first time pondization;
Second group of convolution: the convolution operation of two continuous 3*3*128 is carried out to the Feature Mapping figure of first time Chi Huahou;Then Batch standardization processing is carried out, and passes through network activation function;Finally carry out second of pondization operation;
Third group convolution: the convolution operation of two continuous 3*3*256 is carried out to the Feature Mapping figure of second of Chi Huahou;Then Batch standardization processing is carried out, and passes through network activation function;Finally carry out the operation of third time pondization;
4th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of third time Chi Huahou;Then Batch standardization processing is carried out, and passes through network activation function;Finally carry out the 4th pondization operation;
5th group of convolution: the convolution operation of two continuous 3*3*512 is carried out to the Feature Mapping figure of the 4th Chi Huahou;Then Batch standardization processing is carried out, and passes through network activation function;Finally carry out the 5th pondization operation;
First group of deconvolution: carry out to the Feature Mapping figure of the 5th group of Chi Huahou: up-sampling operates for the first time, uses two 3*3* 512 convolution carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation function;
Second group of deconvolution: carry out to the Feature Mapping figure of first group of deconvolution: second of up-sampling operates, and uses two 3*3* 512 convolution carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation function;
Third group deconvolution: carry out to the Feature Mapping figure of second group of deconvolution: third time up-sampling operates, and uses two 3*3* 256 convolution carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation function;
4th group of deconvolution: carry out to the Feature Mapping figure of third group deconvolution: the 4th up-sampling operation uses two 3*3* 128 convolution carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation function;
5th group of deconvolution: carry out to the Feature Mapping figure of the 4th group of deconvolution: the 5th up-sampling operation uses two 3*3* 64 convolution carries out convolution operation to Feature Mapping figure, then carries out batch standardization processing, and pass through network activation function;
Deconvolution: convolution operation is carried out to the Feature Mapping figure of the 5th group of deconvolution using the convolution of 3*3*64.
6. being based on deep learning character image automatic division method as claimed in claim 2, which is characterized in that the 6th step Personage's mask picture is synthesized with personage's original image, the personage's picture for obtaining having divided specifically includes: creation one and personage's original image The pixel value of the identical i.e. 600*800*3 of wide height is all 0 picture, mask picture all pixels point is traversed, in mask picture For 0 pixel, it is constant that position pixel pixel value in new picture is set to 0;It is not 0 picture for pixel value in mask picture The position pixel pixel value in new picture is set to the position pixel point value in personage's original image by vegetarian refreshments.
7. a kind of personage using based on deep learning character image automatic division method described in claim 1~6 any one It automates and scratches figure BACKGROUNDA processing system.
8. a kind of personage using based on deep learning character image automatic division method described in claim 1~6 any one Photo background replaces processing system.
9. a kind of background using based on deep learning character image automatic division method described in claim 1~6 any one Fuzzy Processing system.
10. a kind of information using based on deep learning character image automatic division method described in claim 1~6 any one Data processing terminal.
CN201910144508.1A 2019-02-27 2019-02-27 Based on deep learning character image automatic division method, information data processing terminal Pending CN110008832A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910144508.1A CN110008832A (en) 2019-02-27 2019-02-27 Based on deep learning character image automatic division method, information data processing terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910144508.1A CN110008832A (en) 2019-02-27 2019-02-27 Based on deep learning character image automatic division method, information data processing terminal

Publications (1)

Publication Number Publication Date
CN110008832A true CN110008832A (en) 2019-07-12

Family

ID=67166074

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910144508.1A Pending CN110008832A (en) 2019-02-27 2019-02-27 Based on deep learning character image automatic division method, information data processing terminal

Country Status (1)

Country Link
CN (1) CN110008832A (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110363728A (en) * 2019-07-25 2019-10-22 上海尊溢商务信息咨询有限公司 A kind of image trimap generation method based on semantic segmentation
CN110414480A (en) * 2019-08-09 2019-11-05 威盛电子股份有限公司 Training image production method and electronic device
CN110458918A (en) * 2019-08-16 2019-11-15 北京百度网讯科技有限公司 Method and apparatus for output information
CN110705558A (en) * 2019-09-29 2020-01-17 郑州阿帕斯科技有限公司 Image instance segmentation method and device
CN110992374A (en) * 2019-11-28 2020-04-10 杭州趣维科技有限公司 Hair refined segmentation method and system based on deep learning
CN111222440A (en) * 2019-12-31 2020-06-02 江西开心玉米网络科技有限公司 Portrait background separation method, device, server and storage medium
CN111339338A (en) * 2020-02-29 2020-06-26 西安理工大学 Text picture matching recommendation method based on deep learning
CN111353069A (en) * 2020-02-04 2020-06-30 清华珠三角研究院 Character scene video generation method, system, device and storage medium
CN111368853A (en) * 2020-02-04 2020-07-03 清华珠三角研究院 Label construction method, system, device and storage medium
CN111462161A (en) * 2020-03-31 2020-07-28 厦门亿联网络技术股份有限公司 System, method, storage medium and equipment for extracting real-time video portrait
CN112581480A (en) * 2020-12-22 2021-03-30 深圳市雄帝科技股份有限公司 Automatic image matting method, system and readable storage medium thereof
CN112614149A (en) * 2020-12-11 2021-04-06 北京影谱科技股份有限公司 Semantic synthesis method based on instance segmentation
CN113052755A (en) * 2019-12-27 2021-06-29 杭州深绘智能科技有限公司 High-resolution image intelligent matting method based on deep learning
CN113052242A (en) * 2021-03-29 2021-06-29 北京达佳互联信息技术有限公司 Training method and device of image processing network, and image processing method and device
CN113973178A (en) * 2021-10-24 2022-01-25 云景文旅科技有限公司 Interactive photographing processing method and device in travel process
CN114140547A (en) * 2021-12-07 2022-03-04 北京百度网讯科技有限公司 Image generation method and device
WO2023239299A1 (en) * 2022-06-10 2023-12-14 脸萌有限公司 Image processing method and apparatus, electronic device, and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107123088A (en) * 2017-04-21 2017-09-01 山东大学 A kind of method of automatic replacing photo background color
JP2018136789A (en) * 2017-02-22 2018-08-30 日本電信電話株式会社 Image processing device, image processing method and image processing program
CN108961303A (en) * 2018-07-23 2018-12-07 北京旷视科技有限公司 A kind of image processing method, device, electronic equipment and computer-readable medium
CN108986132A (en) * 2018-07-04 2018-12-11 华南理工大学 A method of certificate photo Trimap figure is generated using full convolutional neural networks

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018136789A (en) * 2017-02-22 2018-08-30 日本電信電話株式会社 Image processing device, image processing method and image processing program
CN107123088A (en) * 2017-04-21 2017-09-01 山东大学 A kind of method of automatic replacing photo background color
CN108986132A (en) * 2018-07-04 2018-12-11 华南理工大学 A method of certificate photo Trimap figure is generated using full convolutional neural networks
CN108961303A (en) * 2018-07-23 2018-12-07 北京旷视科技有限公司 A kind of image processing method, device, electronic equipment and computer-readable medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SU扬帆启航: "《深度学习语义分割(二)SegNet论文解读》", 《HTTPS://BLOG.CSDN.NET/ORANGE_LITTLEGIRL/ARTICLE/DETAILS/80785589》 *

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110363728A (en) * 2019-07-25 2019-10-22 上海尊溢商务信息咨询有限公司 A kind of image trimap generation method based on semantic segmentation
CN110414480A (en) * 2019-08-09 2019-11-05 威盛电子股份有限公司 Training image production method and electronic device
CN110458918A (en) * 2019-08-16 2019-11-15 北京百度网讯科技有限公司 Method and apparatus for output information
CN110458918B (en) * 2019-08-16 2023-05-09 北京百度网讯科技有限公司 Method and device for outputting information
CN110705558A (en) * 2019-09-29 2020-01-17 郑州阿帕斯科技有限公司 Image instance segmentation method and device
CN110705558B (en) * 2019-09-29 2022-03-08 郑州阿帕斯科技有限公司 Image instance segmentation method and device
CN110992374A (en) * 2019-11-28 2020-04-10 杭州趣维科技有限公司 Hair refined segmentation method and system based on deep learning
CN110992374B (en) * 2019-11-28 2023-09-05 杭州小影创新科技股份有限公司 Hair refinement segmentation method and system based on deep learning
CN113052755A (en) * 2019-12-27 2021-06-29 杭州深绘智能科技有限公司 High-resolution image intelligent matting method based on deep learning
CN111222440A (en) * 2019-12-31 2020-06-02 江西开心玉米网络科技有限公司 Portrait background separation method, device, server and storage medium
CN111368853A (en) * 2020-02-04 2020-07-03 清华珠三角研究院 Label construction method, system, device and storage medium
CN111353069A (en) * 2020-02-04 2020-06-30 清华珠三角研究院 Character scene video generation method, system, device and storage medium
CN111339338B (en) * 2020-02-29 2023-03-07 西安理工大学 Text picture matching recommendation method based on deep learning
CN111339338A (en) * 2020-02-29 2020-06-26 西安理工大学 Text picture matching recommendation method based on deep learning
CN111462161A (en) * 2020-03-31 2020-07-28 厦门亿联网络技术股份有限公司 System, method, storage medium and equipment for extracting real-time video portrait
CN111462161B (en) * 2020-03-31 2023-09-26 厦门亿联网络技术股份有限公司 System, method, storage medium and equipment for extracting real-time video image
CN112614149A (en) * 2020-12-11 2021-04-06 北京影谱科技股份有限公司 Semantic synthesis method based on instance segmentation
CN112581480A (en) * 2020-12-22 2021-03-30 深圳市雄帝科技股份有限公司 Automatic image matting method, system and readable storage medium thereof
CN113052242A (en) * 2021-03-29 2021-06-29 北京达佳互联信息技术有限公司 Training method and device of image processing network, and image processing method and device
CN113973178A (en) * 2021-10-24 2022-01-25 云景文旅科技有限公司 Interactive photographing processing method and device in travel process
CN114140547B (en) * 2021-12-07 2023-03-14 北京百度网讯科技有限公司 Image generation method and device
CN114140547A (en) * 2021-12-07 2022-03-04 北京百度网讯科技有限公司 Image generation method and device
WO2023239299A1 (en) * 2022-06-10 2023-12-14 脸萌有限公司 Image processing method and apparatus, electronic device, and storage medium

Similar Documents

Publication Publication Date Title
CN110008832A (en) Based on deep learning character image automatic division method, information data processing terminal
CN109034208B (en) High-low resolution combined cervical cell slice image classification system
CN108985181B (en) End-to-end face labeling method based on detection segmentation
CN109154978B (en) System and method for detecting plant diseases
CN110400323B (en) Automatic cutout system, method and device
CN107862698B (en) Light field foreground segmentation method and device based on K mean cluster
EP3819859B1 (en) Sky filter method for panoramic images and portable terminal
CN109035269B (en) Cervical cell pathological section pathological cell segmentation method and system
CN107452010A (en) A kind of automatically stingy nomography and device
CN108986132B (en) Method for generating certificate photo Trimap graph by using full convolution neural network
WO2018145470A1 (en) Image detection method and device
CN107256558A (en) The cervical cell image automatic segmentation method and system of a kind of unsupervised formula
CN106384112A (en) Rapid image text detection method based on multi-channel and multi-dimensional cascade filter
CN107240084A (en) A kind of removing rain based on single image method and device
JP2010511215A (en) How to indicate an object in an image
CN110956681A (en) Portrait background automatic replacement method combining convolutional network and neighborhood similarity
CN111462162B (en) Foreground segmentation algorithm for specific class pictures
CN109886170A (en) A kind of identification of oncomelania intelligent measurement and statistical system
CN109636807A (en) A kind of grape disease blade split plot design of image segmentation and pixel recovery
CN110969631A (en) Fine photo hair dyeing method and system
CN111325688B (en) Unmanned aerial vehicle image defogging method for optimizing atmosphere light by fusion morphology clustering
CN111476794A (en) UNET-based cervical pathological tissue segmentation method
CN113160185A (en) Method for guiding cervical cell segmentation by using generated boundary position
CN111160194B (en) Static gesture image recognition method based on multi-feature fusion
CN112651929A (en) Medical image organ segmentation method and system based on three-dimensional full-convolution neural network and region growth

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination