CN110414428A - A method of generating face character information identification model - Google Patents
A method of generating face character information identification model Download PDFInfo
- Publication number
- CN110414428A CN110414428A CN201910686308.9A CN201910686308A CN110414428A CN 110414428 A CN110414428 A CN 110414428A CN 201910686308 A CN201910686308 A CN 201910686308A CN 110414428 A CN110414428 A CN 110414428A
- Authority
- CN
- China
- Prior art keywords
- segmentation
- network
- training
- mask images
- character information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/179—Human faces, e.g. facial parts, sketches or expressions metadata assisted face recognition
Abstract
The invention discloses a kind of methods for generating face character information identification model, comprising: generates the mask images for being labeled with multiple face character information corresponding with original image, and multiple attribute tags of characterization face character information;It will be handled in the segmentation network of original image input pre-training, the mask images predicted, and based on the segmentation penalty values between mask images generated and the mask images predicted, training segmentation network obtains the first segmentation network;Generate the characteristic image for identifying at least one attributive character;Each characteristic image is inputted in the sorter network of each pre-training respectively and handled, the Classification Loss value between the attribute tags of attribute tags and prediction with the attribute tags predicted, and based on mark, training sorter network obtains multiple first sorter networks;And based on segmentation penalty values and Classification Loss value, network and multiple first sorter networks are divided in training first, generate face character information identification model with coupling.
Description
Technical field
The present invention relates to depth learning technology field more particularly to a kind of sides for generating face character information identification model
Method, face character information identifying method, mobile terminal and storage medium.
Background technique
Face character information is identified by detection facial image, obtains the information of multiple attributes in the facial image, such as
Age, gender, expression etc..Face character identification has important role in terms of customer analysis, virtual image.For people
Face attribute, such as eyelid, glasses, beard carry out identifying that there are problems.For example, eyelid is divided into single-edge eyelid, interior double and simple eye
Skin etc., needs consider that whether there is or not eyeglasses when describing glasses, and size, the color of eyeglass, whether there is or not frame, shape, the thickness of frame etc. are retouched
The position of beard is not only considered when stating beard, it is also contemplated that the length of beard, dense degree etc..This kind of identification mission key exists
In given label, how different classes of under same label is distinguished.And multiple labels of same attribute frequently result in data
It is unbalanced.And the position of attribute region in the picture be it is uncertain, be easy by face, other attributes are influenced, such as carry out eyelid
By the interference of false eyelashes, eye shadow, false double-edged eyelid and other background areas when Attribute Recognition.Such as eyelid Attribute Recognition, no
Avoidable ground can cut out eye portion, for beard Attribute Recognition, inevitably cut out mouth part, individual attribute
Recognition effect is easy to be influenced by these factors.
Existing face character recognition methods generally utilize traditional characteristic extracting method combination supporting vector machine scheduling algorithm into
Row classification, such methods often have the shortcomings that nicety of grading is low computationally intensive.
Therefore, it is necessary to a kind of face character recognition methods, and multiple categories of face can be predicted by single Neural
Property, improve the accuracy and efficiency of face character identification.
Summary of the invention
For this purpose, the present invention provides a kind of methods for generating face character information identification model and the identification of face character information
Method, with try hard to solve the problems, such as or at least alleviate above it is existing at least one.
According to an aspect of the invention, there is provided a kind of method for generating face character information identification model, this method
Suitable for executing in the terminal.Firstly, generating the mask artwork for being labeled with multiple face character information corresponding with original image
Picture, and characterization face character information multiple attribute tags then, by original image input pre-training segmentation network in into
Row processing, the mask images predicted, and based on the segmentation between mask images generated and the mask images predicted
Penalty values, training segmentation network, obtain the first segmentation network.And generate the characteristic image for identifying at least one attributive character.
Then, each characteristic image is inputted in the sorter network of each pre-training respectively and is handled, with the attribute tags predicted, and
Attribute tags based on mark and the Classification Loss value between the attribute tags of prediction, training sorter network, obtain multiple first
Sorter network.Finally, based on segmentation penalty values and Classification Loss value, the first segmentation network of training and multiple first sorter networks,
To utilize the first segmentation network and multiple first sorter networks coupling generation face character information identification model after training.
Optionally, in the above-mentioned methods, mask images are generated based on semantic segmentation, and are shown using different colours to characterize
Different face characters.
Optionally, in the above-mentioned methods, segmentation network includes multiple lightweight process of convolution modules, lightweight process of convolution
Module separates convolution using depth.
Optionally, in the above-mentioned methods, the mask images of the mask images based on generation and prediction calculate segmentation loss
Value.And based on segmentation penalty values, inverse iteration updates the parameter of segmentation network, training knot when meeting the first predetermined condition
Beam obtains the first segmentation network.
Optionally, in the above-mentioned methods, the first predetermined condition is that segmentation penalty values no longer decline.
Optionally, in the above-mentioned methods, segmentation penalty values are calculated based on following formula:
Wherein, H (p, q) is segmentation penalty values, and p is the mask images generated, and q is the mask images of prediction, and p (x) makes a living
At mask images pixel Distribution value, q (x) be prediction mask images pixel Distribution value, x indicate image in pixel
Point, n are the quantity of pixel.
Optionally, in the above-mentioned methods, it will handle, obtained at least in the segmentation network of original image input pre-training
One feature channel figure;The mask images of first segmentation neural network forecast and at least one feature channel figure are subjected to point multiplication operation,
Obtain at least one characteristic image.
Optionally, in the above-mentioned methods, the attribute tags of the attribute tags based on mark and prediction calculate Classification Loss
Value;And it is based on Classification Loss value, inverse iteration updates the parameter of sorter network, training knot when meeting the second predetermined condition
Beam obtains multiple first sorter networks.
Optionally, in the above-mentioned methods, the second predetermined condition is that Classification Loss value no longer declines or front and back iteration twice
The difference of the penalty values of the classification function of calculating reaches the first pre-determined number less than the first predetermined threshold or the number of iterations.
Optionally, in the above-mentioned methods, Classification Loss value is calculated by following formula:
Wherein, l (y, z) is Classification Loss value, and y is the attribute tags of mark, and z is the attribute tags of prediction, zjIndicate pre-
The label value for j-th of the attribute tags surveyed, zyIndicate that the label value of the attribute tags of mark, m indicate the quantity of attribute tags.
Optionally, in the above-mentioned methods, weight based on segmentation penalty values and Classification Loss value and, inverse iteration update the
The parameter of one segmentation network and multiple first sorter networks, when meeting third predetermined condition, training terminates, wherein third
Predetermined condition be weight and no longer decline or the difference of weight sum that front and back iterates to calculate twice less than the second predetermined threshold or
Person's the number of iterations reaches the second pre-determined number.
Optionally, in the above-mentioned methods, face character includes eyelid, glasses, beard, cap, hair.Attribute tags include
One or more of following label: the label of eyelid, the label of glasses, the label of beard, the label of cap, hair mark
Label.
Optionally, in the above-mentioned methods, the label of eyelid includes single-edge eyelid, interior double and double-edged eyelid;The label of glasses includes
The presence or absence of eyeglass, size, color, the presence or absence of spectacle-frame, shape, size, thickness;The label of beard includes the length of beard, class
Type, dense degree;The label of cap includes the style of cap, color;The label of hair includes the length, color, hair style of hair.
Optionally, the step of generating the mask images for being labeled with multiple face character information corresponding with original image it
Before, human face region is carried out to original image and is cut out.
Optionally, in the above-mentioned methods, it is based on characteristic point detection model, obtains position and the characteristic point coordinate of face;Base
Face is carried out in face location and characteristic point coordinate to ajust, and obtains the consistent facial image of size to cut.
According to another aspect of the present invention, a kind of face character information identifying method is provided, is suitable in mobile terminal
Middle execution, firstly, obtaining facial image to be identified.Then, by facial image input face character information identification model in into
Row processing, face character information identification model includes the segmentation network and sorter network being mutually coupled, wherein facial image is defeated
Enter to divide in network and handle, obtains the mask images and feature channel figure of multiple face characters;By mask images and feature channel
Figure carries out point multiplication operation, obtains each attribute region figure.Finally, each attribute region figure is inputted corresponding sorter network respectively
In handled, obtain the information of each face character in facial image.
Optionally, in the above-mentioned methods, face character includes eyelid, glasses, beard, cap, hair;The letter of face character
Breath includes one or more of following information: single-edge eyelid, interior double and double-edged eyelid and/or eyeglass size, colors, frame have
Nothing, shape, size, the length of thickness and/or beard, type, dense degree and/or cap style, color and/or hair
Length, color, hair style.
Optionally, in the above-mentioned methods, human face region cutting is carried out to the facial image of acquisition, it is consistent to obtain size
Facial image.
Another aspect according to the present invention, provides a kind of mobile terminal, including one or more processors, memory and
One or more programs, wherein one or more programs store in memory and are configured as being held by one or more processors
Row, one or more programs include the instruction for executing the above method.
Another aspect according to the present invention provides a kind of computer readable storage medium for storing one or more programs.
Here one or more programs include instruction, when these instructions are by mobile terminal execution, so that in the mobile terminal execution
State method.
According to the solution of the present invention, by carrying out semantic segmentation to facial image, each face character can effectively be obtained
Semantic information can be improved the accuracy of face character information identification;Image segmentation is carried out by using lightweight convolution, it can
Meet the needs of mobile terminal is to speed and calculation amount, improves the efficiency of face character information identification.
Detailed description of the invention
To the accomplishment of the foregoing and related purposes, certain illustrative sides are described herein in conjunction with following description and drawings
Face, these aspects indicate the various modes that can practice principles disclosed herein, and all aspects and its equivalent aspect
It is intended to fall in the range of theme claimed.Read following detailed description in conjunction with the accompanying drawings, the disclosure it is above-mentioned
And other purposes, feature and advantage will be apparent.Throughout the disclosure, identical appended drawing reference generally refers to identical
Component or element.
Fig. 1 shows the organigram of mobile terminal 100 according to an embodiment of the invention;
Fig. 2 shows the methods 200 according to an embodiment of the invention for generating face character information identification model
Schematic flow chart;
Fig. 3 shows the schematic diagram of mask images according to an embodiment of the invention;
Fig. 4 shows the process that training according to an embodiment of the invention generates face character information identification model and shows
It is intended to;
Fig. 5 shows the schematic flow of face character information identifying method 500 according to an embodiment of the invention
Figure.
Specific embodiment
Exemplary embodiments of the present disclosure are described in more detail below with reference to accompanying drawings.Although showing the disclosure in attached drawing
Exemplary embodiment, it being understood, however, that may be realized in various forms the disclosure without should be by embodiments set forth here
It is limited.On the contrary, these embodiments are provided to facilitate a more thoroughly understanding of the present invention, and can be by the scope of the present disclosure
It is fully disclosed to those skilled in the art.
Face character identification is to the local feature (eyelid, hair, beard etc.) of such as face, gender, age, expression, decorations
The information such as object are identified.In order to improve the precision and efficiency of face character identification, present solution provides one kind based on semantic point
The face character information identifying method cut can effectively be obtained the semantic information in facial image by semantic segmentation, be reduced
The influence of fitting and data nonbalance to accuracy of identification, improves the accuracy of face character information identification.
Fig. 1 shows the organigram of mobile terminal 100 according to an embodiment of the invention.Mobile terminal 100
It may include memory interface 102, one or more data processors, image processor and/or central processing unit 104, aobvious
Display screen curtain (not shown in figure 1) and peripheral interface 106.
Memory interface 102, one or more processors 104 and/or peripheral interface 106 either discrete component,
It can integrate in one or more integrated circuits.In the mobile terminal 100, various elements can pass through one or more communication
Bus or signal wire couple.Sensor, equipment and subsystem may be coupled to peripheral interface 106, a variety of to help to realize
Function.
For example, motion sensor 110, light sensor 112 and range sensor 114 may be coupled to peripheral interface 106,
To facilitate the functions such as orientation, illumination and ranging.Other sensors 116 can equally be connected with peripheral interface 106, such as positioning system
System (such as GPS receiver), temperature sensor, biometric sensor or other sensor devices, it is possible thereby to help to implement phase
The function of pass.
Camera sub-system 120 and optical sensor 122 can be used for the camera of convenient such as record photos and video clips
The realization of function, wherein the camera sub-system and optical sensor for example can be charge-coupled device (CCD) or complementary gold
Belong to oxide semiconductor (CMOS) optical sensor.It can help to realize by one or more radio communication subsystems 124
Communication function, wherein radio communication subsystem may include radio-frequency transmitter and transmitter and/or light (such as infrared) receiver
And transmitter.The particular design and embodiment of radio communication subsystem 124 can depend on mobile terminal 100 is supported one
A or multiple communication networks.For example, mobile terminal 100 may include be designed to support LTE, 3G, GSM network, GPRS network,
EDGE network, Wi-Fi or WiMax network and BluetoothTMThe communication subsystem 124 of network.
Audio subsystem 126 can be coupled with loudspeaker 128 and microphone 130, to help to implement to enable voice
Function, such as speech recognition, speech reproduction, digital record and telephony feature.I/0 subsystem 140 may include touch screen control
Device 142 processed and/or other one or more input controllers 144.Touch screen controller 142 may be coupled to touch screen 146.It lifts
For example, any one of a variety of touch-sensing technologies are can be used to detect in the touch screen 146 and touch screen controller 142
The contact and movement or pause carried out therewith, wherein detection technology includes but is not limited to capacitive character, resistive, infrared and table
Face technology of acoustic wave.Other one or more input controllers 144 may be coupled to other input/control devicess 148, such as one
Or the pointer device of multiple buttons, rocker switch, thumb wheel, infrared port, USB port, and/or stylus etc.It is described
One or more button (not shown)s may include the up/down for controlling 130 volume of loudspeaker 128 and/or microphone
Button.
Memory interface 102 can be coupled with memory 150.The memory 150 may include that high random access is deposited
Reservoir and/or nonvolatile memory, such as one or more disk storage equipments, one or more optical storage apparatus, and/
Or flash memories (such as NAND, NOR).Memory 150 can store an operating system 172, for example, Android, iOS or
The operating system of Windows Phone etc.The operating system 172 may include for handling basic system services and execution
The instruction of task dependent on hardware.Memory 150 can also store one or more programs 174.In mobile device operation,
Meeting load operating system 172 from memory 150, and executed by processor 104.Program 174 at runtime, also can be from storage
It loads in device 150, and is executed by processor 104.Program 174 operates on operating system, utilizes operating system and bottom
The interface that hardware provides realizes the various desired functions of user, such as instant messaging, web page browsing, pictures management.Program 174 can
To be independently of operating system offer, it is also possible to what operating system carried.In addition, program 174 is mounted to mobile terminal
When in 100, drive module can also be added to operating system.Program 174 may be arranged on an operating system by one or more
A processor 104 executes relevant instruction.In some embodiments, mobile terminal 100 is configured as executing according to the present invention
Generate the method 200 and face character information identifying method 500 of face character information identification model.Wherein, mobile terminal 100
One or more programs 174 include for executing method 200 He according to the present invention for generating face character information identification model
The instruction of face character information identifying method 500.
Image, semantic segmentation is exactly to be partitioned into subject area from image automatically, and by each pixel generic of image
It is labeled, different colours can represent different classifications, divide an image into semantically meaningful part.Since face has
Multiple attributes, such as glasses, eyelid, beard, each attribute have multiple labels according to different classification methods, as beard can be by
According to length, straight, shape progress multi-tag classification is rolled up, so face character identification is the learning process of a multitask.
Fig. 2 shows the methods 200 according to an embodiment of the invention for generating face character information identification model
Schematic flow chart.This method can execute in the mobile terminal 100.Face character information identification model may include segmentation net
Network and multiple sorter networks arranged side by side.The corresponding face character of each sorter network, is a multi-tag sorter network, has
Multiple outputs, each output correspond to an attribute tags.
As shown in Fig. 2, method 200 starts from step S210, generation is corresponding with original image to be labeled with multiple face characters
The mask images of information, and multiple attribute tags of characterization face character information.
Wherein, generating mask images is exactly to be blocked with selected image to the whole of image to be processed or part, is come
Control region or the treatment process of image procossing.Specific image or object for covering are known as mask.Mask images can be with base
It generates, can be shown using different colours to characterize different face characters in semantic segmentation.
Wherein, attribute tags are second level label, and the first order is different latitude (such as size, shape), and the second level is each
The tag along sort of latitude.Such as the label of glasses are as follows: and type: eyeglasses, size: in, shape: rectangular }.Other attributes
Label is similar.According to one embodiment of present invention, face character includes eyelid, glasses, beard, cap, hair etc..Table 1 shows
Attribute tags according to an embodiment of the invention are gone out.As shown in table 1, glasses can according to when whether there is or not classification, having a glasses by
Look in the mirror piece whether there is or not, size, color, frame whether there is or not, shape, thickness, material etc. be labeled.For example, frame can be divided into framed, nothing
Frame can be further subdivided into thick frame, thin frame if framed, further according to color classification, the thin frame of black, red thick frame
Deng.For hair attribute, can classify according to hair lengths, color, hair style etc..It can be according to the face of cap for cap attribute
Color, style etc. are classified.For beard attribute, can classify according to the type of beard, position, length, dense degree etc..
For eyelid attribute, single-edge eyelid, double-edged eyelid, interior double-edged eyelid can be divided into according to eyelid type.Attribute tags in table 1 are only to show
Example property, it can be labeled according to actual needs.It should be noted that not intersecting between each attribute tags, for example, In
When classifying to beard attribute, it cannot be placed on by the Xiao Hu's slag classified according to length and according to the whiskers of Shape Classification same
Classification grade, the same label can carry out numeralization differentiation, press, it should be apparent that for example, when classifying to eyelid attribute
Double-edged eyelid and interior pair are distinguished according to the distance between eyelid.Same label covers all situations as far as possible, is avoided as far as possible by remaining institute
There is less classification to be classified as same class.The case where certain label classifications are not present, such as no-frame glasses do not need to judge mirror
The thickness of frame, can independent one classification of another column, without the label should be allowed default.
The attribute tags according to an embodiment of the present invention of table 1
When obtaining training dataset, the picture of some possible classifications is more, and the picture of some classifications is especially few, makes in this way
At the imbalance of training dataset.In order to solve the problems, such as to state data nonbalance, it can use the methods of sampling or weighting, may be used also
To generate more samples using existing sample by Data Synthesis/enhancing method, such as generation confrontation network can be sampled
The new data of a batch are generated, but should ensure that the data proportion of generation cannot be too many, the feature for avoiding e-learning from arriving is not
It is the feature of necessary being.The weighting loss value between each classification can also be made approximately equal by the way that reasonable weight is arranged.With
The unbalanced method of upper solution data set is merely exemplary, and can solve training number using different methods according to the actual situation
According to unbalanced problem.
Fig. 3 shows the schematic diagram of mask images according to an embodiment of the invention.As shown in figure 3, in order to mark
Convenient, all properties are labeled in same mask images, and different colours display indicates different people in the mask images of generation
Face attribute.
According to one embodiment of present invention, before generating mask images, human face region can be carried out to original image
It cuts out.For example, firstly, obtaining position and the characteristic point coordinate of face based on characteristic point detection model.Then, it is based on face position
It sets and carries out face normalization with characteristic point coordinate, obtain the consistent facial image of size to cut.It is, for example, possible to use two eyes
Coordinate as the standard ajusted, acquisition human face region is cut out to original image after ajusting: 1, obtaining two eyes shapes
At straight line angle, the rotation of image is carried out according to the opposite direction of the angle, a crucial step is according to rotation after rotation
Formula, the coordinate of the characteristic point after being rotated;2, after ajusting, according to eyes, nose, mouth fore-and-aft distance, estimation is wanted
The height of the face of cutting;3, after determining height, setting depth-width ratio etc., the width of the face to be determined of determination;4, basis
The distance ratio of nose and two eyes transverse directions calculates the specific coordinate of the right and left, rotation angle to positive face, last pantograph
Spending human face region is 128*128.The method that the above progress human face region is cut out is merely exemplary, can also be using at image
Reason software is cut out processing, and this programme is not construed as limiting this.
Then in step S220, it will handle, predicted in the segmentation network of original image input pre-training
Mask images, and based on the segmentation penalty values between mask images generated and the mask images predicted, training segmentation net
Network obtains the first segmentation network.
In the method 200 of execution, the network structure of the segmentation network of embodiment according to the present invention can be constructed in advance.Table 2
Show the subnetwork parameter of the segmentation network of pre-training according to an embodiment of the invention.As shown in table 2, wherein kh,
Kw indicates the height and width of convolution kernel, and padding is Filling power, and stride is step-length, and what expand was indicated is the logical of expansion
Road number, the duplicate structure number that n is indicated.The scale of input picture is triple channel 128*128, defeated in embodiments of the present invention
Entering image is the consistent colorized face images of size.In order to run on the mobile apparatus, so the network structure of selection must
Palpus calculation amount is small, and lightweight process of convolution module is used in this programme, is reduced on its basis.It includes more for dividing network
A lightweight process of convolution module MobilenetV2 Block, process of convolution layer process of convolution layer include convolutional layer, batch normalizing
Change layer and active coating.As shown in table 1, conv indicates that convolutional layer, BN are batch normalization layer, PReLU is the activation letter with parameter
Number, can also be any kinds such as ReLU, tanh, sigmoid, LeakyReLU, it is not limited here.Lightweight process of convolution
Module MobilenetV2 Block separates convolution using depth, and it is one corresponding that depth separates each channel in convolutional coding structure
Different filters, rather than all channels correspond to the same filter, can be realized the separation in channel and region, reduce parameter
Calculation amount, obtained characteristic mass is more preferably.By taking Block1 as an example, convolutional coding structure is separated by 4 depth and is formed.Block2,
Block3 is similar.
The subnetwork parameter of the segmentation network of 2 pre-training of table
According to one embodiment of present invention, can mask images and prediction based on generation mask images, calculate point
Cut penalty values.Then, based on segmentation penalty values, inverse iteration updates the parameter of segmentation network, until meeting the first predetermined condition
Shi Xunlian terminates, and obtains the first segmentation network.Wherein, the first predetermined condition can be segmentation penalty values and no longer decline.In training
In the process, it needs to resolve into the mask images of generation into k mask, each mask represents the response of an attribute.If no
There are some attributes, then the mask is all 0.
Segmentation loss can use the intersection entropy loss of Pixel-level, segmentation penalty values can be calculated based on following formula:
Wherein, H (p, q) is segmentation penalty values, and p is the mask images generated, and q is the mask images of prediction, and p (x) makes a living
At mask images pixel Distribution value, q (x) be prediction mask images pixel Distribution value, x indicate image in pixel
Point, n are the quantity of pixel.
Then in step S230, the characteristic image for identifying at least one attributive character is generated.
According to one embodiment of present invention, original image can be inputted in the segmentation network of pre-training and is handled,
Obtain at least one feature channel figure.Then, by the mask images and at least one feature channel figure of the first segmentation neural network forecast
Point multiplication operation is carried out, at least one characteristic image is obtained.
In some embodiments, characteristic pattern output layer, Lai Shengcheng can be increased after the exposure mask output layer of segmentation network
Characteristic image.As shown in table 2, divide includes segmentation mask output layer Mask and characteristic pattern output layer featuremap in network.
Divide the feature channel figure of conv_4 output 128x16x16 size in network, conv_5 exports the mask artwork of k prediction, each
What mask artwork indicated is the semantic segmentation figure an of attribute.This k mask artwork point is taken on the figure of feature channel respectively, feature is logical
Road figure will become the characteristic image of k 128x16x16.What the characteristic pattern of each 128x16x16 represented is an attributive character.Cause
It is 128x128 for input size, output is 16*16, it is down-sampled there are 8 times between pixel, so when label is super in the small lattice of 8x8
Crossing 20 has value, then the label value of this small lattice is set to 1, is otherwise set to 0.
Then in step S240, each characteristic image is inputted in the sorter network of each pre-training respectively and handled, with
The attribute tags predicted, and the Classification Loss value between the attribute tags based on mark and the attribute tags of prediction, training
Sorter network obtains multiple first sorter networks.
According to one embodiment of present invention, can attribute tags and prediction based on mark attribute tags, calculate point
Class penalty values.Then, it is based on Classification Loss value, inverse iteration updates the parameter of sorter network, until meeting the second predetermined condition
Shi Xunlian terminates, and obtains multiple first sorter networks.Wherein, the second predetermined condition may include: that Classification Loss value no longer declines,
Perhaps the difference of the penalty values for the classification function that front and back iterates to calculate twice reaches less than the first predetermined threshold or the number of iterations
First pre-determined number.
Classification Loss can be lost using SoftMaxWithLoss, can calculate Classification Loss value by following formula:
Wherein, l (y, z) is Classification Loss value, and y is the attribute tags of mark, and z is the attribute tags of prediction, zjIndicate pre-
The label value for j-th of the attribute tags surveyed, zyIndicate that the label value of the attribute tags of mark, m indicate the quantity of attribute tags.
Finally in step s 250, based on segmentation penalty values and Classification Loss value, the first segmentation network of training and multiple the
One sorter network, to utilize the first segmentation network after training and the coupling of multiple first sorter networks to generate the knowledge of face character information
Other model.
In training process, all-network is opened, the weight of segmentation loss and classified weight 1: 10 can be tuned into, combined
At all training.Fig. 4 shows training according to an embodiment of the invention and generates face character information identification model
Flow diagram.As shown in figure 4, the first segmentation network exports the mask artwork and feature channel figure of each attribute.Then pass through
Mask artwork and feature channel figure carry out dot product, obtain characteristic image.The essence of dot product is that mask artwork is bianry image, portion interested
Dividing pixel value is 1, and feature channel figure is multiplied with mask artwork respective pixel, and the pixel of area-of-interest remains unchanged, remaining
Pixel is 0, to show area-of-interest.Obtained each characteristic image is finally inputted into corresponding first classification net respectively
In network, the attribute tags predicted, such as obtaining the attribute tags of hair is [black, long, curly hair].Based on segmentation penalty values
With the weight of Classification Loss value and, inverse iteration updates the parameter of the first segmentation network and multiple first sorter networks, Zhi Daoman
When sufficient third predetermined condition, training terminates, and first after training is divided network and the coupling of multiple first sorter networks generates people
Face attribute information identification model.Wherein, third predetermined condition may include: weight and no longer decline or front and back iteration twice
The difference of the weight sum of calculating reaches the second pre-determined number less than the second predetermined threshold or the number of iterations.
According to the method for the present invention 200, sorter network is fixed first, training segmentation network;Then segmentation network is consolidated
It is fixed, the multiple sorter networks of training;Finally, joint training is carried out to the segmentation network trained and the sorter network trained again,
The segmentation network and multiple sorter networks that final training obtains are mutually coupled, and generate face character information identification model.
Fig. 5 shows the schematic flow of face character information identifying method 500 according to an embodiment of the invention
Figure.As shown in figure 5, obtaining facial image to be identified first in step S510.The facial image of acquisition can also be carried out
Human face region is cut.Then in step S520, facial image is inputted in face character information identification model and is handled, people
Face attribute information identification model includes the segmentation network and sorter network being mutually coupled, wherein facial image is inputted segmentation net
It is handled in network, obtains the mask images and feature channel figure of multiple face characters;Mask images and feature channel figure are carried out a little
Multiplication obtains each attribute region figure.A kind of embodiment according to the present invention, face character information identification model can lead to
The above method 200 is crossed to generate, details are not described herein again.Finally in step S530, each attribute region figure is inputted pair respectively
It is handled in the sorter network answered, obtains the information of each face character in facial image.
According to one embodiment of present invention, face character may include eyelid, glasses, beard, cap, hair.Face
The information of attribute may include one or more of following information: the size of single-edge eyelid, interior pair and double-edged eyelid and/or eyeglass,
Color, the presence or absence of frame, shape, size, length, type, style, the face of dense degree and/or cap of thickness and/or beard
The length of color and/or hair, color, hair style.
According to the solution of the present invention, by carrying out semantic segmentation to facial image, the language of each attribute can effectively be obtained
Adopted information can be improved the accuracy of face character information identification;Image segmentation, Neng Gouman are carried out by using lightweight convolution
Demand of the sufficient mobile terminal to speed and calculation amount greatly improves the efficiency of face character information identification.
A8, the method as described in A6, wherein calculate Classification Loss value by following formula:
Wherein, I (y, z) is Classification Loss value, and y is the attribute tags of mark, and z is the attribute tags of prediction, zjIndicate pre-
The label value for j-th of the attribute tags surveyed, zyIndicate that the label value of the attribute tags of mark, m indicate the quantity of attribute tags.
A9, method as described in a1, wherein the face character includes eyelid, glasses, beard, cap, hair;It is described
Attribute tags include one or more of following label: the label of eyelid, the label of glasses, the label of beard, cap mark
The label of label, hair.
A10, the method as described in A9, wherein
The label of eyelid includes single-edge eyelid, interior double and double-edged eyelid;
The label of glasses includes the presence or absence of eyeglass, size, color, the presence or absence of spectacle-frame, shape, size, thickness;
The label of beard includes the length, type, dense degree of beard;
The label of cap includes the style of cap, color;
The label of hair includes the length, color, hair style of hair.
A11, method as described in a1, wherein the segmentation network includes multiple lightweight process of convolution modules, described light
Quantify process of convolution module and convolution is separated using depth.
A12, method as described in a1, wherein the mask images are generated based on semantic segmentation, and aobvious using different colours
Show to characterize different face characters.
A13, method as described in a1, wherein it is described based on segmentation penalty values and Classification Loss value, train described first point
The step of cutting network and multiple first sorter networks include:
Based on segmentation penalty values and Classification Loss value weight and, inverse iteration update first divide network and multiple first
The parameter of sorter network, when meeting third predetermined condition, training terminate, wherein the third predetermined condition be weight and
No longer the difference of weight sum that iterates to calculate twice of decline or front and back reaches the less than the second predetermined threshold or the number of iterations
Two pre-determined numbers.
A14, method as described in a1, wherein corresponding with original image be labeled with multiple face character information generating
Mask images the step of before, the method also includes:
Human face region is carried out to original image to cut out.
A15, the method as described in A14, wherein it is described to original image carry out human face region cut out the step of include:
Based on characteristic point detection model, position and the characteristic point coordinate of face are obtained;
Face normalization is carried out based on face location and characteristic point coordinate, obtains the consistent facial image of size to cut.
B17, the method as described in B16, wherein
The face character includes eyelid, glasses, beard, cap, hair,
The information of face character includes one or more of following information: single-edge eyelid, interior double and double-edged eyelid and/or mirror
Size, the color of piece, the presence or absence of frame, shape, size, the length of thickness and/or beard, type, dense degree and/or cap
Style, the length of color and/or hair, color, hair style.
B18, the method as described in B16, wherein the method also includes:
Human face region cutting is carried out to the facial image of acquisition, to obtain the consistent facial image of size.
It should be appreciated that in order to simplify the disclosure and help to understand one or more of the various inventive aspects, it is right above
In the description of exemplary embodiment of the present invention, each feature of the invention be grouped together into sometimes single embodiment, figure or
In person's descriptions thereof.However, the disclosed method should not be interpreted as reflecting the following intention: i.e. claimed hair
Bright requirement is than feature more features expressly recited in each claim.More precisely, as the following claims
As book reflects, inventive aspect is all features less than single embodiment disclosed above.Therefore, it then follows specific real
Thus the claims for applying mode are expressly incorporated in the specific embodiment, wherein each claim itself is used as this hair
Bright separate embodiments.
Those skilled in the art should understand that the module of the equipment in example disclosed herein or unit or groups
Part can be arranged in equipment as depicted in this embodiment, or alternatively can be positioned at and the equipment in the example
In different one or more equipment.Module in aforementioned exemplary can be combined into a module or furthermore be segmented into multiple
Submodule.
Those skilled in the art will understand that can be carried out adaptively to the module in the equipment in embodiment
Change and they are arranged in one or more devices different from this embodiment.It can be the module or list in embodiment
Member or component are combined into a module or unit or component, and furthermore they can be divided into multiple submodule or subelement or
Sub-component.Other than such feature and/or at least some of process or unit exclude each other, it can use any
Combination is to all features disclosed in this specification (including adjoint claim, abstract and attached drawing) and so disclosed
All process or units of what method or apparatus are combined.Unless expressly stated otherwise, this specification is (including adjoint power
Benefit require, abstract and attached drawing) disclosed in each feature can carry out generation with an alternative feature that provides the same, equivalent, or similar purpose
It replaces.
In addition, it will be appreciated by those of skill in the art that although some embodiments described herein include other embodiments
In included certain features rather than other feature, but the combination of the feature of different embodiments mean it is of the invention
Within the scope of and form different embodiments.For example, in the following claims, embodiment claimed is appointed
Meaning one of can in any combination mode come using.
Various technologies described herein are realized together in combination with hardware or software or their combination.To the present invention
Method and apparatus or the process and apparatus of the present invention some aspects or part can take insertion tangible media, such as it is soft
The form of program code (instructing) in disk, CD-ROM, hard disk drive or other any machine readable storage mediums,
Wherein when program is loaded into the machine of such as computer etc, and is executed by the machine, the machine becomes to practice this hair
Bright equipment.
In the case where program code executes on programmable computers, mobile terminal generally comprises processor, processor
Readable storage medium (including volatile and non-volatile memory and or memory element), at least one input unit, and extremely
A few output device.Wherein, memory is configured for storage program code;Processor is configured for according to the memory
Instruction in the said program code of middle storage executes method of the present invention.
By way of example and not limitation, computer-readable medium includes computer storage media and communication media.It calculates
Machine readable medium includes computer storage media and communication media.Computer storage medium storage such as computer-readable instruction,
The information such as data structure, program module or other data.Communication media is generally modulated with carrier wave or other transmission mechanisms etc.
Data-signal processed passes to embody computer readable instructions, data structure, program module or other data including any information
Pass medium.Above any combination is also included within the scope of computer-readable medium.
In addition, be described as herein can be by the processor of computer system or by executing by some in the embodiment
The combination of method or method element that other devices of the function are implemented.Therefore, have for implementing the method or method
The processor of the necessary instruction of element forms the device for implementing this method or method element.In addition, Installation practice
Element described in this is the example of following device: the device be used for implement as in order to implement the purpose of the invention element performed by
Function.
As used in this, unless specifically stated, come using ordinal number " first ", " second ", " third " etc.
Description plain objects, which are merely representative of, is related to the different instances of similar object, and is not intended to imply that the object being described in this way must
Must have the time it is upper, spatially, sequence aspect or given sequence in any other manner.
Although the embodiment according to limited quantity describes the present invention, above description, the art are benefited from
It is interior it is clear for the skilled person that in the scope of the present invention thus described, it can be envisaged that other embodiments.Additionally, it should be noted that
Language used in this specification primarily to readable and introduction purpose and select, rather than in order to explain or limit
Determine subject of the present invention and selects.Therefore, without departing from the scope and spirit of the appended claims, for this
Many modifications and changes are obvious for the those of ordinary skill of technical field.For the scope of the present invention, done
Disclosure be it is illustrative and not restrictive, it is intended that the scope of the present invention be defined by the claims appended hereto.
Claims (10)
1. a kind of method for generating face character information identification model, the method is suitable for executing in the terminal, comprising:
Generate the mask images for being labeled with multiple face character information corresponding with original image, and characterization face character information
Multiple attribute tags;
It will be handled in the segmentation network of original image input pre-training, the mask images predicted, and be based on being generated
Mask images and the mask images predicted between segmentation penalty values, training segmentation network obtains the first segmentation network;
Generate the characteristic image for identifying at least one attributive character;
Each characteristic image is inputted in the sorter network of each pre-training respectively and is handled, with the attribute tags predicted, and
Attribute tags based on mark and the Classification Loss value between the attribute tags of prediction, training sorter network, obtain multiple first
Sorter network;And
Based on segmentation penalty values and Classification Loss value, the training first segmentation network and multiple first sorter networks, to utilize
The first segmentation network and the coupling of multiple first sorter networks after training generate face character information identification model.
2. the method for claim 1, wherein it is described based on mask images generated and the mask images predicted it
Between segmentation penalty values, training segmentation network the step of include:
The mask images of mask images and prediction based on generation calculate segmentation penalty values;
Based on the segmentation penalty values, inverse iteration updates the parameter of segmentation network, training when meeting the first predetermined condition
Terminate, obtains the first segmentation network.
3. method according to claim 2, wherein first predetermined condition is that segmentation penalty values no longer decline.
4. method according to claim 2, wherein calculate segmentation penalty values based on following formula:
Wherein, H (p, q) is segmentation penalty values, and p is the mask images generated, and q is the mask images of prediction, and p (x) is to generate
The pixel Distribution value of mask images, q (x) are the pixel Distribution value of the mask images of prediction, and x indicates that the pixel in image, n are
The quantity of pixel.
5. the method for claim 1, wherein described the step of generating the characteristic image for identifying at least one attributive character
Include:
It will be handled in the segmentation network of original image input pre-training, obtain at least one feature channel figure;
The mask images of first segmentation neural network forecast and at least one feature channel figure are subjected to point multiplication operation, obtain at least one
Characteristic image.
6. the method for claim 1, wherein attribute tags based on mark and between the attribute tags of prediction
Classification Loss value, training sorter network the step of include:
The attribute tags of attribute tags and prediction based on mark calculate Classification Loss value;
Based on the Classification Loss value, inverse iteration updates the parameter of sorter network, training when meeting the second predetermined condition
Terminate, obtains multiple first sorter networks.
7. method as claimed in claim 6, wherein second predetermined condition no longer declines for Classification Loss value or front and back
The difference of the penalty values of the classification function iterated to calculate twice reaches first less than the first predetermined threshold or the number of iterations and makes a reservation for
Number.
8. a kind of face character information identifying method, suitable for executing in the terminal, which comprises
Obtain facial image to be identified;
Facial image is inputted in face character information identification model and is handled, the face character information identification model includes
The segmentation network and sorter network being mutually coupled, wherein
Facial image is inputted in segmentation network and is handled, the mask images and feature channel figure of multiple face characters are obtained;
Mask images and feature channel figure are subjected to point multiplication operation, obtain each attribute region figure;
Each attribute region figure is inputted respectively in corresponding sorter network and is handled, each face character in facial image is obtained
Information.
9. a kind of mobile terminal, comprising:
Memory;
One or more processors;
One or more programs, wherein one or more of programs are stored in the memory and are configured as by described one
A or multiple processors execute, and one or more of programs include for executing in -8 the methods according to claim 1
The instruction of either method.
10. a kind of computer readable storage medium for storing one or more programs, one or more of programs include instruction,
Described instruction is when mobile terminal execution, so that the mobile terminal execution appointing in method described in -8 according to claim 1
One method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910686308.9A CN110414428A (en) | 2019-07-26 | 2019-07-26 | A method of generating face character information identification model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910686308.9A CN110414428A (en) | 2019-07-26 | 2019-07-26 | A method of generating face character information identification model |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110414428A true CN110414428A (en) | 2019-11-05 |
Family
ID=68363580
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910686308.9A Pending CN110414428A (en) | 2019-07-26 | 2019-07-26 | A method of generating face character information identification model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110414428A (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110991299A (en) * | 2019-11-27 | 2020-04-10 | 中新国际联合研究院 | Confrontation sample generation method aiming at face recognition system in physical domain |
CN111027474A (en) * | 2019-12-09 | 2020-04-17 | Oppo广东移动通信有限公司 | Face area acquisition method and device, terminal equipment and storage medium |
CN111079659A (en) * | 2019-12-19 | 2020-04-28 | 武汉水象电子科技有限公司 | Face feature point positioning method |
CN111291700A (en) * | 2020-02-20 | 2020-06-16 | 苏州科达科技股份有限公司 | Face attribute identification method, device and equipment and readable storage medium |
CN112163545A (en) * | 2020-10-12 | 2021-01-01 | 北京易华录信息技术股份有限公司 | Head feature extraction method and device, electronic equipment and storage medium |
CN112598007A (en) * | 2021-03-04 | 2021-04-02 | 浙江所托瑞安科技集团有限公司 | Method, device and equipment for screening picture training set and readable storage medium |
CN112785595A (en) * | 2019-11-07 | 2021-05-11 | 北京市商汤科技开发有限公司 | Target attribute detection, neural network training and intelligent driving method and device |
CN112825117A (en) * | 2019-11-20 | 2021-05-21 | 北京眼神智能科技有限公司 | Behavior attribute judgment method, behavior attribute judgment device, behavior attribute judgment medium and behavior attribute judgment equipment based on head features |
CN113076850A (en) * | 2021-03-29 | 2021-07-06 | Oppo广东移动通信有限公司 | Multitask prediction method, multitask prediction device and electronic equipment |
CN113269781A (en) * | 2021-04-21 | 2021-08-17 | 青岛小鸟看看科技有限公司 | Data generation method and device and electronic equipment |
CN114093011A (en) * | 2022-01-12 | 2022-02-25 | 北京新氧科技有限公司 | Hair classification method, device, equipment and storage medium |
WO2023009058A1 (en) * | 2021-07-30 | 2023-02-02 | 脸萌有限公司 | Image attribute classification method and apparatus, electronic device, medium, and program product |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102324036A (en) * | 2011-09-02 | 2012-01-18 | 北京新媒传信科技有限公司 | Obtain the method and apparatus of face complexion area in the image |
CN103914683A (en) * | 2013-12-31 | 2014-07-09 | 闻泰通讯股份有限公司 | Gender identification method and system based on face image |
CN107169455A (en) * | 2017-05-16 | 2017-09-15 | 中山大学 | Face character recognition methods based on depth local feature |
US20170301121A1 (en) * | 2013-05-02 | 2017-10-19 | Emotient, Inc. | Anonymization of facial images |
CN108596839A (en) * | 2018-03-22 | 2018-09-28 | 中山大学 | A kind of human-face cartoon generation method and its device based on deep learning |
CN108596248A (en) * | 2018-04-23 | 2018-09-28 | 上海海洋大学 | A kind of classification of remote-sensing images model based on improvement depth convolutional neural networks |
CN109117760A (en) * | 2018-07-27 | 2019-01-01 | 北京旷视科技有限公司 | Image processing method, device, electronic equipment and computer-readable medium |
CN109145983A (en) * | 2018-08-21 | 2019-01-04 | 电子科技大学 | A kind of real-time scene image, semantic dividing method based on lightweight network |
CN109344693A (en) * | 2018-08-13 | 2019-02-15 | 华南理工大学 | A kind of face multizone fusion expression recognition method based on deep learning |
CN109685740A (en) * | 2018-12-25 | 2019-04-26 | 努比亚技术有限公司 | Method and device, mobile terminal and the computer readable storage medium of face normalization |
CN109829356A (en) * | 2018-12-05 | 2019-05-31 | 科大讯飞股份有限公司 | The training method of neural network and pedestrian's attribute recognition approach neural network based |
CN109858439A (en) * | 2019-01-30 | 2019-06-07 | 北京华捷艾米科技有限公司 | A kind of biopsy method and device based on face |
CN109886273A (en) * | 2019-02-26 | 2019-06-14 | 四川大学华西医院 | A kind of CMR classification of image segmentation system |
-
2019
- 2019-07-26 CN CN201910686308.9A patent/CN110414428A/en active Pending
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102324036A (en) * | 2011-09-02 | 2012-01-18 | 北京新媒传信科技有限公司 | Obtain the method and apparatus of face complexion area in the image |
US20170301121A1 (en) * | 2013-05-02 | 2017-10-19 | Emotient, Inc. | Anonymization of facial images |
CN103914683A (en) * | 2013-12-31 | 2014-07-09 | 闻泰通讯股份有限公司 | Gender identification method and system based on face image |
CN107169455A (en) * | 2017-05-16 | 2017-09-15 | 中山大学 | Face character recognition methods based on depth local feature |
CN108596839A (en) * | 2018-03-22 | 2018-09-28 | 中山大学 | A kind of human-face cartoon generation method and its device based on deep learning |
CN108596248A (en) * | 2018-04-23 | 2018-09-28 | 上海海洋大学 | A kind of classification of remote-sensing images model based on improvement depth convolutional neural networks |
CN109117760A (en) * | 2018-07-27 | 2019-01-01 | 北京旷视科技有限公司 | Image processing method, device, electronic equipment and computer-readable medium |
CN109344693A (en) * | 2018-08-13 | 2019-02-15 | 华南理工大学 | A kind of face multizone fusion expression recognition method based on deep learning |
CN109145983A (en) * | 2018-08-21 | 2019-01-04 | 电子科技大学 | A kind of real-time scene image, semantic dividing method based on lightweight network |
CN109829356A (en) * | 2018-12-05 | 2019-05-31 | 科大讯飞股份有限公司 | The training method of neural network and pedestrian's attribute recognition approach neural network based |
CN109685740A (en) * | 2018-12-25 | 2019-04-26 | 努比亚技术有限公司 | Method and device, mobile terminal and the computer readable storage medium of face normalization |
CN109858439A (en) * | 2019-01-30 | 2019-06-07 | 北京华捷艾米科技有限公司 | A kind of biopsy method and device based on face |
CN109886273A (en) * | 2019-02-26 | 2019-06-14 | 四川大学华西医院 | A kind of CMR classification of image segmentation system |
Non-Patent Citations (2)
Title |
---|
MADBI M.KALAYEH ET AL.: "Improving Facial Attribute Prediction Using Semantic Segmentation", 《PROCESSING OF THE IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION(CVPR)》 * |
欧仁·查尼阿克: "《统计语言学习》", 31 August 2016 * |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112785595B (en) * | 2019-11-07 | 2023-02-28 | 北京市商汤科技开发有限公司 | Target attribute detection, neural network training and intelligent driving method and device |
CN112785595A (en) * | 2019-11-07 | 2021-05-11 | 北京市商汤科技开发有限公司 | Target attribute detection, neural network training and intelligent driving method and device |
CN112825117A (en) * | 2019-11-20 | 2021-05-21 | 北京眼神智能科技有限公司 | Behavior attribute judgment method, behavior attribute judgment device, behavior attribute judgment medium and behavior attribute judgment equipment based on head features |
CN110991299A (en) * | 2019-11-27 | 2020-04-10 | 中新国际联合研究院 | Confrontation sample generation method aiming at face recognition system in physical domain |
CN110991299B (en) * | 2019-11-27 | 2023-03-14 | 中新国际联合研究院 | Confrontation sample generation method aiming at face recognition system in physical domain |
CN111027474A (en) * | 2019-12-09 | 2020-04-17 | Oppo广东移动通信有限公司 | Face area acquisition method and device, terminal equipment and storage medium |
CN111027474B (en) * | 2019-12-09 | 2024-03-15 | Oppo广东移动通信有限公司 | Face region acquisition method and device, terminal equipment and storage medium |
CN111079659A (en) * | 2019-12-19 | 2020-04-28 | 武汉水象电子科技有限公司 | Face feature point positioning method |
CN111291700A (en) * | 2020-02-20 | 2020-06-16 | 苏州科达科技股份有限公司 | Face attribute identification method, device and equipment and readable storage medium |
CN112163545A (en) * | 2020-10-12 | 2021-01-01 | 北京易华录信息技术股份有限公司 | Head feature extraction method and device, electronic equipment and storage medium |
CN112598007B (en) * | 2021-03-04 | 2021-05-18 | 浙江所托瑞安科技集团有限公司 | Method, device and equipment for screening picture training set and readable storage medium |
CN112598007A (en) * | 2021-03-04 | 2021-04-02 | 浙江所托瑞安科技集团有限公司 | Method, device and equipment for screening picture training set and readable storage medium |
CN113076850A (en) * | 2021-03-29 | 2021-07-06 | Oppo广东移动通信有限公司 | Multitask prediction method, multitask prediction device and electronic equipment |
CN113269781A (en) * | 2021-04-21 | 2021-08-17 | 青岛小鸟看看科技有限公司 | Data generation method and device and electronic equipment |
WO2023009058A1 (en) * | 2021-07-30 | 2023-02-02 | 脸萌有限公司 | Image attribute classification method and apparatus, electronic device, medium, and program product |
CN114093011A (en) * | 2022-01-12 | 2022-02-25 | 北京新氧科技有限公司 | Hair classification method, device, equipment and storage medium |
CN114093011B (en) * | 2022-01-12 | 2022-05-06 | 北京新氧科技有限公司 | Hair classification method, device, equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110414428A (en) | A method of generating face character information identification model | |
JP6997309B2 (en) | Image processing system and processing method using deep neural network | |
JP7413400B2 (en) | Skin quality measurement method, skin quality classification method, skin quality measurement device, electronic equipment and storage medium | |
US9864901B2 (en) | Feature detection and masking in images based on color distributions | |
CN108062526A (en) | A kind of estimation method of human posture and mobile terminal | |
US9111375B2 (en) | Evaluation of three-dimensional scenes using two-dimensional representations | |
CN108961279A (en) | Image processing method, device and mobile terminal | |
CN106295533B (en) | A kind of optimization method, device and the camera terminal of self-timer image | |
US10769718B1 (en) | Method, medium, and system for live preview via machine learning models | |
CN110110118A (en) | Dressing recommended method, device, storage medium and mobile terminal | |
CN105405157B (en) | Portrait generation device, Portrait generation method | |
CN108875540A (en) | Image processing method, device and system and storage medium | |
CN106897659A (en) | The recognition methods of blink motion and device | |
KR20160142742A (en) | Device and method for providing makeup mirror | |
CN111008935B (en) | Face image enhancement method, device, system and storage medium | |
CN109117760A (en) | Image processing method, device, electronic equipment and computer-readable medium | |
CN109271930A (en) | Micro- expression recognition method, device and storage medium | |
CN109741338A (en) | A kind of face dividing method, device and equipment | |
JP2019016114A (en) | Image processing device, learning device, focus controlling device, exposure controlling device, image processing method, learning method and program | |
CN113661520A (en) | Modifying the appearance of hair | |
KR20230025906A (en) | Systems and methods for improved facial attribute classification and its use | |
CN110351597A (en) | A kind of method, apparatus and electronic equipment of video clipping | |
WO2019142127A1 (en) | Method and system of creating multiple expression emoticons | |
Lienhard et al. | How to predict the global instantaneous feeling induced by a facial picture? | |
CN117152566A (en) | Classification model training method, model, classification method and product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20191105 |
|
RJ01 | Rejection of invention patent application after publication |