CN109359618A - A kind of image processing method and its device, equipment and storage medium - Google Patents
A kind of image processing method and its device, equipment and storage medium Download PDFInfo
- Publication number
- CN109359618A CN109359618A CN201811281028.1A CN201811281028A CN109359618A CN 109359618 A CN109359618 A CN 109359618A CN 201811281028 A CN201811281028 A CN 201811281028A CN 109359618 A CN109359618 A CN 109359618A
- Authority
- CN
- China
- Prior art keywords
- point
- adjustment
- distance
- subset
- image processing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the present invention provides a kind of image processing method and its device, equipment and storage medium, wherein the described method includes: obtaining Q key point in facial image, wherein Q is the integer greater than 1;Default key point to be processed is filtered out from the Q key point;Target area to be processed in facial image is determined according to the default key point to be processed, includes the forehead region of facial image in the target area;The target area is divided into N number of subregion, N is the integer more than or equal to 2;Stretching is carried out to the pixel in each subregion respectively, the image that obtains that treated.
Description
Technical field
The present invention relates to technical field of image processing, a kind of image processing method and its device, equipment are related to, but are not limited to
And storage medium.
Background technique
With the popularization and application of the development of mobile communication technology, especially mobile network and intelligent terminal, mobile Internet
Field development is swift and violent.For at present, the function of intelligent terminal is not simple already for meeting connecting each other for people, and is become
Highly important mobile personal recreation terminal in people's daily life.
Nowadays camera function has become the essential function of intelligent terminal, and people can record excellent wink anywhere or anytime
Between.And current intelligent terminal all includes front camera mostly, and user is facilitated to carry out self-timer.Liking to be beautiful is the nature of people, greatly
It is more good-looking that family all wants the photo oneself taken, therefore has just expedited the emergence of various U.S. face methods.One good face phase, forehead are wanted
Full, circumference, bright profit.Traditional full algorithm of 2D forehead is mainly by means of human face detection tech and simple deformation algorithm to figure
The forehead of people carries out " filling " operation in piece, to reach face picture " forehead is full " effect.
But there are also significant limitations for the full algorithm of 2D forehead traditional at present.For example, the effect of deformation algorithm very according to
Rely the accuracy in human face detection tech, subtle deviation this may result in " shaping failure ";High-precision, key point is intensive
Face datection model has high time-consuming, and it is unacceptable with live preview function institute to be that U.S. face camera is taken pictures;In addition, people
The three-dimensional multiplicity of forehead, traditional algorithm often only carry out forehead profile to be simply filled with operation, it is difficult to handle different angle, size and
The forehead of shape;And 2D U.S. face is difficult to make the face deformation of " three-dimensional sense ", simple deformation can only simply stretch with
Forehead profile is pushed, three-dimensional full effect is not achieved.
Summary of the invention
In view of this, an embodiment of the present invention is intended to provide a kind of image processing method and its device, equipment and storage medium,
It solves in the full algorithm of forehead in prior art that accuracy rate is low, be easy to cause U.S. face failure and treatment effeciency is low
Technical problem can not only be fitted various three-dimensional foreheads, achieve the effect that three-dimensional U.S. face, so that treated image is more
Add naturally, and also simplifying implementation to greatly improve treatment effeciency.
The technical solution of the embodiment of the present invention is achieved in that
The embodiment of the present invention provides a kind of image processing method, which comprises
Obtain Q key point in facial image, wherein Q is the integer greater than 1;
Default key point to be processed is filtered out from the Q key point;
Target area to be processed in the facial image is determined according to the default key point to be processed, wherein
It include the forehead region of facial image in the target area;
The target area is divided into N number of subregion, wherein N is the integer greater than 2;
Stretching is carried out to the pixel in each subregion respectively, the image that obtains that treated.
In embodiments of the present invention, the default key point to be processed includes forehead region in the facial image
Fisrt feature point set accordingly determines target area to be processed in facial image according to the default key point to be processed
Domain, comprising:
Obtain the facial angle information and adjusting parameter of the facial image;
According to the fisrt feature point set, facial angle information and adjusting parameter determine the forehead region central point,
The first point set on original contour and the second point set on objective contour;
According to target area described in the central point, the second point set and preset second adjustment ratio-dependent.
In embodiments of the present invention, described to be determined according to the fisrt feature point set, facial angle information and adjusting parameter
The second point set on the first point set and objective contour on the central point in the forehead region, original contour, comprising:
The central point of the target area is determined according to the fisrt feature point set and the facial angle information;
Second feature point set is determined according to the central point, the fisrt feature point set and adjusting parameter;
Interpolation, phase are carried out according to preset interpolation algorithm respectively to the fisrt feature point set and the second feature point set
It should obtain first point set and the second point set.
In embodiments of the present invention, described to determine according to the central point, the fisrt feature point set and adjusting parameter
Two feature point sets, comprising:
Determine the first distance between the central point and fisrt feature point;
The first adjustment distance is determined according to the first distance and the adjusting parameter;
The point that the point of fisrt feature described in the distance along the first adjustment direction is the first adjustment distance is determined as accordingly
Second feature point, wherein centered on the first adjustment direction point arrive fisrt feature point direction;
It obtains the fisrt feature and concentrates the corresponding second feature point of each fisrt feature point, obtain second feature point
Collection.
In embodiments of the present invention, it is described to the fisrt feature point set and the second feature point set respectively according to default
Interpolation algorithm carry out interpolation, accordingly obtain first point set and the second point set, comprising:
It is first by the first subset of the fisrt feature point set and the third subset division of corresponding second feature point set
The second subset of the fisrt feature point set and the 4th subset division of corresponding second feature point set are the by target subregion
Two target subregions;
First subset of the first object subregion and the second subset are calculated according to preset interpolation respectively
Method carries out interpolation, obtains corresponding 5th subset and the 6th subset;
The intersection of 5th subset and the 6th subset is determined as the first point set;
The third subset and the 4th subset to the second target subregion are calculated according to preset interpolation respectively
Method carries out interpolation, obtains corresponding 7th subset and the 8th subset;
The intersection of 7th subset and the 8th subset is determined as second subset.
In embodiments of the present invention, described according to the central point, the second point set and preset second adjustment ratio-dependent
The target area, comprising:
Determine the second distance between the central point and i-th of second point of second point concentration, wherein i=1,
2 ..., N+1, (N+1) are the sum of second point;
According to the second distance and the second adjustment ratio-dependent second adjustment distance;
It is that the point of the second adjustment distance is determined as i-th by i-th of second point described in the distance along second adjustment direction
Adjust point, wherein the second adjustment direction is the central point to the direction of i-th of second point;
The target area is determined according to (N+1) a adjustment point.
It is in embodiments of the present invention, described that the target area is divided into N number of subregion, comprising:
The central point, the i-th adjustment point and i+1 adjustment point are sequentially connected and constitute i-th of subregion, wherein i=1,
2 ..., N, (N+1) are the sum of second point.
In embodiments of the present invention, described that stretching is carried out to the pixel in each subregion respectively, it obtains
Treated image, comprising:
Obtain the location information of j-th of pixel in i-th of subregion;
According to the location information of j-th of pixel, central point, it is described first point concentration i-th first point, i-th+
1 first point, i-th of second point of second point concentration, i+1 second point, the i-th adjustment point and i+1 adjustment
Point determines stretching function;
J-th of target position is determined according to the location information of j-th of pixel and the stretching function;
The target pixel value of j-th of pixel is determined according to j-th of target position;
The pixel value of j-th of pixel is updated to the target pixel value, the image that obtains that treated.
In embodiments of the present invention, it is described according to the location information of j-th of pixel, it is central point, first point described
First point of i-th concentrated, i+1 first point, i-th of second point of second point concentration, the i+1 second
Point, the i-th adjustment point and i+1 adjustment point determine stretching function, comprising:
By the 4th line of the central point and j-th of pixel along the central point to j-th of pixel
Direction extended, with described i-th first point, the i+1 first point of line intersect at the first intersection point, with it is described
I-th of second point, the i+1 second point line intersect at the second intersection point, with the 5th line intersect at third friendship
Point, wherein the 5th line is the line of the i-th adjustment point with i+1 adjustment point;
Stretching function is determined according to the 4th distance, the 5th distance and the 6th distance, wherein the 4th distance is institute
The distance between central point and first intersection point are stated, the 5th distance is between the central point and second intersection point
Distance, the 6th distance is the distance between the central point and the third intersection point.
It is in embodiments of the present invention, described that stretching function is determined according to the 4th distance, the 5th distance and the 6th distance,
Include:
Determine the 4th distance and it is described 6th distance between the first ratio, the 5th distance with the described 6th away from
The second ratio between;
The first coordinate is determined according to first ratio and second ratio;
The linear equation of first coordinate and the line of origin is determined as the first piecewise function;
The linear equation of first coordinate and the line of preset second coordinate is determined as the second piecewise function;
Stretching function is determined according to first piecewise function and second piecewise function.
In embodiments of the present invention, the location information and the stretching function according to j-th of pixel
Determine j-th of target position, comprising:
Between j-th of pixel and the central point is determined according to the location information of j-th of pixel
Seven distances;
Determine the third ratio between the 7th distance and the 6th distance;
Using the third ratio as the input of the stretching function, output valve is calculated;
The 8th distance is determined according to the output valve and the 6th distance, wherein the 8th distance is j-th of target
The distance between position and the central point;
J-th of mesh is determined according to the location information of the 8th distance, the central point and j-th of pixel
Cursor position.
In embodiments of the present invention, the target picture that j-th of pixel is determined according to j-th of target position
Element value, comprising:
Coordinate value in response to the target position is integer, and the pixel value of the target position is determined as the jth
The target pixel value of a pixel;
Coordinate value in response to the target position is not integer, determines that the target position is corresponding according to preset algorithm
Pixel value;
The corresponding pixel value in the target position is determined as to the target pixel value of j-th of pixel.
The embodiment of the present invention provides a kind of image processing apparatus, and described image processing unit includes: the first acquisition module, sieve
Modeling block, the first determining module, division module and stretching module, in which:
Described first obtains module, for obtaining Q key point in facial image, wherein Q is the integer greater than 1;
The screening module, for filtering out default key point to be processed from the Q key point;
First determining module, for according to the default key point to be processed determine in the facial image to
The target area of processing, wherein include the forehead region of facial image in the target area;
The division module, for the target area to be divided into N number of subregion, wherein N is the integer greater than 2;
The stretching module is obtained for carrying out stretching to the pixel in each subregion respectively
Treated image.
In embodiments of the present invention, the default key point to be processed includes forehead region in the facial image
Fisrt feature point set, accordingly, the first determining module includes:
First acquisition unit, for obtaining fisrt feature point set, the facial angle letter in forehead region in the facial image
Breath and adjusting parameter;
First determination unit, for according to the determination of the fisrt feature point set, facial angle information and adjusting parameter
The second point set on the first point set and objective contour on the central point in forehead region, original contour;
Second determination unit, for according to the central point, the second point set and preset second adjustment ratio-dependent
Target area.
In embodiments of the present invention, first determination unit, comprising:
First determines subelement, for determining the target according to the fisrt feature point set and the facial angle information
The central point in region;
Second determines subelement, for determining second according to the central point, the fisrt feature point set and adjusting parameter
Feature point set;
Interpolation subelement is used for the fisrt feature point set and the second feature point set respectively according to preset interpolation
Algorithm carries out interpolation, accordingly obtains first point set and the second point set.
In embodiments of the present invention, described second determine that subelement is also used to:
Determine the first distance between the central point and each fisrt feature point;
The first adjustment distance is determined according to the first distance and the adjusting parameter;
The point that the point of fisrt feature described in the distance along the first adjustment direction is the first adjustment distance is determined as accordingly
Second feature point, wherein centered on the first adjustment direction point arrive fisrt feature point direction;
It obtains the fisrt feature and concentrates the corresponding second feature point of each fisrt feature point, obtain second feature point
Collection.
In embodiments of the present invention, the interpolation subelement is also used to:
It is first by the first subset of the fisrt feature point set and the third subset division of corresponding second feature point set
The second subset of the fisrt feature point set and the 4th subset division of corresponding second feature point set are the by target subregion
Two target subregions;
First subset of the first object subregion and the second subset are calculated according to preset interpolation respectively
Method carries out interpolation, obtains corresponding 5th subset and the 6th subset;
The intersection of 5th subset and the 6th subset is determined as the first point set;
The third subset and the 4th subset to the second target subregion are calculated according to preset interpolation respectively
Method carries out interpolation, obtains corresponding 7th subset and the 8th subset;
The intersection of 7th subset and the 8th subset is determined as second subset.
In embodiments of the present invention, second determination unit, comprising:
Third determines subelement, for determining between the central point and i-th of second point of second point concentration
Second distance, wherein i=1,2 ..., N+1, (N+1) are the sum of second point;
4th determines subelement, for according to the second distance and the second adjustment ratio-dependent second adjustment away from
From;
Adjust subelement, for will along second adjustment direction i-th second point described in distance be the second adjustment away from
From point be determined as the i-th adjustment point, wherein the second adjustment direction is the central point to the side of i-th of second point
To;
5th determines subelement, determines the target area for according to (N+1) a adjustment point.
In embodiments of the present invention, the division module, comprising:
Division unit, for the central point, the i-th adjustment point and i+1 adjustment point to be sequentially connected and constitute i-th of sub-district
Domain, wherein i=1,2 ..., N, (N+1) are the sum of second point;
In embodiments of the present invention, the stretching module includes:
Second acquisition unit, for obtaining the location information of j-th of pixel in i-th of subregion;
Third determination unit, for according to the location information of j-th of pixel, central point, first point of concentration
I-th first point, i+1 first point, the second point i-th of second point, the i+1 second point, i-th concentrated
Adjustment point and i+1 adjustment point determine stretching function;
4th determination unit, for being determined according to the location information of j-th of pixel and the stretching function
J-th of target position;
5th determination unit, for determining the object pixel of j-th of pixel according to j-th of target position
Value;
Updating unit, for the pixel value of j-th of pixel to be updated to the target pixel value, after obtaining processing
Image.
In embodiments of the present invention, the third determination unit, comprising:
Extend subelement, for arriving the 4th line of the central point and j-th of pixel along the central point
The direction of j-th of pixel is extended, with described i-th first point, the i+1 first point of line intersects at
First intersection point mutually gives the second intersection point with the line of i-th of second point, the i+1 second point, connects with the described 5th
Line intersects at third intersection point, wherein the 5th line is the line of the i-th adjustment point with i+1 adjustment point;
6th determines subelement, for determining stretching function according to the 4th distance, the 5th distance and the 6th distance,
In, the 4th distance is the distance between the central point and first intersection point, and the 5th distance is the central point
The distance between described second intersection point, the 6th distance is the distance between the central point and the third intersection point.
In embodiments of the present invention, the described 6th determine that subelement is also used to:
Determine the 4th distance and it is described 6th distance between the first ratio, the 5th distance with the described 6th away from
The second ratio between;
The first coordinate is determined according to first ratio and second ratio;
The linear equation of first coordinate and the line of origin is determined as the first piecewise function;
The linear equation of first coordinate and the line of preset second coordinate is determined as the second piecewise function;
Stretching function is determined according to first piecewise function and second piecewise function.
In embodiments of the present invention, the 4th determination unit, comprising:
7th determine subelement, for according to the location information of j-th of pixel determine j-th of pixel with
The 7th distance between the central point;
8th determines subelement, determines the third ratio between the 7th distance and the 6th distance;
Computation subunit, it is described for being calculated using the third ratio as the input of the stretching function
The corresponding output valve of third ratio;
9th determines subelement, for determining the 8th distance according to the output valve and the 6th distance, wherein described
8th distance is the distance between j-th of target position and described central point;
Tenth determines subelement, for the position according to the 8th distance, the central point and j-th of pixel
Information determines j-th of target position.
In embodiments of the present invention, the 5th determination unit, comprising:
11st determines subelement, is integer for the coordinate value in response to the target position, by the target position
Pixel value be determined as the target pixel value of j-th of pixel;
12nd determines subelement, for not being integer in response to the coordinate value of the target position, according to preset algorithm
Determine the corresponding pixel value in the target position;The corresponding pixel value in the target position is determined as j-th of pixel
Target pixel value.
The embodiment of the present invention provides a kind of image processing equipment, and described image processing equipment includes at least: memory, communication
Bus and processor, in which:
The memory, for storing image processing program;
The communication bus, for realizing the connection communication between processor and memory;
The processor, for obtaining the image processing program stored in memory by the communication bus, to execute
The step of image processing method as described above.
The embodiment of the present invention provides a kind of storage medium, and image processing program, the figure are stored on the storage medium
The step of image processing method as described above is realized when being executed by processor as processing routine.
The embodiment of the present invention provides a kind of image processing method and its device, equipment and storage medium, wherein obtains first
Q key point in facial image, wherein Q is the integer greater than 1;Then filtered out from the Q key point default to
The key point of processing, this makes it possible to the calculation amounts of less image procossing;And then further according to the default key to be processed
Point determines target area to be processed in facial image, includes the forehead region of facial image in the target area;And by institute
It states target area and is divided into N number of subregion, N is the integer more than or equal to 2;Finally respectively to the pixel in each subregion
Point carries out stretching, the image that obtains that treated;In this way, leading to too small amount of key point to be processed determines target area,
And target area is divided into multiple subregions, then stretching is individually carried out to each subregion, it can not only be fitted various
The three-dimensional forehead of various kinds achievees the effect that three-dimensional U.S. face, so that treated image is more naturally, and also simplify realization side
Formula is to greatly improve treatment effeciency.
Detailed description of the invention
Figure 1A is the application scenarios schematic diagram of image processing method of the embodiment of the present invention;
Figure 1B is the schematic diagram of the another application scenarios of the embodiment of the present invention;
Fig. 2 is the implementation process schematic diagram of image processing method of the embodiment of the present invention;
Fig. 3 A is the implementation process schematic diagram of image processing method of the embodiment of the present invention;
Fig. 3 B is the schematic diagram that the embodiment of the present invention carries out tri patch when forehead plentiful adjustment;
Fig. 3 C is the schematic diagram that the embodiment of the present invention carries out tri patch when forehead reduces adjustment;
Fig. 3 D is the coordinate schematic diagram that the embodiment of the present invention carries out piecewise function when forehead plentiful adjustment;
Fig. 3 E is the coordinate schematic diagram that the embodiment of the present invention carries out piecewise function when forehead reduces adjustment;
Fig. 4 is the implementation process schematic diagram of image processing method of the embodiment of the present invention;
Fig. 5 is the schematic diagram of the fitting broken line of the original forehead of the embodiment of the present invention and target forehead profile;
Fig. 6 is the schematic diagram that the embodiment of the present invention is split as forehead region multiple continuous tri patch;
Fig. 7 is the comparison diagram that the embodiment of the present invention carries out tri patch stretching when forehead full adjustment;
Fig. 8 is the adjustment effect comparison diagram that the embodiment of the present invention carries out facial image when forehead full adjustment;
Fig. 9 A is the comparison diagram that the embodiment of the present invention carries out tri patch stretching when the adjustment of forehead area reduction;
Fig. 9 B is the adjustment effect comparison diagram that the embodiment of the present invention carries out facial image when the adjustment of forehead area reduction;
Figure 10 is the composed structure schematic diagram of image processing apparatus of the embodiment of the present invention;
Figure 11 is the composed structure schematic diagram of image processing equipment of the embodiment of the present invention.
Specific embodiment
In order to make the object, technical scheme and advantages of the embodiment of the invention clearer, below in conjunction with the embodiment of the present invention
In attached drawing, the specific technical solution of invention is described in further detail.The following examples are intended to illustrate the invention, but does not have to
To limit the scope of the invention.
Figure 1A is the application scenarios schematic diagram of image processing method of the embodiment of the present invention, as shown in Figure 1A, the application scenarios
In include terminal 101 and server 102, terminal 101 and server 102 establish communication connection by network and go forward side by side row information friendship
Mutually.Terminal 101 obtains image 1011 to be processed first, includes facial image region in image 1011, then will be to be processed
Image 1011 is sent to server 102, and the image processing method provided according to embodiments of the present invention by server 102 is to be processed
Image 1011 handled, the image 1012 that obtains that treated, then will treated that image 1012 is sent to terminal, by terminal
101 output display treated images 1012.Wherein, image to be processed can be adopted in real time by the image collecting device of terminal
Collect image, can also be the image for being stored in terminal local.
Figure 1B is the schematic diagram of the another application scenarios of the embodiment of the present invention, only includes in application scenarios as shown in Figure 1B
Terminal 111, terminal 111 obtain image 1111 to be processed first, include facial image region in image 1111, then terminal
111 image processing methods provided according to embodiments of the present invention handle image 1111 to be processed, obtain that treated
Image 1112, finally again by terminal output display treated image 1112.
The application scenarios schematic diagram in conjunction with shown in Figure 1A and Figure 1B, below to image processing method and image processing apparatus,
Each embodiment of equipment is illustrated.
The present embodiment provides a kind of image processing method, Fig. 2 is the implementation process of image processing method of the embodiment of the present invention
Schematic diagram, as shown in Fig. 2, this method comprises:
Step S201 obtains Q key point in facial image.
Here, step S201 can be by terminal realization, can also be by server realization.In the present embodiment, Q
For the integer greater than 1.Facial image can be the image including forehead region shot under different angle, for example can be people
The facial image of face face camera lens, can be with lateral face to the facial image of camera lens.
When step S201 is realized by terminal, facial image can be to be adopted in real time using the image collecting device in terminal
The image collected, such as can be preview image when taking pictures using camera applications, it can also be the face of terminal local storage
Image, facial image can be the image shot by camera applications at this time, can also be the face downloaded from network
Image, naturally it is also possible to be the facial image that other terminals are sent to the terminal, when the face figure that user stores terminal local
When as being unsatisfied with, it can use image processing method provided in an embodiment of the present invention and being adjusted.
When step S201 is realized by server, facial image can be terminal and be sent to server, at this time in order to
Guarantee the good experience of taking pictures of user, facial image is generally the image being locally stored in terminal.
Step S201 can be the human face region identified in facial image by face recognition technology when realizing
Characteristic point namely Q key point.In the present embodiment, the characteristic point that forehead region is included at least in Q key point, can be with
Characteristic point including regions such as eye, nose, mouth, chins.
Step S202 filters out default key point to be processed from Q key point.
Here, step S202 can be by terminal realization, can also be by server realization.
In the present embodiment, the characteristic point including forehead region in a key point to be processed, therefore, step S202 are preset
When realizing, the characteristic point that forehead region is filtered out from Q key point can be.In this way, due to being screened from multiple key points
A small amount of key point to be processed out, to reduce the calculation amount of image procossing, and then can be improved image processing efficiency.
Step S203 determines target area to be processed in the facial image according to the default key point to be processed
Domain.
Here, step S203 can be by terminal realization, can also be by server realization.In target area at least
Forehead region including facial image.
Due to including the characteristic point in forehead region in the default key point to be processed, step S203 when realizing,
Can with when first the characteristic point in forehead region is carried out carrying out interpolation obtaining more key points (the first point set), and according to adjustment
Ratio-dependent adjusts distance, then is adjusted more key points that interpolation obtains are carried out according to adjustment distance, is adjusted
Multiple key points (the second point set) afterwards, and multiple key points adjusted are sequentially connected, determine target area.
Target area is divided into N number of subregion by step S204.
Here, step S204 can be by terminal realization, can also be by server realization.
In a kind of enforceable mode, subregion can be triangular shape.It include multiple passes on the profile of target area
Key point, when target area is divided into N number of subregion, can be will be on the central point of target area and the profile of target area
Each key point (adjustment point) be sequentially connected to obtain N number of subregion.
In the present embodiment, N is the integer greater than 2;Embedded the first sub- triangular facet with same apex angle of each subregion
Piece and the second sub- tri patch, wherein the bottom edge of the first sub- tri patch is two neighboring pass on the original contour in forehead region
The line of key point (first point), the bottom edge of the second sub- tri patch are two neighboring key point on the objective contour in forehead region
The line of (second point), two ends on the bottom edge of two endpoints and the second sub- tri patch on the bottom edge of the first sub- tri patch
Point is on tri patch by the point on two bevel edges on vertex.
Step S205 carries out stretching to the pixel in each subregion respectively, the image that obtains that treated.
Here, step S205 can be by terminal realization, can also be by server realization.
Pixel progress stretching in subregion may be considered, the pixel value of pixel is replaced with into another picture
The pixel value of vegetarian refreshments, wherein another pixel mentioned here can be the central point along target area to the side of the pixel
The pixel that certain distance obtains is stretched out to by pixel, can also be pixel along the direction of the pixel to central point
The point pixel that retraction certain distance obtains inwards.It can be realized by carrying out stretching to the pixel in subregion to volume
The stretching of head region profile or retraction, to realize the adjustment in forehead region.
Step S205 can be and be clicked through using preset stretching algorithm to the pixel in each subregion when realizing
Row stretching, in the present embodiment, preset stretching algorithm are linear function, therefore have linear computation complexity,
Efficiency faster, is adapted to camera live preview function to the U.S. efficient demand of face algorithm.
When step S205 is realized by terminal, after step S205, this method further include: after exporting and showing processing
Image;When step S205 is realized by server, after step S205, this method further include: after server will be handled
Image be sent to terminal;Terminal shows and exports the image that receives that treated.
In image processing method provided in an embodiment of the present invention, Q key point in facial image is obtained first, then
Default key point to be processed is filtered out from the Q key point, this makes it possible to the calculation amounts for reducing image procossing;Into
And target area to be processed in facial image is determined further according to the default key point to be processed, in the target area
Forehead region including facial image;And the target area is divided into N number of subregion, N is the integer more than or equal to 2;Most
Stretching is carried out to the pixel in each subregion respectively afterwards, the image that obtains that treated;In this way, leading to too small amount of
Key point to be processed determines target area, and target area is divided into multiple subregions, can not only be fitted various each
The three-dimensional forehead of sample achievees the effect that three-dimensional U.S. face, so that treated, image is more natural;In addition, to each subregion
It individually carries out stretching and uses the stretching function with linear computation complexity, implementation can be simplified, and
And computational efficiency is high, to greatly improve treatment effeciency.
Based on the above embodiments, the embodiment of the present invention provides a kind of image processing method again, is applied to terminal, and Fig. 3 A is
The implementation process schematic diagram of image processing method of the embodiment of the present invention, as shown in Figure 3A, this method comprises:
Step S301, terminal obtain Q key point in facial image.
Here, Q is the integer greater than 1.Facial image can be to be collected in real time using the image collecting device in terminal
Image, such as can be preview image when taking pictures using camera applications, can also be the facial image of terminal local storage,
Facial image can be the image shot by camera applications at this time, can also be the facial image downloaded from network,
It can certainly be the facial image that other terminals are sent to the terminal, when the facial image that user stores terminal local is discontented
When meaning, it can use image processing method provided in an embodiment of the present invention and be adjusted.
Step S301 can be the human face region identified in facial image by face recognition technology when realizing
Characteristic point namely Q key point.In the present embodiment, the characteristic point that forehead region is included at least in Q key point, can be with
Characteristic point including regions such as eye, nose, mouth, chins.
Step S302, terminal filter out default key point to be processed from Q key point.
Here, in the present embodiment, it includes each on the original contour of forehead region for presetting in a key point to be processed
Characteristic point, therefore, step S202 can be from the original contour for filtering out forehead region in Q key point when realizing
Characteristic point.In this way, due to filtering out a small amount of key point to be processed from multiple key points, to reduce the meter of image procossing
Calculation amount, and then can be improved image processing efficiency.
Step S303, terminal obtain the facial angle information and adjusting parameter of facial image to be processed.
Here, step S303 can be facial image inputting a preset Face datection mould during realization
Type, to obtain the facial angle information in facial image, wherein when face is face camera lens, facial angle information is 0 degree,
When face is biased to 30 degree of the left side, facial angle information is 30 degree, when face is biased to 30 degree of the right, facial angle information for-
30 degree.
Adjusting parameter can be the real number between 0 to 1, can also be the real number between -1 to 0, wherein adjusting parameter
Absolute value is bigger, then adjusts dynamics or adjustment amount is bigger.
When adjusting parameter be positive number when, be to extend to the outside forehead region, adjustment it is fuller, when adjusting parameter is negative
It is by forehead region to contract when number.
Step S304, terminal determine in forehead region according to fisrt feature point set, facial angle information and adjusting parameter
The second point set on the first point set and objective contour on heart point, original contour.
Here, step S304 can be realized by following steps:
Step S3041 determines the central point of target area according to fisrt feature point set and facial angle information.
Here, step S3041 realize when be utilized fisrt feature point set location information and facial angle information it is true
Set the goal the central point in region.
Step S3042 determines second feature point set according to central point, fisrt feature point set and adjusting parameter.
Here, step S3042 can be during realization and be sequentially connected central point and each fisrt feature point,
And determine the distance between central point and each fisrt feature point, then an adjustment ratio is determined further according to adjusting parameter
Example, and then determine adjustment distance, then the line of central point and fisrt feature point is adjusted according to the adjustment distance, it determines
Second feature point.Second feature point set may be considered the set of the characteristic point on the objective contour in forehead region.
Step S3043 carries out interpolation according to preset interpolation algorithm respectively to fisrt feature point set and second feature point set,
Accordingly obtain the first point set on original contour and the second point set on objective contour.
Here, due to image processing method provided in an embodiment of the present invention be for forehead plentiful degree and/or
What the width at left and right sides of forehead was adjusted, therefore institute's fisrt feature point set and second feature point set are being inserted according to preset
When value-based algorithm carries out interpolation, the characteristic point and second feature point set that the fisrt feature point at left and right sides of forehead region is concentrated can be
In characteristic point carry out interpolation respectively.Interpolation can be carried out using Catmull-Rom polygon approach method when realizing.
Step S305, terminal is according in central point, the second point set and preset second adjustment ratio-dependent facial image
Target area.
Here, the implementation method of step S305 is similar with step S3042, and central point and each second point are sequentially connected,
And determine the distance between central point and each second point, then further according to second adjustment ratio-dependent second adjustment distance,
The line of central point and second point is adjusted according to the second adjustment distance finally, adjustment point is determined, by each adjustment
Point is sequentially connected, and determines target area.
It in embodiments of the present invention, include the original contour and objective contour in forehead region in target area.It needs
It is bright, when the second adjustment ratio used when the second adjustment ratio used when reducing forehead region is than plentiful forehead region
Greatly, to guarantee the original contour and objective contour in the target area determined including forehead region.
Target area is divided into N number of subregion by step S306, terminal.
Here, N is integer greater than 2, and subregion is the tri patch in other embodiments, therefore this step can also be with
It is interpreted as target area being divided into N number of continuous tri patch, embedded the with same apex angle of each tri patch
One sub- tri patch and the second sub- tri patch, wherein the bottom edge of the first sub- tri patch is on the original contour in forehead region
Two neighboring first point of line, the bottom edge of the second sub- tri patch are two neighboring second point on the objective contour in forehead region
Line.
Step S306 can be when realizing is sequentially connected composition i-th for central point, the i-th adjustment point and i+1 adjustment point
Sub-regions are to obtain N number of subregion, wherein i=1,2 ..., N+1, (N+1) are the sum of adjustment point.
Step S307, terminal obtain the location information of j-th of pixel in i-th of subregion.
Here, j=1,2 ..., M, wherein M is the pixel sum in i-th of subregion.The position of j-th of pixel
Information can be the coordinate of j-th of pixel.
Step S308, terminal is according to the location information of j-th of pixel, central point, i-th the of first point of concentration
A bit, i-th of second point, i+1 second point, the i-th adjustment point and the i+1 that first point of i+1, the second point are concentrated
Adjustment point determines stretching function.
Here, i-th first point and i+1s first point be the first sub- tri patch bottom edge endpoint, i-th the
2 points and i+1 second point are the endpoint on the bottom edge of the second sub- tri patch.
One stretching function will be determined according to the position of the pixel for each pixel, pass through the flexible change
The pixel value of pixel on original contour can be mapped as the pixel value of the pixel on objective contour by exchange the letters number, and should
Stretching function is linear function, has linear complexity, therefore treatment effeciency is high.
Step S309, terminal determine j-th of target position according to the location information and stretching function of j-th pixel
It sets.
Here, step S309 is during realization, can first according to the location information of j-th of pixel determine one it is defeated
Enter value, which is input in stretching function, the corresponding output valve of the input value is obtained, then according to the output valve
Determine corresponding j-th of the target position of j-th of pixel.
Step S310, terminal determine the target pixel value of j-th of pixel according to j-th of target position.
Here, when the coordinate value of j-th of target position is integer, the corresponding pixel value in j-th of target position is determined
For target pixel value;When the coordinate value of j-th of target position is floating number, j-th of target can be determined according to preset algorithm
The corresponding pixel value in position can use bilinear interpolation method and determine the corresponding picture in j-th of target position during realization
Element value.
The pixel value of j-th of pixel is updated to target pixel value by step S311, terminal, the image that obtains that treated.
In image processing method provided in an embodiment of the present invention, firstly, terminal obtains Q key in facial image
Then point filters out default key point to be processed, in this way according only to a small amount of pass filtered out from the Q key point
Key point is handled, it will be able to reduce the calculation amount of image procossing;Then it based on default key point to be processed and obtains again
The facial angle information and adjusting parameter got determine the central point in forehead region, the first point set on original contour and target wheel
The second point set on exterior feature, and then according to the mesh in central point, the second point set and preset second adjustment ratio-dependent facial image
Region is marked, then target area is divided into N number of continuous subregion, for example, target area is divided into N number of tri patch, and
The location information of each pixel in each tri patch is obtained, then the location information further according to each pixel, center
Point determines a stretching function with, the second point of second point concentration at first point of first point of concentration, and then terminal is further according to three
The location information of each pixel and stretching function determine the target pixel value of each pixel in edged surface piece, finally by
The pixel value of j pixel is updated to target pixel value, the image that obtains that treated, in this way, due to that can pass through tri patch
It is fitted arbitrary polygon, therefore by implementing image processing method provided in this embodiment, various solids can be fitted
Forehead achievees the effect that three-dimensional U.S. face, so that treated image is more naturally, and the stretching function in the present embodiment
For linear function, there is linear complexity, computation complexity, to greatly improve image processing efficiency.
In other embodiments, step S3042 can be realized by following steps:
Step 21, the first distance between central point and fisrt feature point is determined.
Step 22, the first adjustment distance is determined according to first distance and adjusting parameter.
Here, step 22 realize when, can be and the first adjustment ratio is determined according to adjusting parameter first, then further according to
First distance and the first adjustment ratio-dependent the first adjustment distance.
Since adjusting parameter can be positive value, or negative value, thus determine that the first adjustment distance also correspondingly
For positive value or negative value.
Step 23, will be determined as accordingly along the first adjustment direction apart from the point that fisrt feature point is the first adjustment distance
Second feature point.
Here, the direction for arriving fisrt feature point is put centered on the first adjustment direction.When the first adjustment distance is positive value, step
Rapid 23, when realizing, are that fisrt feature point is extended the adjustment distance along the first adjustment direction, obtain second feature point;Assuming that the
One apart from being 10, and the first adjustment distance is 2, then being exactly that fisrt feature point will be extended distance 2 and obtained along the first adjustment direction
To second feature point, that is to say, that second feature point is by distance center point is 12 along the first adjustment direction point.
When the first adjustment distance is negative value, step 23 is by fisrt feature point along the first adjustment direction phase when realizing
The absolute value of anti-direction retraction the first adjustment distance, obtains second feature point;Assuming that first distance is 10, the first adjustment distance
It is -2, then being exactly that the point along the contrary direction of the first adjustment at a distance from fisrt feature point for 2 is determined as second
Characteristic point, that is to say, that second feature point is by distance center point is 8 along the first adjustment direction point.
Step 24, it obtains the fisrt feature and concentrates the corresponding second feature point of each fisrt feature point, obtain the
Two feature point sets.
In other embodiments, step S3043 can be realized by following steps:
Step 241, the first subset of the fisrt feature point set and the third subset of corresponding second feature point set are drawn
It is divided into first object subregion, by the 4th subset of the second subset of the fisrt feature point set and corresponding second feature point set
It is divided into the second target subregion.
Here, the first subset in first object subregion including fisrt feature point set and the third of second feature point set
Collect, includes the second subset of fisrt feature point set and the 4th subset of second feature point set in the second target subregion.
It should be noted that first object subregion and the second target subregion are the region of left and right two in forehead region.
Step 242, interpolation is carried out according to preset interpolation algorithm respectively to the first subset and second subset, obtained corresponding
5th subset and the 6th subset.
Step 243, the intersection of the 5th subset and the 6th subset is determined as the first point set.
Step 244, interpolation is carried out according to preset interpolation algorithm respectively to third subset and the 4th subset, obtained corresponding
7th subset and the 8th subset.
Step 245, the intersection of the 7th subset and the 8th subset is determined as second subset.
In the embodiment where step 241 to step 245, being will be in the sub- target area in left and right two in target area
Feature point set carry out interpolation arithmetic respectively, that is to say, that in embodiments of the present invention, there is no to forehead region top edge
Profile carries out interpolation arithmetic, i.e., is not changed to the hair line of user, so that more aesthetically pleasing to effect of the forehead after flexible.
In other embodiments, step S305 can be realized by following steps:
Step 31, the second distance between central point and i-th of second point of second point concentration is determined.
Here, i=1,2 ..., N+1, wherein (N+1) is the sum of second point, due to the sum of second point and first point
Sum and adjustment point sum it is all identical, so first point sum also be N+1, adjust a little sum also be N+1.
Step 32, according to second distance and second adjustment ratio-dependent second adjustment distance.
Here, second adjustment ratio is positive value.Second adjustment ratio is typically less than 1 real number, such as second adjustment ratio
Example can be 0.15.
Step 33, by along second adjustment direction i-th of second point of distance be second adjustment distance point be determined as the i-th tune
Integral point.
Here, to the direction of i-th of second point centered on second adjustment direction o'clock;I-th first point in central point and i-th
On the second line for adjusting point.
Step 34, target area is determined according to (N+1) a adjustment point.
Here it is possible to which adjustment point adjacent in (N+1) a adjustment point is sequentially connected, target area is determined.
In other embodiments, step S308 can be realized by following steps:
Step 41, the 4th line of central point and j-th of pixel is carried out along central point to the direction of j-th of pixel
Extend, with i-th first point, i+1 first point of line intersect at the first intersection point, with i-th of second point, the i+1
The line of a second point intersects at the second intersection point, intersects at third intersection point with the 5th line.
Here, the 5th line is the line of the i-th adjustment point and i+1 adjustment point namely the bottom edge of tri patch.Fig. 3 B is
The embodiment of the present invention carries out the schematic diagram of tri patch when forehead plentiful adjustment, and as shown in Figure 3B, A, B, C are respectively tri patch
Three vertex, A, D, E be the first sub- tri patch three vertex, A, F, G be the second sub- tri patch three vertex, three
Edged surface piece ABC, the first sub- tri patch ADE, the second sub- tri patch AFG share an apex angle.BC is the 5th line, DE the
The bottom edge of one sub- tri patch, and DE is a part of forehead region original contour, FG is the bottom edge of the second sub- tri patch, and
FG is a part of forehead regional aim profile.Assuming that j-th of pixel is point P, the first intersection point is point I at this time, and second hands over
Point is point J, and third intersection point is point K.
Step 42, stretching function is determined according to the 4th distance, the 5th distance and the 6th distance.
Here, point is the distance between with first intersection point centered on the 4th distance, i.e. the distance between AI, during the 5th distance is
The distance between heart point and the second intersection point, i.e. the distance between AJ, the 6th puts the distance between third intersection point centered on,
That is the distance between AK.
Fig. 3 C is the schematic diagram that the embodiment of the present invention carries out tri patch when forehead reduces adjustment, as shown in Figure 3 C, A, B, C
Respectively three vertex of tri patch, A, D, E are three vertex of the first sub- tri patch, and A, F, G are the second sub- triangular facet
Three vertex of piece, tri patch ABC, the first sub- tri patch ADE, the second sub- tri patch AFG share an apex angle.BC is
5th line, DE is the bottom edge of the first sub- tri patch, and DE is a part of forehead region original contour, and FG is the second son three
The bottom edge of edged surface piece, and FG is a part of forehead regional aim profile.Assuming that j-th of pixel is point P, first hand at this time
Point is point I, and the second intersection point is point J, and third intersection point is point K.
In other embodiments, step 42 can be realized by following steps:
Step 421, it determines between the first ratio, the 5th distance and the 6th distance between the 4th distance and the 6th distance
Second ratio.
Step 422, the first coordinate is determined according to the first ratio and the second ratio.
Here, the abscissa of the first coordinate is the second ratio, and the ordinate of the first coordinate is the first ratio.
Step 423, the linear equation of the first coordinate and the line of origin is determined as the first piecewise function.
Here, the first piecewise function can use formula (1-1) expression:
Wherein, in formula (1-1), DIS_AI is the 4th distance, and DIS_AJ is the 5th distance.
Step 424, the linear equation of the first coordinate and the line of preset second coordinate is determined as the second piecewise function.
Here, preset second coordinate can be (1,1), and the second piecewise function can be indicated with formula (1-2):
In formula (1-2), DIS_AK is the 6th distance.
Step 425, stretching function is determined according to the first piecewise function and the second piecewise function.
Here, stretching function can be indicated with formula (1-3):
Wherein, in formula (1-3), DIS_AK is the 6th distance.
In the full adjustment of progress forehead, since second point is on central point and first point of extended line, that is to say, that the
The area of two sub- tri patch is greater than the first sub- tri patch, and the 5th distance is greater than the 4th distance, at this time the coordinate of piecewise function
Schematic diagram is as shown in Figure 3 C, wherein 301 be point represented by the first coordinate, and 302 be point represented by the second coordinate.
When carrying out forehead diminution adjustment, since second point is between central point and first point, that is to say, that the second son three
The area of edged surface piece is less than the first sub- tri patch, and the 5th distance is less than the 4th distance, the coordinate schematic diagram of piecewise function at this time
As shown in FIGURE 3 E, wherein 303 be point represented by the first coordinate, and 302 be point represented by the second coordinate.
In other embodiments, step S309 can be realized by following steps:
Step 51, according to the location information of j-th of pixel determine between j-th of pixel and central point the 7th away from
From.
Here, suppose that j-th of pixel, between central point A and the first intersection I, the 7th distance is indicated with DIS_AP.
Step 52, the third ratio between the 7th distance and the 6th distance is determined.
Here, third ratio is
Step 53, using third ratio as the input of stretching function, output valve is obtained.
Here, output valve
Step 54, the 8th distance is determined according to output valve and the 6th distance.
Here, the 8th distance is the distance between j-th of target position and central point, it is assumed that j-th of target position is P ',
8th distance
Step 55, j-th of target position is determined according to the location information of the 8th distance, central point and j-th of pixel.
Here, it is known that the coordinate of the length of AP ', central point A, and the coordinate of point P being located on AP ' can be determined
The coordinate of j-th of target position P '.
In other embodiments, step S310 can be realized by following steps:
Step 61, the coordinate value in response to target position is integer, and the pixel value of target position is determined as j-th of pixel
The target pixel value of point;
Step 62, the coordinate value in response to target position is not integer, determines that target position is corresponding according to preset algorithm
Pixel value;The corresponding pixel value in target position is determined as to the target pixel value of j-th of pixel.
Here, step 62 can determine the corresponding pixel value in target position when realizing according to bilinear interpolation method,
In other embodiments, the pixel of four pixel nearest with target position can also be determined according to the coordinate value of target position
Value, then mean value computation is carried out to four pixel values, using obtained mean value as the corresponding pixel value in j-th of target position.
In other embodiments, step 42 can also be realized by following steps:
Step 426, third coordinate is determined according to the 4th distance and the 5th distance;
Step 427,4-coordinate is determined according to the 6th distance;
Step 428, the linear equation of third coordinate and the line of origin is determined as third piecewise function;
Step 429, the linear equation of third coordinate and the line of 4-coordinate is determined as the 4th piecewise function.
Step 430, stretching function is determined according to third piecewise function and the 4th piecewise function.
Here, in the embodiment where step 426 to step 430, the table of third piecewise function and the first piecewise function
Be up to formula it is identical, the expression formula of the 4th piecewise function and the second piecewise function be it is identical, what is determined in step 430 is flexible
Transforming function transformation function is also stretching function identical, that only step 430 obtains with the stretching function determined in step 425
In stretching function obtained in step 425, the value range of x is different, the stretching function that step 430 obtains
It can be indicated with formula (1-4):
Correspondingly, when the method provided using step 426 to step 430 determines stretching function, step S309 can
To be realized by following steps:
Step 71, according to the location information of j-th of pixel determine between j-th of pixel and central point the 7th away from
From.
Here, suppose that j-th of pixel, between central point A and the first intersection I, the 7th distance is indicated with DIS_AP.
Step 72, the input by the 7th distance as stretching function, is calculated output valve.
Here, output valve is the distance between j-th of target position and central point, it is assumed that and j-th of target position is P ',
8th distance
Step 73, j-th of target position is determined according to the location information of the 8th distance, central point and j-th of pixel.
Here, it is known that the coordinate of the length of AP ', central point A, and the coordinate of point P being located on AP ' can be determined
The coordinate of j-th of target position P '.
Based on the above embodiments, the embodiment of the present invention provides a kind of image processing method again, and Fig. 4 is the embodiment of the present invention
The implementation process schematic diagram of image processing method, as shown in figure 4, method includes the following steps:
Step S401, input face picture and " forehead is full " dynamics parameter.
Here, " forehead is full " dynamics parameter reflects the plentiful degree of forehead after image procossing, " forehead is full " dynamics ginseng
Number is bigger, and finally obtained forehead is more plentiful.
Step S402 carries out Face datection to face picture, exports forehead key point and POSE face information.
Here, Face datection is carried out to face picture using preset Face datection model, forehead region can be obtained
Key point and POSE facial angle information.Wherein, when face face camera lens, facial angle information is 0 degree, to the left to be positive
Angle, the angle to the right being negative.
Step S403 is fitted forehead profile using Catmull-Rom polygon approach algorithm and the forehead key point of input.
Step S404 determines " forehead " objective contour by incoming " forehead is full " dynamics parameter and forehead contour line,
Using the objective contour of the POSE attribute amendment all angles of Face datection, the full center of forehead is determined using forehead key point.
Here, the center-filled direction of " forehead is full " is determined using POSE angle information and forehead key point, and then really
One filling center of fixed " forehead is full ";Then " forehead is full " filling center and forehead key point are sequentially connected and is extended,
The point on extended line is taken according to a certain percentage according to " forehead is full " dynamics parameter to determine " forehead is full " forehead key point afterwards
Position;Finally using Catmull-Rom polygon approach method and forehead it is full " afterwards forehead key point carry out interpolation obtain more
Point on more forehead profiles connects the fitting broken line of these point composition volume target forehead profiles.
Forehead region is split as multiple continuous tri patch by step S405.
It here, will be corresponding on each point and target forehead profile on " forehead is full " filling center, original forehead profile
The extension certain proportion of line of each point obtain vertex of a triangle set, center and vertex are filled according to " forehead is full "
Forehead region is split as multiple continuous tri patch by set, is fitted three-dimensional forehead using continuous tri patch.
Here the subregion in tri patch, that is, other embodiments.
Step S406, for each tri patch, using the stretching formula and bilinear interpolation ground certainly to triangle
Pixel in dough sheet does stretching.
Step S407 judges whether to have handled all tri patch.
Here, if having handled all tri patch, S408 is entered step;If not handled all triangular facets
Piece enters step S406.
Step S408, the effect picture after output " forehead is full " processing.
Image processing method provided in an embodiment of the present invention is on the basis of a small amount of key point of Face datection model calibration
It is upper further to use polygon approach forehead profile using Catmull-Rom polygon approach method.Organic U.S. face camera is to inspection
The precision and execution efficiency for surveying model have high requirement, and detection model can be effectively relieved using fitted polygon process
Performance pressures.Meanwhile the deformation algorithm used in stretching has linear complexity, efficiency faster, is adapted to camera
Live preview function is to the U.S. efficient demand of face algorithm, and the deformation method has certain fault-tolerance, can be directed to forehead
A certain range region carries out whole deformation around profile, can slow down crucial point tolerance bring negative effect, overall effect is more
Stablize.
Include following three techniqueflows in image processing method provided in an embodiment of the present invention, is to utilize folding respectively
Line segment is fitted the profile of forehead and target forehead, forehead region is split as multiple continuous tri patch and based on triangular facet
The figure stretching of piece.Wherein:
The first step utilizes the profile of broken line fitting forehead and target forehead.
Here, in this step, forehead key point, the posture (POSE) of face picture, the output of Face datection algorithm are inputted
Facial angle information and " forehead is full " dynamics parameter.Export the fitting broken line of forehead and target forehead profile.Fig. 5 is this hair
The schematic diagram for being fitted broken line of bright embodiment original forehead and target forehead profile, as shown in figure 5,501 be original forehead profile
Fitting broken line, 502 be target forehead profile fitting broken line.
During realization, which can be realized by following steps:
Step 11, the center-filled direction of " forehead is full " is determined using POSE angle information and forehead key point, in turn
Determine a filling center of " forehead is full ";
Step 12, " forehead is full " filling center and forehead key point are sequentially connected and is extended, according to " forehead is full " power
Degree parameter takes the point on extended line according to a certain percentage to determine the position of " forehead is full " forehead key point afterwards;
Step 13, using Catmull-Rom polygon approach method and the more forehead wheels of new and old forehead key point interpolation
Point on exterior feature connects the fitting broken line that these points constitute original forehead and target forehead profile.
In other embodiments, after step 13, it is also necessary to be sequentially connected " forehead is full " filling center and target volume
The point obtained after interpolation in head contour and extension take the point on extended line according to a certain percentage to determine adjustment point, successively connect
Connect the fitting broken line that adjacent adjustment point constitutes a transition profile.
Forehead region is split as multiple continuous tri patch by second step.
Here, the fitting for being fitted broken line and transition profile of original forehead, target forehead profile is inputted in this step
The filling center of broken line and " forehead is full " exports several continuous tri patch.
Fig. 6 is the schematic diagram that the embodiment of the present invention is split as forehead region multiple continuous tri patch, such as Fig. 6 institute
Show, be sequentially connected filling center and adjust the fitting broken line of profile.
Third step, the figure stretching based on tri patch.
Here, tri patch and Control point are inputted in this step, the tri patch after exporting stretching.
Fig. 3 B is the signal that the embodiment of the present invention carries out the tri patch and Control point that input when the full adjustment of forehead
Figure, as shown in Figure 3B, the stretching for controlling tri patch need altogether seven points, and wherein tri- points of ABC are tri patch
Three vertex, four additional point D, E and F, G are the point on original forehead profile and the corresponding broken line of target forehead profile respectively.
The core of stretching is that arbitrary point P in input tri patch is constructed mapping relations P '=f (P), so that defeated
The point P of tri patch corresponding position takes the pixel value at the position point P ' out, to complete the shift transformation that point P ' arrives point P.This hair
The figure stretching method that bright embodiment provides can deforming tri patch with fast and flexible.The stretching method
The picture at other based on control point can also be applied to deform field.
For each pixel P, tie point P in tri patch and extend and DE, FG and BC intersect at I respectively, J and
K.Length DIS_AP, DIS_AI, DIS_AJ and the DIS_AK of AP, AI, AJ and AK are found out respectively.Structural segmentation function y=g (x),
So that the corresponding piecewise function image passing point (0,0) of y=g (x), (1,1) and (DIS_AI/DIS_AK, DIS_AJ/DIS_
AK).X=DIS_AP/DIS_AK substitution equation y=g (x) is obtained into y=DIS_AP '/DIS_AK (piecewise function refers to Fig. 3 D).
In view of point A is filling center, so point P ' should be on line segment AK, and then the point P ' on available line segment AK, so that
The length of AP ' is DIS_AP '.
Finally, for arbitrary point P in tri patch, the pixel of the corresponding position point P ' is all taken, to complete stretching.
The case where being floating number for point P ' coordinate, bilinear interpolation algorithm can be used and get corresponding pixel value.Fig. 7 is the present invention
Embodiment carries out the comparison diagram of tri patch stretching when forehead full adjustment, as shown in fig. 7,701 be original tri patch,
702 be the tri patch after stretching, wherein hatched example areas is to be filled part, and white space is elongated portion.
Fig. 8 is the adjustment effect comparison diagram that the embodiment of the present invention carries out facial image when forehead full adjustment, wherein 801 are
Facial image to be processed, 802 be facial image adjusted, by comparing 801 and 802 it is found that people adjusted
The forehead of face image is fuller.
Tri patch fitting and stretching method are comprehensively utilized in the image processing method, and to complete camera, " forehead is full
It is full " function, to achieve the effect that three-dimensional U.S. face.Three-dimensional full face meet the aesthetic of Asians, and 3D transform can remold volume
Head, effect are more natural.
In the diminution adjustment for carrying out forehead region, the tri patch and Control point of input are as shown in Figure 3 C, control
The stretching of tri patch needs altogether seven points, and wherein tri- points of ABC are three vertex of tri patch, four additional point
D, E and F, G are the point on original forehead profile and the corresponding broken line of target forehead profile respectively, at this time D, E to A distance be greater than
F, the distance of G to A.
When carrying out the treatment process of stretching in the diminution adjustment for carrying out forehead region and carry out the full adjustment of forehead
Process it is similar, the piecewise function of use is as shown in FIGURE 3 E.When Fig. 9 A is that the embodiment of the present invention carries out the adjustment of forehead area reduction
The comparison diagram of tri patch stretching, as shown in Figure 9 A, 901 be original tri patch, and 902 be the triangular facet after stretching
Piece, wherein hatched example areas is to be filled part, and blank region is elongated portion.
Fig. 9 B is the adjustment effect comparison diagram that the embodiment of the present invention carries out facial image when the adjustment of forehead area reduction, wherein
911 be facial image to be processed, and 912 be facial image adjusted, after comparing 911 and 912 it is found that adjusting
The forehead region of facial image be obviously reduced.
3D " forehead is full " U.S. face can be carried out to human face photo using image processing method provided in an embodiment of the present invention,
Firstly, being fitted forehead wheel using Face datection model calibration forehead key point, and using Catmull-Rom polygon approach method
It is wide.Later, divide three-dimensional forehead using continuous tri patch according to clockwise direction.Finally, being directed to each triangular facet
Piece carries out stretching using stretching algorithm, reaches 3D " forehead is full " effect.In image provided in an embodiment of the present invention
In processing method, be fitted complicated, miscellaneous solid forehead using tri patch, then to each tri patch individually into
Row deformation algorithm simplifies implementation and greatly improves efficiency of algorithm.
Tri patch fitting, which can be played, breaks the whole up into parts, and simplifies deformation process and quickly establishes the work of 3D mathematical model
With can flexibly cope with different angle, the forehead of size and shape.Image processing method tool provided in an embodiment of the present invention simultaneously
There is whole the characteristics of deforming in proportion, so that deformation effect is more natural.
The embodiment of the present invention provides a kind of image processing apparatus, and Figure 10 is the composition of image processing apparatus of the embodiment of the present invention
Structural schematic diagram, as shown in Figure 10, the image processing apparatus 1000 include: the first acquisition module 1001, screening module 1002,
One determining module 1003, division module 1004 and stretching module 1005, in which:
Described first obtains module 1001, for obtaining Q key point in facial image, wherein Q is whole greater than 1
Number;
The screening module 1002, for filtering out default key point to be processed from the Q key point;
First determining module 1003, for according to the default key point to be processed determine in the facial image to
The target area of processing, wherein include the forehead region of facial image in target area;
Division module 1004, for target area to be divided into N number of subregion, wherein N is the integer greater than 2;
Stretching module 1005 is obtained for carrying out stretching to the pixel in each subregion respectively
Treated image.
In other embodiments, the default key point to be processed includes the of forehead region in the facial image
One feature point set, accordingly, the first determining module 1003 includes:
First acquisition unit, for obtaining the facial angle information and adjusting parameter of facial image;
First determination unit, for determining forehead region according to fisrt feature point set, facial angle information and adjusting parameter
Central point, the second point set on the first point set and objective contour on original contour;
Second determination unit, for according to central point, the second point set and preset second adjustment ratio-dependent target area.
In other embodiments, the first determination unit, comprising:
First determines subelement, for determining the center of target area according to fisrt feature point set and facial angle information
Point;
Second determines subelement, for determining second feature point set according to central point, fisrt feature point set and adjusting parameter;
Interpolation subelement, for being carried out respectively according to preset interpolation algorithm to fisrt feature point set and second feature point set
Interpolation accordingly obtains the first point set and the second point set.
In other embodiments, second determine that subelement is also used to:
Determine the first distance between central point and fisrt feature point;
The first adjustment distance is determined according to first distance and adjusting parameter;
It is special by being determined as corresponding second apart from the point that fisrt feature point is the first adjustment distance along the first adjustment direction
Levy point, wherein point arrives the direction of fisrt feature point centered on the first adjustment direction;
It obtains the fisrt feature and concentrates the corresponding second feature point of each fisrt feature point, obtain second feature point
Collection.
In other embodiments, interpolation subelement is also used to:
It is first object by the first subset of fisrt feature point set and the third subset division of corresponding second feature point set
The second subset of fisrt feature point set and the 4th subset division of corresponding second feature point set are the second target by subregion
Region;Interpolation is carried out according to preset interpolation algorithm respectively to the first subset and second subset of first object subregion, is obtained
Corresponding 5th subset and the 6th subset;
The intersection of 5th subset and the 6th subset is determined as the first point set;
Third subset and the 4th subset to the second target subregion carry out interpolation according to preset interpolation algorithm respectively, obtain
To corresponding 7th subset and the 8th subset;
The intersection of 7th subset and the 8th subset is determined as second subset.
In other embodiments, the second determination unit, comprising:
Third determines subelement, for determining second between central point and i-th of second point of second point concentration
Distance, wherein i=1,2 ..., N+1, (N+1) are the sum of second point;
4th determines subelement, for according to second distance and second adjustment ratio-dependent second adjustment distance;
Adjust subelement, for by along second adjustment direction i-th of second point of distance be second adjustment distance point it is true
It is set to the i-th adjustment point, wherein to the direction of i-th of second point centered on second adjustment direction o'clock;
5th determines subelement, for determining target area according to (N+1) a adjustment point.
In other embodiments, division module 1004, comprising:
Division unit, for central point, the i-th adjustment point and i+1 adjustment point to be sequentially connected and constitute i-th of subregion,
Wherein, i=1,2 ..., N, (N+1) are the sum of second point;
In other embodiments, stretching module 1005 includes:
Second acquisition unit, for obtaining the location information of j-th of pixel in i-th of subregion;
Third determination unit, for according to the location information of j-th of pixel, central point, it is described first point concentration i-th
A first point, i-th of second point concentrating of i+1 first point, the second point, i+1 second point, the i-th adjustment put and
I+1 adjustment point determines stretching function;
4th determination unit, for determining j-th of target according to the location information and stretching function of j-th pixel
Position;
5th determination unit, for determining the target pixel value of j-th of pixel according to j-th of target position;
Updating unit, for the pixel value of j-th of pixel to be updated to target pixel value, the image that obtains that treated.
In other embodiments, third determination unit, comprising:
Extend subelement, for by the 4th line of central point and j-th of pixel along central point to j-th pixel
Direction is extended, with described i-th first point, the i+1 first point of line intersect at the first intersection point, with described the
I second point, the i+1 second point line mutually give the second intersection point, with the 5th line intersect at third intersection point,
In, the 5th line is the line of the i-th adjustment point with i+1 adjustment point;
6th determines subelement, for determining stretching function according to the 4th distance, the 5th distance and the 6th distance,
In, the distance between point and the first intersection point centered on the 4th distance, the 5th puts the distance between second intersection point centered on,
The distance between point and third intersection point centered on 6th distance.
In other embodiments, the 6th determine that subelement is also used to:
Determine the second ratio between the first ratio, the 5th distance and the 6th distance between the 4th distance and the 6th distance
Value;
The first coordinate is determined according to the first ratio and the second ratio;
The linear equation of first coordinate and the line of origin is determined as the first piecewise function;
The linear equation of first coordinate and the line of preset second coordinate is determined as the second piecewise function;
Stretching function is determined according to the first piecewise function and the second piecewise function.
In other embodiments, the 4th determination unit, comprising:
7th determines subelement, for according to the location information of j-th of pixel determine j-th of pixel and central point it
Between the 7th distance;
8th determines subelement, determines the third ratio between the 7th distance and the 6th distance;
Computation subunit, for it is corresponding that third ratio to be calculated using third ratio as the input of stretching function
Output valve;
9th determines subelement, for determining the 8th distance according to output valve and the 6th distance, wherein the 8th distance is the
The distance between j target position and central point;
Tenth determines subelement, for determining jth according to the location information of the 8th distance, central point and j-th of pixel
A target position.
In other embodiments, the 5th determination unit, comprising:
11st determines subelement, is integer for the coordinate value in response to target position, by the pixel value of target position
It is determined as the target pixel value of j-th of pixel;
12nd determines subelement, is not integer for the coordinate value in response to target position, is determined according to preset algorithm
The corresponding pixel value in target position;The corresponding pixel value in target position is determined as to the target pixel value of j-th of pixel.
It should be noted that the description of apparatus above embodiment, be with the description of above method embodiment it is similar, have
The similar beneficial effect with embodiment of the method.For undisclosed technical detail in apparatus of the present invention embodiment, this hair is please referred to
The description of bright embodiment of the method and understand.
It should be noted that if realize above-mentioned image processing method in the form of software function module, and as only
Vertical product when selling or using, also can store in a computer readable storage medium.Based on this understanding, originally
Substantially the part that contributes to existing technology can be in the form of software products in other words for the technical solution of inventive embodiments
It embodies, which is stored in a storage medium, including some instructions are used so that a computer
Equipment (can be personal computer, server or network equipment etc.) execute each embodiment method of the present invention whole or
Part.And storage medium above-mentioned includes: USB flash disk, mobile hard disk, read-only memory (ROM, Read Only Memory), magnetic disk
Or the various media that can store program code such as CD.In this way, the embodiment of the present invention is not limited to any specific hardware
It is combined with software.
Correspondingly, the embodiment of the present invention provides a kind of computer storage medium again, is stored in the computer storage medium
Computer executable instructions, the computer executable instructions realize image procossing provided by the above embodiment when being executed by processor
The step of method.
Correspondingly, the embodiment of the present invention provides a kind of image processing equipment, and Figure 11 is that image procossing of the embodiment of the present invention is set
Standby composed structure schematic diagram, as shown in figure 11, the image processing equipment 1100 include: a processor 1101, at least one
Communication bus 1102, user interface 1103, at least one external communication interface 1104 and memory 1105.Wherein, communication bus
1102 are arranged for carrying out the connection communication between these components.Wherein, user interface 1103 may include display screen, PERCOM peripheral communication
Interface 1104 may include the wireline interface and wireless interface of standard.Wherein processor 1101, configuration pass through the communication bus
The image processing program that stores in memory is obtained, the step of to execute image processing method provided by the above embodiment.
The description of images above processing equipment and computer storage medium embodiment, the description with above method embodiment are
Similar, there is with embodiment of the method similar beneficial effect.For computer program product of the present invention, computer equipment and meter
Undisclosed technical detail in calculation machine storage medium embodiment, please refers to the description of embodiment of the present invention method and understands.
It should be understood that " one embodiment " or " embodiment " that specification is mentioned in the whole text mean it is related with embodiment
A particular feature, structure, or characteristic is included at least one embodiment of the present invention.Therefore, occur everywhere in the whole instruction
" in one embodiment " or " in one embodiment " not necessarily refer to identical embodiment.In addition, these specific features, knot
Structure or characteristic can combine in any suitable manner in one or more embodiments.It should be understood that in various implementations of the invention
In example, magnitude of the sequence numbers of the above procedures are not meant that the order of the execution order, and the execution sequence of each process should be with its function
It can determine that the implementation process of the embodiments of the invention shall not be constituted with any limitation with internal logic.The embodiments of the present invention
Serial number is for illustration only, does not represent the advantages or disadvantages of the embodiments.
It should be noted that, in this document, the terms "include", "comprise" or its any other variant are intended to non-row
His property includes, so that the process, method, article or the device that include a series of elements not only include those elements, and
And further include other elements that are not explicitly listed, or further include for this process, method, article or device institute it is intrinsic
Element.In the absence of more restrictions, the element limited by sentence "including a ...", it is not excluded that including being somebody's turn to do
There is also other identical elements in the process, method of element, article or device.
In several embodiments provided herein, it should be understood that disclosed device and method can pass through it
Its mode is realized.Apparatus embodiments described above are merely indicative, for example, the division of the unit, only
A kind of logical function partition, there may be another division manner in actual implementation, such as: multiple units or components can combine, or
It is desirably integrated into another system, or some features can be ignored or not executed.In addition, shown or discussed each composition portion
Mutual coupling or direct-coupling or communication connection is divided to can be through some interfaces, the INDIRECT COUPLING of equipment or unit
Or communication connection, it can be electrical, mechanical or other forms.
Above-mentioned unit as illustrated by the separation member, which can be or may not be, to be physically separated, aobvious as unit
The component shown can be or may not be physical unit;Both it can be located in one place, and may be distributed over multiple network lists
In member;Some or all of units can be selected to achieve the purpose of the solution of this embodiment according to the actual needs.
In addition, each functional unit in various embodiments of the present invention can be fully integrated in one processing unit, it can also
To be each unit individually as a unit, can also be integrated in one unit with two or more units;It is above-mentioned
Integrated unit both can take the form of hardware realization, can also realize in the form of hardware adds SFU software functional unit.
Those of ordinary skill in the art will appreciate that: realize that all or part of the steps of above method embodiment can pass through
The relevant hardware of program instruction is completed, and program above-mentioned can store in computer-readable storage medium, which exists
When execution, step including the steps of the foregoing method embodiments is executed;And storage medium above-mentioned includes: movable storage device, read-only deposits
The various media that can store program code such as reservoir (Read Only Memory, ROM), magnetic or disk.
If alternatively, the above-mentioned integrated unit of the present invention is realized in the form of software function module and as independent product
When selling or using, it also can store in a computer readable storage medium.Based on this understanding, the present invention is implemented
Substantially the part that contributes to existing technology can be embodied in the form of software products the technical solution of example in other words,
The computer software product is stored in a storage medium, including some instructions are used so that computer equipment (can be with
It is personal computer, server or network equipment etc.) execute all or part of each embodiment the method for the present invention.
And storage medium above-mentioned includes: various Jie that can store program code such as movable storage device, ROM, magnetic or disk
Matter.
The above description is merely a specific embodiment, but scope of protection of the present invention is not limited thereto, any
Those familiar with the art in the technical scope disclosed by the present invention, can easily think of the change or the replacement, and should all contain
Lid is within protection scope of the present invention.Therefore, protection scope of the present invention should be based on the protection scope of the described claims.
Claims (10)
1. a kind of image processing method, which is characterized in that the described method includes:
Obtain Q key point in facial image, wherein Q is the integer greater than 1;
Default key point to be processed is filtered out from the Q key point;
Target area to be processed in the facial image, the target area are determined according to the default key point to be processed
It include the forehead region of facial image in domain;
The target area is divided into N number of subregion, N is the integer more than or equal to 2;
Stretching is carried out to the pixel in each subregion respectively, the image that obtains that treated.
2. image processing method according to claim 1, which is characterized in that the default key point packet to be processed
The fisrt feature point set in forehead region in the facial image is included, it is described that people is determined according to the default key point to be processed
Target area to be processed in face image, comprising:
Obtain the facial angle information and adjusting parameter of the facial image;
The central point, original in the forehead region is determined according to the fisrt feature point set, facial angle information and adjusting parameter
The first point set on profile and the second point set on objective contour;
According to target area described in the central point, the second point set and preset second adjustment ratio-dependent.
3. image processing method according to claim 2, which is characterized in that it is described according to the fisrt feature point set,
Facial angle information and adjusting parameter determine the central point in the forehead region, the first point set and objective contour on original contour
On the second point set, comprising:
The central point of the target area is determined according to the fisrt feature point set and the facial angle information;
Second feature point set is determined according to the central point, the fisrt feature point set and adjusting parameter;
Interpolation is carried out according to preset interpolation algorithm respectively to the fisrt feature point set and the second feature point set, it is mutually deserved
To first point set and the second point set.
4. image processing method according to claim 3, which is characterized in that it is described according to the central point, described
One feature point set and adjusting parameter determine second feature point set, comprising:
Determine the first distance between the central point and fisrt feature point;
The first adjustment distance is determined according to the first distance and the adjusting parameter;
The point that the point of fisrt feature described in the distance along the first adjustment direction is the first adjustment distance is determined as corresponding the
Two characteristic points, wherein point arrives the direction of fisrt feature point centered on the first adjustment direction;
It obtains the fisrt feature and concentrates the corresponding second feature point of each fisrt feature point, obtain second feature point set.
5. image processing method according to claim 3, which is characterized in that described to the fisrt feature point set and institute
It states second feature point set and carries out interpolation according to preset interpolation algorithm respectively, accordingly obtain first point set and the second point set,
Include:
It is first object by the first subset of the fisrt feature point set and the third subset division of corresponding second feature point set
The second subset of the fisrt feature point set and the 4th subset division of corresponding second feature point set are the second mesh by subregion
Mark subregion;
To first subset of the first object subregion and the second subset respectively according to preset interpolation algorithm into
Row interpolation obtains corresponding 5th subset and the 6th subset;
The intersection of 5th subset and the 6th subset is determined as the first point set;
The third subset and the 4th subset to the second target subregion respectively according to preset interpolation algorithm into
Row interpolation obtains corresponding 7th subset and the 8th subset;
The intersection of 7th subset and the 8th subset is determined as second subset.
6. image processing method according to claim 2, which is characterized in that described according to the central point, second point
Target area described in collection and preset second adjustment ratio-dependent, comprising:
Determine the second distance between the central point and i-th of second point of second point concentration, wherein i=1,2 ...,
N+1, (N+1) are the sum of second point;
According to the second distance and the second adjustment ratio-dependent second adjustment distance;
It is that the point of the second adjustment distance is determined as the i-th adjustment by i-th of second point described in the distance along second adjustment direction
Point, wherein the second adjustment direction is the central point to the direction of i-th of second point;
The target area is determined according to (N+1) a described adjustment point.
7. image processing method according to claim 6, which is characterized in that described that the target area is divided into N
Sub-regions, comprising:
The central point, the i-th adjustment point and i+1 adjustment point are sequentially connected and constitute i-th of subregion, wherein i=1,
2 ..., N, (N+1) are the sum of second point.
8. a kind of image processing apparatus, which is characterized in that described image processing unit include: the first acquisition module, screening module,
First determining module, division module and stretching module, in which:
Described first obtains module, for obtaining Q key point in facial image, wherein Q is the integer greater than 1;
The screening module, for filtering out default key point to be processed from the Q key point;
First determining module, it is to be processed in the facial image for being determined according to the default key point to be processed
Target area, wherein in the target area include facial image forehead region;
The division module, for the target area to be divided into N number of subregion, wherein N is the integer more than or equal to 2;
The stretching module is handled for carrying out stretching to the pixel in each subregion respectively
Image afterwards.
9. a kind of image processing equipment, which is characterized in that described image processing equipment includes at least: memory, communication bus and
Processor, in which:
The memory, for storing image processing program;
The communication bus, for realizing the connection communication between processor and memory;
The processor, for obtaining the image processing program stored in memory by the communication bus, with perform claim
It is required that the step of image processing method described in 1 to 7 any one.
10. a kind of storage medium, which is characterized in that be stored with image processing program, described image processing on the storage medium
The step of image processing method described in any one of claim 1 to 7 is realized when program is executed by processor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811281028.1A CN109359618B (en) | 2018-10-30 | 2018-10-30 | Image processing method and device, equipment and storage medium thereof |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811281028.1A CN109359618B (en) | 2018-10-30 | 2018-10-30 | Image processing method and device, equipment and storage medium thereof |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109359618A true CN109359618A (en) | 2019-02-19 |
CN109359618B CN109359618B (en) | 2022-02-22 |
Family
ID=65347211
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811281028.1A Active CN109359618B (en) | 2018-10-30 | 2018-10-30 | Image processing method and device, equipment and storage medium thereof |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109359618B (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110097506A (en) * | 2019-05-07 | 2019-08-06 | 北京市商汤科技开发有限公司 | Image processing method and device, vision facilities and storage medium |
CN110223220A (en) * | 2019-06-14 | 2019-09-10 | 北京百度网讯科技有限公司 | A kind of method and apparatus handling image |
CN110288521A (en) * | 2019-06-29 | 2019-09-27 | 北京字节跳动网络技术有限公司 | Image beautification method, device and electronic equipment |
CN110334653A (en) * | 2019-07-08 | 2019-10-15 | 聚好看科技股份有限公司 | Image processing method, device and equipment in video communication |
CN111126344A (en) * | 2019-12-31 | 2020-05-08 | 杭州趣维科技有限公司 | Method and system for generating key points of forehead of human face |
CN112348736A (en) * | 2020-10-12 | 2021-02-09 | 武汉斗鱼鱼乐网络科技有限公司 | Method, storage medium, device and system for removing black eye |
CN112767288A (en) * | 2021-03-19 | 2021-05-07 | 北京市商汤科技开发有限公司 | Image processing method and device, electronic equipment and storage medium |
CN113591562A (en) * | 2021-06-23 | 2021-11-02 | 北京旷视科技有限公司 | Image processing method, image processing device, electronic equipment and computer readable storage medium |
CN113822804A (en) * | 2021-09-23 | 2021-12-21 | 深圳万兴软件有限公司 | Gray scale map optimization method and device, computer equipment and storage medium |
CN115619410A (en) * | 2022-10-19 | 2023-01-17 | 闫雪 | Self-adaptive financial payment platform |
WO2023173953A1 (en) * | 2022-03-15 | 2023-09-21 | 华为技术有限公司 | Probe data processing and coding methods and devices |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5811690B2 (en) * | 2011-08-24 | 2015-11-11 | カシオ計算機株式会社 | Image processing apparatus, image processing method, and program |
CN105894446A (en) * | 2016-05-09 | 2016-08-24 | 西安北升信息科技有限公司 | Automatic face outline modification method for video |
CN107203963A (en) * | 2016-03-17 | 2017-09-26 | 腾讯科技(深圳)有限公司 | A kind of image processing method and device, electronic equipment |
CN107330868A (en) * | 2017-06-26 | 2017-11-07 | 北京小米移动软件有限公司 | image processing method and device |
CN107527034A (en) * | 2017-08-28 | 2017-12-29 | 维沃移动通信有限公司 | A kind of face contour method of adjustment and mobile terminal |
CN107578371A (en) * | 2017-09-29 | 2018-01-12 | 北京金山安全软件有限公司 | Image processing method and device, electronic equipment and medium |
CN107657590A (en) * | 2017-09-01 | 2018-02-02 | 北京小米移动软件有限公司 | Image processing method and device |
CN107730449A (en) * | 2017-11-07 | 2018-02-23 | 深圳市云之梦科技有限公司 | A kind of method and system of human face five-sense-organ landscaping treatment |
CN107818543A (en) * | 2017-11-09 | 2018-03-20 | 北京小米移动软件有限公司 | Image processing method and device |
-
2018
- 2018-10-30 CN CN201811281028.1A patent/CN109359618B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5811690B2 (en) * | 2011-08-24 | 2015-11-11 | カシオ計算機株式会社 | Image processing apparatus, image processing method, and program |
CN107203963A (en) * | 2016-03-17 | 2017-09-26 | 腾讯科技(深圳)有限公司 | A kind of image processing method and device, electronic equipment |
CN105894446A (en) * | 2016-05-09 | 2016-08-24 | 西安北升信息科技有限公司 | Automatic face outline modification method for video |
CN107330868A (en) * | 2017-06-26 | 2017-11-07 | 北京小米移动软件有限公司 | image processing method and device |
CN107527034A (en) * | 2017-08-28 | 2017-12-29 | 维沃移动通信有限公司 | A kind of face contour method of adjustment and mobile terminal |
CN107657590A (en) * | 2017-09-01 | 2018-02-02 | 北京小米移动软件有限公司 | Image processing method and device |
CN107578371A (en) * | 2017-09-29 | 2018-01-12 | 北京金山安全软件有限公司 | Image processing method and device, electronic equipment and medium |
CN107730449A (en) * | 2017-11-07 | 2018-02-23 | 深圳市云之梦科技有限公司 | A kind of method and system of human face five-sense-organ landscaping treatment |
CN107818543A (en) * | 2017-11-09 | 2018-03-20 | 北京小米移动软件有限公司 | Image processing method and device |
Non-Patent Citations (2)
Title |
---|
MARCO MAGGINI ET AL.: "Representation of Facial Features by Catmull-Rom Splines", 《COMPUTER ANALYSIS OF IMAGE AND PATTERNS》 * |
刘洁娟: "基于MPEG-4的人脸建模和动画实现研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110097506A (en) * | 2019-05-07 | 2019-08-06 | 北京市商汤科技开发有限公司 | Image processing method and device, vision facilities and storage medium |
CN110223220A (en) * | 2019-06-14 | 2019-09-10 | 北京百度网讯科技有限公司 | A kind of method and apparatus handling image |
CN110288521A (en) * | 2019-06-29 | 2019-09-27 | 北京字节跳动网络技术有限公司 | Image beautification method, device and electronic equipment |
CN110334653A (en) * | 2019-07-08 | 2019-10-15 | 聚好看科技股份有限公司 | Image processing method, device and equipment in video communication |
CN111126344B (en) * | 2019-12-31 | 2023-08-01 | 杭州趣维科技有限公司 | Method and system for generating key points of forehead of human face |
CN111126344A (en) * | 2019-12-31 | 2020-05-08 | 杭州趣维科技有限公司 | Method and system for generating key points of forehead of human face |
CN112348736A (en) * | 2020-10-12 | 2021-02-09 | 武汉斗鱼鱼乐网络科技有限公司 | Method, storage medium, device and system for removing black eye |
CN112767288A (en) * | 2021-03-19 | 2021-05-07 | 北京市商汤科技开发有限公司 | Image processing method and device, electronic equipment and storage medium |
WO2022193466A1 (en) * | 2021-03-19 | 2022-09-22 | 北京市商汤科技开发有限公司 | Image processing method and apparatus, and electronic device and storage medium |
CN113591562A (en) * | 2021-06-23 | 2021-11-02 | 北京旷视科技有限公司 | Image processing method, image processing device, electronic equipment and computer readable storage medium |
CN113822804A (en) * | 2021-09-23 | 2021-12-21 | 深圳万兴软件有限公司 | Gray scale map optimization method and device, computer equipment and storage medium |
WO2023173953A1 (en) * | 2022-03-15 | 2023-09-21 | 华为技术有限公司 | Probe data processing and coding methods and devices |
CN115619410A (en) * | 2022-10-19 | 2023-01-17 | 闫雪 | Self-adaptive financial payment platform |
CN115619410B (en) * | 2022-10-19 | 2024-01-26 | 闫雪 | Self-adaptive financial payment platform |
Also Published As
Publication number | Publication date |
---|---|
CN109359618B (en) | 2022-02-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109359618A (en) | A kind of image processing method and its device, equipment and storage medium | |
WO2021057027A1 (en) | Human body detection method and apparatus, computer device, and storage medium | |
CN110287846B (en) | Attention mechanism-based face key point detection method | |
CN107577990B (en) | Large-scale face recognition method based on GPU (graphics processing Unit) accelerated retrieval | |
CN110717851B (en) | Image processing method and device, training method of neural network and storage medium | |
CN106067190B (en) | A kind of generation of fast face threedimensional model and transform method based on single image | |
CN104317391B (en) | A kind of three-dimensional palm gesture recognition exchange method and system based on stereoscopic vision | |
CN102156969B (en) | Processing method for correcting deviation of image | |
CN109829437A (en) | Image processing method, text recognition method, device and electronic system | |
EP2706507B1 (en) | Method and apparatus for generating morphing animation | |
US11636668B2 (en) | Bilateral convolution layer network for processing point clouds | |
CN105913051B (en) | A kind of updating device and method of the template library identifying facial image | |
CN109472753A (en) | A kind of image processing method, device, computer equipment and computer storage medium | |
CN107204010A (en) | A kind of monocular image depth estimation method and system | |
CN106022355B (en) | High spectrum image sky based on 3DCNN composes joint classification method | |
CN105825163A (en) | Retrieval system and method of face image | |
CN102567716B (en) | Face synthetic system and implementation method | |
CN113538569B (en) | Weak texture object pose estimation method and system | |
CN102479379A (en) | Image rectification method and relevant image rectification system | |
CN109261528A (en) | Express delivery sorting method and device based on binocular vision | |
CN116583878A (en) | Method and system for personalizing 3D head model deformation | |
CN107704848A (en) | A kind of intensive face alignment method based on multi-constraint condition convolutional neural networks | |
CN106910202A (en) | The image partition method and system of a kind of remote sensing images atural object | |
CN110288715A (en) | Virtual necklace try-in method, device, electronic equipment and storage medium | |
CN112509106A (en) | Document picture flattening method, device and equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |