WO2016011834A1 - Image processing method and system - Google Patents

Image processing method and system Download PDF

Info

Publication number
WO2016011834A1
WO2016011834A1 PCT/CN2015/077353 CN2015077353W WO2016011834A1 WO 2016011834 A1 WO2016011834 A1 WO 2016011834A1 CN 2015077353 W CN2015077353 W CN 2015077353W WO 2016011834 A1 WO2016011834 A1 WO 2016011834A1
Authority
WO
WIPO (PCT)
Prior art keywords
face
person
character
model
image
Prior art date
Application number
PCT/CN2015/077353
Other languages
French (fr)
Chinese (zh)
Inventor
邢小月
姜涌
孟昭龙
Original Assignee
邢小月
姜涌
孟昭龙
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 邢小月, 姜涌, 孟昭龙 filed Critical 邢小月
Publication of WO2016011834A1 publication Critical patent/WO2016011834A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects

Abstract

The present invention provides an image processing method and system. The method comprises: simulating a model of a first person according to at least one image containing the first person; determining a target image containing a second person; determining feature information for displaying the second person in the target image; adjusting the display of the first person according to the feature information in the model of the first person; and replacing the second person by the first person after display adjustment in the target image. By means of the technical solution provided in the present invention, the problem that replaced persons and backgrounds are inconsistent and conflicting due to absence of the relation between the persons before and after replacement in image processing is solved.

Description

一种图像处理方法及系统Image processing method and system 技术领域Technical field
本发明涉及图像处理技术,特别涉及一种图像处理方法及系统。The present invention relates to image processing technologies, and in particular, to an image processing method and system.
背景技术Background technique
现有技术在进行图像处理,将图像中的一个人物替换到另一张图像中时,只是简单的将一个人物的头部或者面部沿着轮廓裁剪下来,叠加在另一个人物的图像相应位置,类似于大头贴的效果。一方面,其所替换的人物会因光照、视角等与背景不一致而会出现人物与背景颜色相冲突、色调不搭配的情况;另一方面,当把人物面部替换到另外一个人的面部时,只能保留原来人物的表情,而这表情通常都与背景不一致,这种按现有技术替换后出现的人物与背景的不协调显然无法满足人们的需求。In the prior art, when image processing is performed to replace one character in an image into another image, the head or face of one character is simply cut along the contour and superimposed on the corresponding position of the image of another character. Similar to the effect of the photo sticker. On the one hand, the characters replaced by the characters may be inconsistent with the background due to illumination, viewing angle, etc., and the characters may be in conflict with the background color and the color tone may not match; on the other hand, when the face of the person is replaced with the face of another person, Only the expression of the original character can be retained, and this expression is usually inconsistent with the background. The disharmony between the character and the background appearing after the replacement of the prior art obviously cannot meet the needs of the people.
现有技术的不足在于:The shortcomings of the prior art are:
替换前后人物之间没有联系,使得替换后的人物与背景出现不协调、冲突等问题。There is no connection between the characters before and after the replacement, which causes the replaced characters to appear inconsistent and conflict with the background.
发明内容Summary of the invention
本发明针对上述问题,提出了一种图像处理方法及系统,用以解决图像模拟替换时,人物图像与所替换的图像背景不协调的问题。The present invention is directed to the above problem, and provides an image processing method and system for solving the problem that the image of a person is inconsistent with the background of the replaced image when the image is simulated and replaced.
本发明实施例中提供了一种图像处理方法,可以包括如下步骤:An embodiment of the present invention provides an image processing method, which may include the following steps:
根据至少一张包含第一人物的图像,模拟出第一人物的模型;Simulating a model of the first character based on at least one image containing the first person;
确定包含第二人物的目标图像;Determining a target image containing the second person;
确定在目标图像中显示第二人物的特征信息; Determining to display feature information of the second person in the target image;
在第一人物的模型中根据所述特征信息调整第一人物的显示;Adjusting the display of the first character according to the feature information in the model of the first character;
在目标图像中,将第二人物替换为显示调整后的第一人物。In the target image, the second character is replaced with the adjusted first character.
本发明实施例中提供了一种图像处理系统,可以包括:An embodiment of the present invention provides an image processing system, which may include:
模型模拟模块,用于根据至少一张包含第一人物的图像,模拟出第一人物的模型;a model simulation module, configured to simulate a model of the first character according to at least one image including the first character;
目标图像确定模块,用于确定包含第二人物的目标图像;a target image determining module, configured to determine a target image that includes the second person;
特征信息确定模块,用于确定在目标图像中显示第二人物的特征信息;a feature information determining module, configured to determine that feature information of the second person is displayed in the target image;
调整显示模块,用于在第一人物的模型中根据所述特征信息调整第一人物的显示;Adjusting a display module, configured to adjust a display of the first character according to the feature information in a model of the first character;
人物替换模块,用于在目标图像中,将第二人物替换为显示调整后的第一人物。a character replacement module for replacing the second character with the adjusted first character in the target image.
本发明有益效果如下:The beneficial effects of the present invention are as follows:
在本发明实施例提供的技术方案中,首先模拟出第一人物的模型,然后根据作为被替代对象的第二人物在目标图像中显示的特征信息进行调整,这样,使得第一人物与第二人物在目标图像中都具备同样的显示特征,从而克服了替换后与目标图像背景等不协调的问题。In the technical solution provided by the embodiment of the present invention, the model of the first character is first simulated, and then the feature information displayed in the target image is adjusted according to the second character as the replaced object, so that the first character and the second character are The characters have the same display characteristics in the target image, thereby overcoming the problem of inconsistency with the background of the target image after replacement.
附图说明DRAWINGS
下面将参照附图描述本发明的具体实施例,其中:Specific embodiments of the present invention will be described below with reference to the accompanying drawings, in which:
图1为本发明实施例中图像处理方法实施流程示意图;1 is a schematic flowchart of an implementation process of an image processing method according to an embodiment of the present invention;
图2为本发明实施例中人脸检测算法实施流程示意图;2 is a schematic flowchart of an implementation process of a face detection algorithm according to an embodiment of the present invention;
图3为本发明实施例中提取Haar-like特征示意图;3 is a schematic diagram of extracting Haar-like features in an embodiment of the present invention;
图4为本发明实施例中积分图的方法实施流程示意图;4 is a schematic flowchart of a method for implementing an integral map in an embodiment of the present invention;
图5为本发明实施例中瀑布型级联检测器示意图;5 is a schematic diagram of a waterfall cascade detector according to an embodiment of the present invention;
图6为本发明实施例中标定的人脸面部示意图;6 is a schematic diagram of a face and a face of a calibration in an embodiment of the present invention;
图7为本发明实施例中局部特征的创建过程示意图;FIG. 7 is a schematic diagram of a process of creating a local feature in an embodiment of the present invention; FIG.
图8为本发明实施例中计算每个特征点的新位置的方法实施流程示意图; FIG. 8 is a schematic flowchart of a method for calculating a new location of each feature point according to an embodiment of the present invention; FIG.
图9为本发明实施例中人脸检测结果示意图;FIG. 9 is a schematic diagram of a face detection result according to an embodiment of the present invention; FIG.
图10为本发明实施例中三维人脸重建实施流程示意图;FIG. 10 is a schematic flowchart of a three-dimensional face reconstruction implementation process according to an embodiment of the present invention; FIG.
图11为本发明实施例中原始图像和三维模型示意图;11 is a schematic diagram of an original image and a three-dimensional model in an embodiment of the present invention;
图12为本发明实施例中模型的表情实例示意图;12 is a schematic diagram of an example of an expression of a model in an embodiment of the present invention;
图13为本发明实施例中人物表情特征点示意图;FIG. 13 is a schematic diagram of character expression points of a person in an embodiment of the present invention; FIG.
图14为本发明实施例中图像处理系统结构示意图。FIG. 14 is a schematic structural diagram of an image processing system according to an embodiment of the present invention.
具体实施方式detailed description
为使本发明实施例的目的、技术方案和优点更加清楚明白,下面结合附图对本发明实施例做进一步详细说明。在此,本发明的示意性实施例及其说明用于解释本发明,但并不作为对本发明的限定。The embodiments of the present invention will be further described in detail below with reference to the accompanying drawings. The illustrative embodiments of the present invention and the description thereof are intended to explain the present invention, but are not intended to limit the invention.
图1为图像处理方法实施流程示意图,如图1所示,可以包括如下步骤:FIG. 1 is a schematic diagram of an implementation process of an image processing method, as shown in FIG. 1 , which may include the following steps:
步骤101:根据至少一张包含第一人物的图像,模拟出第一人物的模型;Step 101: Simulate a model of the first character according to at least one image including the first character;
步骤102:确定包含第二人物的目标图像;Step 102: Determine a target image that includes the second person;
步骤103:确定在目标图像中显示第二人物的特征信息;Step 103: Determine to display feature information of the second person in the target image;
步骤104:在第一人物的模型中根据所述特征信息调整第一人物的显示;Step 104: Adjusting, in the model of the first character, the display of the first character according to the feature information;
步骤105:在目标图像中,将第二人物替换为显示调整后的第一人物。Step 105: In the target image, replace the second character with the displayed first character.
具体的,在对图像进行替换时,可以根据用户提供的图片或图像序列自动完成对人物的编辑工作。例如可以如下:Specifically, when the image is replaced, the editing work of the character can be automatically completed according to the picture or image sequence provided by the user. For example, the following can be:
a.用户提供一张或多张图片或者图像序列作为素材,其所有素材中均含有同一个人物,即:第一人物;a. The user provides one or more pictures or a sequence of images as the material, all of which contain the same person, namely: the first person;
b.系统根据用户提供的素材,模拟出该第一人物的模型。该模型可针对不同的视角、光照等做出相应的调整,并且可做出不同的形变;b. The system simulates the model of the first character based on the material provided by the user. The model can be adjusted for different perspectives, illuminations, etc., and can be deformed differently;
c.用户在图片或图像序列中指定另一个人物,即:第二人物;c. The user specifies another character in the picture or image sequence, ie: the second person;
d.系统检测出该指定的第二人物在每帧图像中的相关特征信息。这些特征信息是指位置、轮廓、相对视角、光照及形变等特征;d. The system detects relevant feature information of the designated second character in each frame of the image. These feature information refers to features such as position, contour, relative viewing angle, illumination and deformation;
e.在每帧图像上,将第一人物模型调整到第二人物在该帧图像的特性上, 并替换第二人物。e. on each frame of the image, the first character model is adjusted to the characteristics of the second character in the frame image, And replace the second person.
实施中,对单张图像进行的提取、处理、调整、转换进行了说明,由于多张图像序列和视频图像的每一帧均是由单张图像构成,因此,以本发明实施例提供的技术方案为基础,可以容易的得出对多张或批量图片组成的图像序列,或者是对视频图像的处理,比如,一种最简单的方式是:对图像序列或视频的每张图像进行替换处理后,再组成替换后的图像序列或视频。如何在单张图像的处理基础上扩展至对整个图像序列或视频的处理,这是本领域技术人员容易理解并作出相应修改的。In the implementation, the extraction, processing, adjustment, and conversion of the single image are described. Since each of the multiple image sequences and the video image is composed of a single image, the technology provided by the embodiment of the present invention is used. Based on the scheme, it is easy to draw a sequence of images composed of multiple or batch pictures, or to process video images. For example, one of the simplest ways is to replace each image sequence or video. After that, the replaced image sequence or video is composed. How to extend the processing of the entire image sequence or video based on the processing of a single image is easily understood and modified by those skilled in the art.
实施中,本发明实施例中的人物可以是拟人化人物,如卡通人物、3D人物等,其不仅限于人类的人物,也不必是自然存在的人物,在实施例中均称为“人物”。下面的实施例中大多也是以人类的图像处理为例,这是因为其最具代表性,也最为复杂。所以这里以人像为例进行说明;但是,本发明实施例提供的技术方案也可以用其它的图像处理,因为其披露的是一种用于图像处理的涉及替换的方案,也即,图像处理领域内对所有图案实现替换这一目的时,都可以采用本发明实施例中的方案,理论上并不仅限于人物,人像仅用于教导本领域技术人员具体如何实施本发明,但不意味仅能使用于人像,实施过程中可以结合实践需要在相应的环境中使用。In an embodiment, the character in the embodiment of the present invention may be an anthropomorphic person, such as a cartoon character, a 3D character, or the like, which is not limited to a human character, nor is it a naturally occurring person, and is referred to as a “person” in the embodiment. Most of the following embodiments are also based on human image processing because they are the most representative and most complex. Therefore, the portrait is taken as an example here; however, the technical solution provided by the embodiment of the present invention can also use other image processing, because it discloses an alternative scheme for image processing, that is, the field of image processing. The solution in the embodiment of the present invention can be adopted for the purpose of replacing all the patterns in the embodiment. The theory is not limited to the characters. The portrait is only used to teach the person skilled in the art how to implement the present invention, but it does not mean that only the invention can be used. For portraits, the implementation process can be used in the corresponding environment in combination with practice.
实施中,在根据至少一张包含第一人物的图像,模拟出第一人物的脸部的模型时,可以包括:In the implementation, when the model of the face of the first person is simulated according to at least one image including the first person, the method may include:
检测出第一人物的脸部的区域;Detecting the area of the face of the first person;
在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;Determining the contours of the facial features and cheeks in the area of the detected face;
将检测到的五官的区域和脸颊的轮廓,贴合到已有的人脸三维3D模型上后获得模拟出的第一人物的脸部的模型。The model of the face of the first character is obtained by fitting the detected area of the facial features and the contour of the cheek to the existing three-dimensional 3D model of the face.
具体的,根据模拟出的第一人物的模型,可以是人物全身的模型,也可以是人物脸部的模型,实施例中将以人物脸部的实施为例,但本领域技术人员应当知晓,采用相应的图像工具进行处理,即可获得不限于脸部的处理方 式,比如人物全身的模型。Specifically, according to the simulated model of the first person, it may be a model of a person's whole body or a model of a person's face. In the embodiment, the implementation of the face of the person will be taken as an example, but those skilled in the art should know that Use the corresponding image tool to process, you can get the processing side not limited to the face , such as the model of a person's whole body.
以人物脸部的实施为例,可以如下:Taking the implementation of a person's face as an example, it can be as follows:
a.从用户提供的图片或图像序列中检测出第一人物的人脸的位置和区域;a. detecting the location and area of the face of the first person from a sequence of pictures or images provided by the user;
b.在检测出的人脸的区域里,确定人脸的五官的区域和脸颊的轮廓,如:眼睛、鼻子、眉毛、嘴和耳朵;b. In the area of the detected face, determine the facial features of the face and the contours of the cheeks, such as: eyes, nose, eyebrows, mouth and ears;
c.将检测到的五官的区域和脸颊的轮廓,贴合到一个已有的人脸3维模型上,使其能够自动的根据参数的设定,呈现不同的视角、光照和表情的变化。c. The detected facial features and cheeks are attached to an existing 3D model of the face, so that it can automatically display different perspectives, illuminations and expression changes according to the parameters.
实施中,在确定在目标图像中显示第二人物的脸部的特征信息时,可以包括:In the implementation, when determining the feature information of displaying the face of the second person in the target image, the method may include:
检测出第二人物的脸部的区域;Detecting the area of the face of the second person;
在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;Determining the contours of the facial features and cheeks in the area of the detected face;
根据检测到的五官的区域和脸颊的轮廓,确定在目标图像中显示第二人物的脸部的特征信息。Based on the detected area of the facial features and the outline of the cheek, it is determined that the feature information of the face of the second person is displayed in the target image.
具体的,检测出指定的第二人物在每帧图像中的相关特性信息,可以如下:Specifically, the related characteristic information of the specified second character in each frame image is detected, which may be as follows:
a.从图片或图像序列中检测出第二人物的人脸的位置和区域;a. detecting the position and area of the face of the second person from the sequence of pictures or images;
b.在检测出的人脸的区域里,确定人脸的五官的区域和脸颊的轮廓,如:眼睛、鼻子、眉毛、嘴和耳朵;b. In the area of the detected face, determine the facial features of the face and the contours of the cheeks, such as: eyes, nose, eyebrows, mouth and ears;
c.通过检测到的五官的区域和脸颊的轮廓,推断第二人物的相关特性信息。这些特性信息包括视角、光照和表情的变化等。c. Inferring the relevant characteristic information of the second person through the detected area of the facial features and the outline of the cheek. These characteristics include changes in perspective, lighting, and expressions.
具体的,在检测出的脸部的区域里,采用人脸识别算法确定五官的区域和脸颊的轮廓,可以采用ASM(Active Shape Model,主动形状模型)算法确定五官的区域和脸颊的轮廓。Specifically, in the detected area of the face, the face recognition algorithm is used to determine the contour of the facial features and the cheeks. The ASM (Active Shape Model) algorithm can be used to determine the contours of the facial features and cheeks.
在实施中,采用ASM算法来进行说明是因为ASM算法在人脸识别算法中比较典型,也较为常用,容易被本领域技术人员理解实施,所以这里以ASM 算法为例;但是,从理论上来说,用其它的算法也是可以的,只要能够达到确定五官的区域和脸颊的轮廓这一目的即可,例如可以采用AAM(Active Appearance Model,主动表现模型)或SDM(Supervised Descent Method,监督梯度下降法)等算法。因此,ASM算法仅用于教导本领域技术人员具体如何实施本发明,但不意味仅能使用ASM算法,实施过程中可以结合实践需要来确定相应的算法。In the implementation, the ASM algorithm is used for explanation because the ASM algorithm is typical in the face recognition algorithm, and is also commonly used, and is easy to be understood by those skilled in the art, so here is ASM. The algorithm is an example; however, in theory, other algorithms are also possible, as long as the contour of the facial features and the cheeks can be determined, for example, an AAM (Active Appearance Model) or SDM (Supervised Descent Method) and other algorithms. Therefore, the ASM algorithm is only used to teach the person skilled in the art how to implement the invention, but it does not mean that only the ASM algorithm can be used, and the corresponding algorithm can be determined in the implementation process in combination with practical needs.
实施中,将第二人物替换为显示调整后的第一人物,是根据第一人物的脸部的区域与第二人物的脸部的区域将第二人物的脸部替换为显示调整后的第一人物的脸部。In the implementation, replacing the second person with the adjusted first character is to replace the face of the second person with the adjusted face according to the area of the face of the first person and the area of the face of the second person The face of a character.
具体的,将第一人物模型进一步替换第二人物,可以如下:Specifically, the first character model is further replaced by the second character, which can be as follows:
a.根据第二人物的相关特征信息,调整第一人物的模型,使其与第二人物的相关特性相仿;a. adjusting the model of the first character according to the related feature information of the second character to make it similar to the related characteristic of the second character;
b.通过检测到的五官的区域和脸颊的轮廓,将每帧图像中的第二人物的脸部区域抹去;b. erasing the face area of the second person in each frame image by detecting the contour of the facial features and the contour of the cheek;
c.在每帧图像中,将调整好的第一人物的模型放置在第二人物的脸部区域。c. In each frame of image, the model of the adjusted first person is placed in the face area of the second person.
实施中,在检测出的脸部的区域里,可以采用人脸识别算法确定五官的区域和脸颊的轮廓。In the implementation, in the detected area of the face, a face recognition algorithm can be used to determine the contour of the facial features and the cheeks.
实施中,在目标图像中,将第二人物替换为显示调整后的第一人物之后,进一步可以包括:In the implementation, after the second character is replaced with the adjusted first character in the target image, the method may further include:
为目标图像中的第一人物添加图像。Add an image to the first person in the target image.
这是便于对替换后的原人物添加道具等图像的,这些道具包括眼镜、帽子、衣服、背包、以及鞋子等。This is convenient for adding images such as props to the original characters after replacement. These items include glasses, hats, clothes, backpacks, and shoes.
实施中,以上根据用户提供的图片或图像序列中检测出第一人物或第二 人物的脸部的位置和区域的方法有很多种,如图3所示。In the implementation, the first person or the second is detected according to the sequence of pictures or images provided by the user. There are many ways to position and area a person's face, as shown in Figure 3.
所列方法中,基于统计模型的方法是目前比较流行的方法,具体可以参见:梁路宏等所著《人脸检测研究综述》(载于计算机学报Vol 25No 5May2002),该方案具有较大的优越性。其优点有:Among the listed methods, the method based on statistical model is a popular method at present. For details, please refer to: “Review of Face Detection Research” by Liang Luhong et al. (Journal of Computers Vol 25No 5May2002), which has great advantages. . Its advantages are:
1、不依赖于人脸的先验知识和参数模型,可以避免不精确或不完整的知识造成的错误;1. A priori knowledge and parameter model that does not depend on the face can avoid errors caused by inaccurate or incomplete knowledge;
2、采用实例学习的方法获取模型的参数,统计意义上更加可靠;2. The method of instance learning is used to obtain the parameters of the model, which is more reliable in statistical sense;
3、通过增加学习的实例可以扩种检测模式范围,提高鲁棒性。3. By increasing the number of learning examples, the range of detection modes can be expanded to improve robustness.
一、统计模型的方法First, the method of statistical model
2001年左右由Viola和Jones提出的基于集成机器学习的人脸检测算法相对于其他方法具有明显优势,具体可参见:艾海舟等所著《人脸检测与检索》(载于自然科学基金项目60273005);武勃等所著《基于连续adaboost算法的多视角人脸检测》(载于计算机研究与发展,2005)。近期文献也表明目前尚未发现优于Viola和Jones方法的其他人脸检测方法,具体可参见:N Degtyarev et al.所著的《Comparative Testing of Face Detection Algorithms》(Image and Signal Processing,2010)。该方法不仅检测精度高,最关键的是其运算速度大大快于其他方法。The face detection algorithm based on integrated machine learning proposed by Viola and Jones around 2001 has obvious advantages over other methods. For details, see: Ai Haizhou et al., "Face Detection and Retrieval" (in the Natural Science Foundation Project 60273005) ); Wu Bo et al., "Multi-view face detection based on continuous adaboost algorithm" (in Computer Research and Development, 2005). Recent literature also indicates that other face detection methods superior to the Viola and Jones methods have not been found yet. For details, see: "Comparative Testing of Face Detection Algorithms" by N Degtyarev et al. (Image and Signal Processing, 2010). This method not only has high detection accuracy, but the most important thing is that its operation speed is much faster than other methods.
Viola和Jones人脸检测方法中几个关键性步骤,具体可参见:Paul Viola and Michael Jones所著《Rapid object detection using a boosted cascade of simple features》(载于Accepted Conference on Computer Vision and Pattern Recognition 2001):Several key steps in the Viola and Jones face detection methods can be found in: "Rapid object detection using a boosted cascade of simple features" by Paul Viola and Michael Jones (available at Accepted Conference on Computer Vision and Pattern Recognition 2001). :
1.提取Haar-like特征(Haar-like features,哈尔特征)1. Extract Haar-like features (Haar-like features)
Haar-like型特征是Viola等人提出的一种简单矩形特征,因为类似Haar小波而得名。Haar型特征的定义是黑色矩形和白色矩形在图像子窗口中对应的区域的权重灰度级总和之差。如图4所示,显示了两种最简单的特征算子。图4中可以看到,在人脸特定结构处,算子计算得到较大的值。 The Haar-like feature is a simple rectangular feature proposed by Viola et al., which is named after the Haar wavelet. The Haar type feature is defined as the difference between the sum of the weight gray levels of the corresponding areas of the black rectangle and the white rectangle in the image sub-window. As shown in Figure 4, the two simplest feature operators are shown. As can be seen in Figure 4, at the specific structure of the face, the operator calculates a larger value.
2.计算积分图2. Calculate the integral map
算子数量庞大时上述计算量显得太大,Viola等人发明了积分图方法,使得计算速度大大加快。如图5所示,点1处的值为A区域的像素积分,点2处的值为AB区域的像素积分。对整张图片进行一次积分操作,便可以方便的计算出任一区域D像素积分值为4+1-2-3。When the number of operators is large, the above calculation is too large. Viola et al. invented the integral graph method, which greatly speeds up the calculation. As shown in FIG. 5, the value at point 1 is the pixel integral of the A region, and the value at the point 2 is the pixel integral of the AB region. By performing an integration operation on the entire picture, it is convenient to calculate the D pixel integral value of any region as 4+1-2-3.
3.训练Adaboost模型3. Training the Adaboost model
在离散Adaboost算法中,Haar-like特征算子计算结果减去某阈值,便可视为一个人脸检测器。因为其准确率不高,称为弱分类器。Adaboost算法的循环中,首先利用各种弱分类器对训练图片库进行分类,准确度最高的弱分类器保留下来,同时提高判断错误的图片的权重,进入下一循环。最终将每次循环所保留的弱分类器组合起来,成为一个准确的人脸检测器,称为强分类器。具体计算流程见,具体可参见:武勃等所著《基于连续adaboost算法的多视角人脸检测》(载于计算机研究与发展,2005);Paul Viola and Michael Jones所著《Rapid object detection using a boosted cascade of simple features》(载于Accepted Conference on Computer Vision and Pattern Recognition 2001)。In the discrete Adaboost algorithm, the Haar-like feature operator subtracts a certain threshold from the calculation result, which can be regarded as a face detector. Because its accuracy is not high, it is called a weak classifier. In the loop of the Adaboost algorithm, the training picture library is first classified by various weak classifiers, and the weak classifier with the highest accuracy is retained, and the weight of the picture that determines the error is increased, and the next cycle is entered. Finally, the weak classifiers retained in each cycle are combined to become an accurate face detector called a strong classifier. For details of the calculation process, see: "Multi-view face detection based on continuous adaboost algorithm" by Wu Bo et al. (in Computer Research and Development, 2005); "Rapid object detection using a by Paul Viola and Michael Jones" Boosted cascade of simple features (available at Accepted Conference on Computer Vision and Pattern Recognition 2001).
4.建立瀑布型级联检测器4. Establish a waterfall cascade detector
瀑布型级联检测器是针对人脸检测速度问题提出的一种检测结构。如图6所示,瀑布的每一层是一个由adaboost算法训练得到的强分类器。设置每层的阈值,使得大多数人脸图像能够通过,在此基础上尽量抛弃反例。位置越靠后的层越复杂,具有越强的分类能力。The waterfall cascade detector is a detection structure proposed for the face detection speed problem. As shown in Figure 6, each layer of the waterfall is a strong classifier trained by the adaboost algorithm. Set the threshold of each layer so that most face images can pass, and on this basis, try to discard the counterexamples. The more complex the layer is, the more complex the classification is.
这样的检测器结构就想一系列筛孔大小递减的筛子,每一步都能筛除一些前面筛子漏下的反例,最终通过所有筛子的样本被接受为人脸。瀑布型检测器训练算法,具体可参见:武勃等所著《基于连续adaboost算法的多视角人脸检测》(载于计算机研究与发展2005)。Such a detector structure would like to have a series of screens with decreasing mesh size. Each step can screen out some of the negative examples of the front screen, and finally the samples passing through all the screens are accepted as human faces. Waterfall type detector training algorithm, see: Wu Bo et al. "Multi-view face detection based on continuous adaboost algorithm" (in Computer Research and Development 2005).
以上算法实现上,采用OpenCV(Open Source Computer Vision Library,开源计算机视觉库)人脸检测程序流程,具体程序源代码可参见如下网址所 记载:http://www.opencv.org.cn/index.php/%E4%BA%BA%E8%84%B8%E6%A3%80%E6%B5%8B。In the above algorithm implementation, OpenCV (Open Source Computer Vision Library) face detection program flow is adopted. The specific program source code can be found at the following website. Record: http://www.opencv.org.cn/index.php/%E4%BA%BA%E8%84%B8%E6%A3%80%E6%B5%8B.
OpenCV是一个基于(开源)发行的跨平台计算机视觉库,可以运行在Linux、Windows和Mac OS操作系统上。它轻量级而且高效-由一系列C函数和少量C++类构成,同时提供了Python、Ruby、MATLAB等语言的接口,实现了图像处理和计算机视觉方面的很多通用算法。OpenCV is a cross-platform computer vision library based on (open source) distribution that runs on Linux, Windows and Mac OS operating systems. It is lightweight and efficient - it consists of a series of C functions and a small number of C++ classes, and provides interfaces to languages such as Python, Ruby, and MATLAB, and implements many general-purpose algorithms for image processing and computer vision.
OpenCV的人脸检测程序采用了Viola和Jones人脸检测方法,主要是调用训练好的瀑布级联分类器cascade来进行模式匹配。OpenCV's face detection program uses Viola and Jones face detection methods, mainly to call the trained waterfall cascade classifier cascade for pattern matching.
cvHaarDetectObjects,先将图像灰度化,根据传入参数判断是否进行canny边缘处理(默认不使用),再进行匹配。匹配后收集找出的匹配块,过滤噪声,计算相邻个数如果超过了规定值(传入的min_neighbors)就当成输出结果,否则删去。cvHaarDetectObjects, first grayscale the image, determine whether to perform canny edge processing (not used by default) according to the incoming parameters, and then match. After matching, the found matching blocks are collected, and the noise is filtered. If the number of adjacent ones exceeds the specified value (incoming min_neighbors), the result is output, otherwise it is deleted.
匹配循环:将匹配分类器放大scale(传入值)倍,同时原图缩小scale倍,进行匹配,直到匹配分类器的大小大于原图,则返回匹配结果。匹配的时候调用cvRunHaarClassifierCascade来进行匹配,将所有结果存入CvSeq*Seq(可动态增长元素序列),将结果传给cvHaarDetectObjects。Matching loop: The matching classifier is enlarged by the scale (input value), and the original image is scaled down to match, until the matching classifier is larger than the original image, and the matching result is returned. When matching, cvRunHaarClassifierCascade is called to match, all results are stored in CvSeq*Seq (a sequence of dynamically growing elements), and the result is passed to cvHaarDetectObjects.
cvRunHaarClassifierCascade函数整体是根据传入的图像和cascade来进行匹配。并且可以根据传入的cascade类型不同(树型、stump(不完整的树)或其他的),进行不同的匹配方式。The cvRunHaarClassifierCascade function is based on the incoming image and cascade. And different matching methods can be performed according to the type of cascade type (tree type, stump (incomplete tree) or other).
函数cvRunHaarClassifierCascade用于对单幅图片的检测。在函数调用前首先利用cvSetImagesForHaarClassifierCascade设定积分图和合适的比例系数(=>窗口尺寸)。当分析的矩形框全部通过级联分类器每一层的时返回正值(这是一个候选目标),否则返回0或负值。The function cvRunHaarClassifierCascade is used to detect a single image. First use cvSetImagesForHaarClassifierCascade to set the integral map and the appropriate scale factor (=> window size) before the function call. A positive value is returned when the analyzed rectangular boxes all pass through each layer of the cascade classifier (this is a candidate target), otherwise 0 or a negative value is returned.
其中分类器的训练采用哈尔分类器,Haar分类器的训练是独立于人脸检测过程的。分类器的训练分为两个阶段: The training of the classifier adopts the Hal classifier, and the training of the Haar classifier is independent of the face detection process. The training of the classifier is divided into two phases:
a.创建样本,用OpenCV自带的creatsamples.exe完成;a. Create a sample, complete with the creatsamples.exe that comes with OpenCV;
b.训练分类器,生成xml文件,由OpenCV自带的haartraining.exe完成。b. Train the classifier to generate the xml file, which is done by haartraining.exe that comes with OpenCV.
训练过程,具体可参见如下1和2:The training process can be seen in the following 1 and 2:
1、http://034080116.blog.163.com/blog/static/334061912009641073715/;1. http://034080116.blog.163.com/blog/static/334061912009641073715/;
2、\OpenCV\apps\HaarTraining\doc\haartraining.doc;2.\OpenCV\apps\HaarTraining\doc\haartraining.doc;
以上地址中,地址1可以在博客中看到,地址2提供的哈尔训练的源文件可以在下载安装后的openCVS安装包目录中找到。In the above address, address 1 can be seen in the blog, and the source file of the Hal training provided by address 2 can be found in the openCVS installation package directory after downloading and installing.
同时,OpenCV中采用的训练算法adaboost是gentle adaboost,为最适合人脸检测的方案。具体可参见:At the same time, the training algorithm adaboost used in OpenCV is gentle adaboost, which is the most suitable solution for face detection. For details, please refer to:
1、http://www.opencv.org.cn/forum/viewtopic.php?f=1&t=4264#p152581. http://www.opencv.org.cn/forum/viewtopic.php? f=1&t=4264#p15258
2、http://www.opencv.org.cn/forum/viewtopic.php?t=38802. http://www.opencv.org.cn/forum/viewtopic.php? t=3880
举例来说,在检测的人脸区域内,确定人脸的五官区域,位置关系及脸颊的轮廓信息,如:眼睛、鼻子、眉毛、嘴和耳朵等,可以通过很多算法实现。本发明优先用ASM算法,以下将对ASM算法进行介绍。For example, in the detected face area, determining the facial features of the face, the positional relationship, and the outline information of the cheeks, such as eyes, nose, eyebrows, mouth, and ears, can be implemented by many algorithms. The invention preferentially uses the ASM algorithm, and the ASM algorithm will be introduced below.
ASM是一种基于分布模型(Point Distribution Model,PDM)的算法,在PDM中,外形相似的物体,例如人脸、人手、心脏、肺部等的几何形状可以通过若干关键特征点(landmarks)的坐标依次串连形成一个形状向量来表示。本发明实施例就以人脸为例来介绍该算法的基本原理和方法。首先给出一个标定好68个关键特征点的人脸面部图片,如图6所示。ASM在实际应用过程中,包括训练和搜索两个部分。ASM is a model based on the Point Distribution Model (PDM). In PDM, geometric shapes of similarly shaped objects such as faces, hands, hearts, lungs, etc. can pass through several key landmarks. The coordinates are successively connected in series to form a shape vector to represent. The embodiment of the present invention introduces the basic principle and method of the algorithm by taking a human face as an example. First, a face image with 68 key feature points is given, as shown in Figure 6. In the actual application process, ASM includes two parts: training and search.
一、ASM的训练First, ASM training
ASM训练包括两个部分。ASM training consists of two parts.
1、建立形状模型:该部分由以下几个步骤组成1. Create a shape model: this part consists of the following steps
1.1搜集n个训练样本1.1 Collect n training samples
如果需要对人脸的面部关键区域进行ASM训练,就需要涉及n个含有人脸面部区域的样本图片。需要提醒的是,搜集的图片只要里面含有人脸面部区 域就可以了,这里不用考虑图像尺寸的归一化等问题。If you need to perform ASM training on the face key area of the face, you need to involve n sample images containing the face area. Need to be reminded that the collected images as long as they contain facial and facial areas The domain is fine, and there is no need to consider the normalization of the image size.
1.2手动记录下每个训练样本中的k个关键特征点1.2 Manually record k key feature points in each training sample
如图7所示,对于训练集中任意一个图片而言,需要记录下若干个(图7中是68个)关键特征点的位置坐标信息,并在文本文件中将该坐标信息保存。该步骤可通过程序员编写的程序完成。程序每次加载一张训练样本,用户依次点击图片中的关键特征点,每点击一次,程序自动记录下当前鼠标点击的位置坐标,予以保存,供后面使用。As shown in FIG. 7, for any picture in the training set, it is necessary to record the position coordinate information of several (68 in FIG. 7) key feature points, and save the coordinate information in the text file. This step can be done by a program written by the programmer. Each time the program loads a training sample, the user clicks on the key feature points in the picture. Each time the program is clicked, the program automatically records the coordinates of the current mouse click position and saves it for later use.
1.3构建训练集的形状向量1.3 Building the shape vector of the training set
将一副图中标定的k个关键特征点组成一个形状向量。The k key feature points calibrated in a picture are combined into a shape vector.
Figure PCTCN2015077353-appb-000001
式(1)
Figure PCTCN2015077353-appb-000001
Formula 1)
其中,
Figure PCTCN2015077353-appb-000002
表示第i个训练样本上第j个特征点的坐标,n表示训练样本的个数。如此一来,n个训练样本,就构成了n个形状向量。
among them,
Figure PCTCN2015077353-appb-000002
Indicates the coordinates of the jth feature point on the i-th training sample, and n represents the number of training samples. In this way, n training samples constitute n shape vectors.
1.4形状归一化1.4 shape normalization
该步骤的目的在于对前面手动标定的人脸形状进行归一化或者对齐操作,从而消除图片中人脸由于不同角度、距离远近、姿态变换等外界因素造成的非形状干扰,从而使得点分布模型更加有效。一般来说,该步骤都采用Procrustes方法进行归一化。简单来说,该方法就是把一系列的点分布模型通过适当的平移、旋转、缩放变换,在不改变点分布模型的基础上对齐到同一个点分布模型,从而改变获取的原始数据杂乱无章的状态,减少非形状因素的干扰。利用Procrustes方法对π={α1,α2,...,αn}这个训练集进行对齐的过程,需要对其中的每个αi计算的参数有4个:旋转角度旋θi,缩放尺度si,水平方向平移量
Figure PCTCN2015077353-appb-000003
垂直方向平移量
Figure PCTCN2015077353-appb-000004
令M(si,θi)[αi]表示对αi做一个旋转角度为θi,缩放尺度为si的变换。αi向αk对齐的过程就是求θi,si
Figure PCTCN2015077353-appb-000005
使得
Figure PCTCN2015077353-appb-000006
最小化的过程。其中
Figure PCTCN2015077353-appb-000007
这里的 W是一个对角矩阵,它可以通过下面的计算来得到:令
Figure PCTCN2015077353-appb-000008
表示一副图像中第k个点和第1个点之间的距离,令
Figure PCTCN2015077353-appb-000009
表示整个训练集中不同图像之间
Figure PCTCN2015077353-appb-000010
的方差,通过计算
Figure PCTCN2015077353-appb-000011
从而得到:
Figure PCTCN2015077353-appb-000012
不难发现,Procrustes方法只是一种求解变换矩阵的方法。而ASM中,正是利用了Procrustes进行点分布模型的对齐操作,具体步骤如下:
The purpose of this step is to normalize or align the face shape manually calibrated in front, so as to eliminate the non-shape interference caused by external factors such as different angles, distances, and posture changes in the picture, so that the point distribution model More effective. In general, this step is normalized using the Procrustes method. To put it simply, the method is to align a series of point distribution models with appropriate translation, rotation, and scaling transformations to the same point distribution model without changing the point distribution model, thereby changing the disordered state of the acquired raw data. To reduce interference from non-shape factors. Using the Procrustes method to align the training set of π={α 1 , α 2 ,..., α n }, there are four parameters that need to be calculated for each α i : rotation angle rotation θ i , scaling Scale s i , horizontal shift
Figure PCTCN2015077353-appb-000003
Vertical shift amount
Figure PCTCN2015077353-appb-000004
Let M(s i , θ i )[α i ] denote a transformation of α i with a rotation angle of θ i and a scaling scale of s i . The process of aligning α i to α k is to find θ i , s i ,
Figure PCTCN2015077353-appb-000005
Make
Figure PCTCN2015077353-appb-000006
Minimize the process. among them
Figure PCTCN2015077353-appb-000007
Here W is a diagonal matrix, which can be obtained by the following calculation:
Figure PCTCN2015077353-appb-000008
Indicates the distance between the kth point and the 1st point in an image,
Figure PCTCN2015077353-appb-000009
Represents the entire training set between different images
Figure PCTCN2015077353-appb-000010
Variance by calculation
Figure PCTCN2015077353-appb-000011
Thereby getting:
Figure PCTCN2015077353-appb-000012
It is not difficult to find that the Procrustes method is just a way to solve the transformation matrix. In ASM, Procrustes is used to perform the alignment of the point distribution model. The specific steps are as follows:
(1)将训练集中的所有人脸模型对齐到第1个人脸模型;(1) Aligning all face models in the training set to the first face model;
(2)计算平均人脸模型
Figure PCTCN2015077353-appb-000013
(2) Calculate the average face model
Figure PCTCN2015077353-appb-000013
(3)将所有人脸模型对齐到平均人脸模型
Figure PCTCN2015077353-appb-000014
(3) Align all face models to the average face model
Figure PCTCN2015077353-appb-000014
(4)重复(2),(3)直到收敛。(4) Repeat (2), (3) until convergence.
1.5将对齐后的形状向量进行PCA处理1.5 The aligned shape vector is processed by PCA
(1)计算平均形状向量:(1) Calculate the average shape vector:
Figure PCTCN2015077353-appb-000015
式(2)
Figure PCTCN2015077353-appb-000015
Formula (2)
(2)计算协方差矩阵:(2) Calculate the covariance matrix:
Figure PCTCN2015077353-appb-000016
式(3)
Figure PCTCN2015077353-appb-000016
Formula (3)
(3)计算协方差矩阵S的特征值并将其按从大到小依次排序:(3) Calculate the eigenvalues of the covariance matrix S and sort them in order from large to small:
这样,便得到λ1,λ2,...,λq,其中λ1>0。选择前t个特征向量P=(p1,p2,...,pt)使得与其对应的特征值满足:Thus, λ 1 , λ 2 , ..., λ q are obtained , where λ 1 >0. The first t eigenvectors P=(p 1 , p 2 , . . . , p t ) are selected such that the corresponding eigenvalues satisfy:
Figure PCTCN2015077353-appb-000017
式(4)
Figure PCTCN2015077353-appb-000017
Formula (4)
这里的fv是一个由特征向量个数来确定的比例系数,通常取值为95%,而VT是所有特征之和。即: Here f v is a scale factor determined by the number of feature vectors, usually 95%, and V T is the sum of all features. which is:
VT=∑λi V T =∑λ i
这样任何一个用于训练的形状向量都可以被表示为:Thus any shape vector used for training can be expressed as:
Figure PCTCN2015077353-appb-000018
式(5)
Figure PCTCN2015077353-appb-000018
Formula (5)
上面的式子当中,bs是包含了t个参数的向量,其中,In the above formula, b s is a vector containing t parameters, where
Figure PCTCN2015077353-appb-000019
Figure PCTCN2015077353-appb-000019
另外,为了确保由bs的变化产生的形状与训练集中的形状类似,需要对bs进行一些限制,即In addition, in order to ensure that the shape resulting from the change in b s is similar to the shape in the training set, some restrictions on b s are required, namely
Figure PCTCN2015077353-appb-000020
Figure PCTCN2015077353-appb-000020
其中Dmax通常为3,如果b在更新过程中Dm>Dmax,则使用Where D max is usually 3, if b is D m >D max during the update process, then use
Figure PCTCN2015077353-appb-000021
Figure PCTCN2015077353-appb-000021
对bs加以约束。Bind b s .
2、为每个特征点构建局部特征2. Construct local features for each feature point
为了能在每一次迭代过程中为每个特征点寻找其新的位置,需要为它们分别建立局部特征。对于第i个特征点,其局部特征的创建过程如图7所示,在第i个训练图像上的第i个特征点的两侧,沿着垂直于该点前后两个特征点连线的方向上分别选择m个像素以构成一个长度为2m+1的向量,对该向量所包含的像素的灰度值求导得到一个局部纹理gij,对训练集中其他训练样本图像上的第i特征点进行同样的操作,便可得到第i个特征点的n个局部纹理gi1,gi2,...,gin。然后,求取它们的均值:In order to be able to find new locations for each feature point during each iteration, local features need to be created separately for them. For the i-th feature point, the creation process of the local feature is as shown in FIG. 7. On both sides of the i-th feature point on the i-th training image, along the line connecting the two feature points perpendicular to the point Select m pixels in the direction to form a vector with length 2m+1, and derive the gray value of the pixel included in the vector to obtain a local texture g ij for the i-th feature on other training sample images in the training set. By performing the same operation, the n local textures g i1 , g i2 , . . . , g in of the i-th feature point are obtained. Then, find their mean:
Figure PCTCN2015077353-appb-000022
式(6)
Figure PCTCN2015077353-appb-000022
Formula (6)
以及方差:And variance:
Figure PCTCN2015077353-appb-000023
式(7)
Figure PCTCN2015077353-appb-000023
Formula (7)
这样就得到了第i个特征点的局部特征。对其他所有的特征点进行相同的操作,就可得到每个特征点的局部特征。这样,一个特征点的新的特征g与其训练好的局部特征之间的相似性度量就可以用马氏距离来表示:This gives the local features of the i-th feature point. By performing the same operation on all other feature points, the local features of each feature point are obtained. Thus, the similarity measure between the new feature g of a feature point and its trained local feature can be expressed in terms of Mahalanobis distance:
Figure PCTCN2015077353-appb-000024
式(8)
Figure PCTCN2015077353-appb-000024
Formula (8)
二、ASM的搜索Second, ASM search
在通过样本集进行训练得到ASM模型建立后即可进行ASM搜索,首先对平均形状进行仿射变换得到一个初始模型:After the ASM model is established through the training of the sample set, the ASM search can be performed. First, the average shape is affine transformed to obtain an initial model:
X=M(s,θ)[αi]+Xc式(9)X=M(s,θ)[α i ]+X c (9)
上面的式子表示对平均形状以其中心逆时针旋转θ缩放S,然后再平移Xc得到初始模型X。The above expression represents scaling the S to the average shape with its center counterclockwise rotation θ, and then shifting X c to obtain the initial model X.
用该初始模型在新的图像中搜索目标形状,使搜索到的最终形状中的特征点和相对应的真正特征点最为接近,这个搜索过程主要是通过仿射变换和参数b的变化来实现。具体算法可以通过反复如下两步来实现:The initial model is used to search for the target shape in the new image, so that the feature points in the searched final shape are closest to the corresponding real feature points. This search process is mainly realized by the affine transformation and the change of the parameter b. The specific algorithm can be implemented by repeating the following two steps:
2.1计算每个特征点的新位置2.1 Calculate the new position of each feature point
首先把初始ASM模型覆盖在图像上,如图8所示,First overlay the initial ASM model on the image, as shown in Figure 8.
对于模型中第i个特征点,在垂直于其前后两个特征点连线方向上以其为中心两边各选择1(1>m)个像素,然后计算这1个像素的灰度值导数并归一化从而得到一个局部特征,其包括2(1-m)+1个子局部特征,然后利用前面的公式计算这些子局部特征与当前特征点的局部特征之间的马氏距离,使得马氏距离最小的那个子局部特征的中心即为当前特征点的新位置,这样就会产生一个位移。为所有的特征点找到其新位置,并把它们的位移组成一个向量:For the i-th feature point in the model, select 1 (1>m) pixels on both sides of the two feature points perpendicular to the front and rear of the feature points, and then calculate the gray value derivative of the one pixel. Normalization to obtain a local feature, including 2(1-m)+1 sub-local features, and then using the previous formula to calculate the Mahalanobis distance between these sub-local features and the local features of the current feature points, so that Markov The center of the smallest sub-local feature is the new position of the current feature point, which produces a displacement. Find their new locations for all feature points and group their displacement into a vector:
dX=(dX1,dX2,...dXk)。dX = (dX 1 , dX 2 , ... dX k ).
2.2仿射变化中的参数和b的更新2.2 Parameters in the affine change and update of b
通过仿射变换并调整其参数使得当前特征点的位置X与对应的新的位置X+dX最为接近。 The position X of the current feature point is closest to the corresponding new position X+dX by affine transformation and adjusting its parameters.
仿射变换后便可以得到仿射变换参数的变化量
Figure PCTCN2015077353-appb-000025
同时由式(9)得:
After the affine transformation, the amount of change in the affine transformation parameters can be obtained.
Figure PCTCN2015077353-appb-000025
At the same time, it is obtained by formula (9):
M(s(1+ds),(θ+dθ))[αi+dαi]+(Xc+dXc)式(10)M(s(1+ds), (θ+dθ))[α i +dα i ]+(X c +dX c ) (10)
同时X又可以由(9)表示,因此,上式又可以表示为:At the same time, X can be represented by (9). Therefore, the above formula can be expressed as:
M(s(1+ds),(θ+dθ))[αi+dai]=M(S,θ)[αi]+dX+Xc-(Xc=dXC)式(11)M(s(1+ds),(θ+dθ))[α i +da i ]=M(S,θ)[α i ]+dX+Xc-(Xc=dX C ) (11)
同时由式(9)可得:At the same time, it can be obtained by formula (9):
M-1(s,θ)=M(s-1,θ)式(12)M -1 (s,θ)=M(s -1 ,θ) (12)
由式(11)以及式(12)可得:From equation (11) and equation (12):
i=M(s(1+ds)-1,-(θ+dθ))[M(S,θ)+dX-dXc]-α式(13)i =M(s(1+ds) -1 , -(θ+dθ))[M(S,θ)+dX-dXc]-α (13)
同时由式(5)可得:At the same time, it can be obtained by formula (5):
Figure PCTCN2015077353-appb-000026
式(14)
Figure PCTCN2015077353-appb-000026
Formula (14)
用式(14)减去式(5)可得:Subtracting equation (5) from equation (14) yields:
i≈P×db式(15)i ≈P×db (15)
即:which is:
db=P-1i式(16)Db=P -1i (16)
db=PTi式(17)Db=P Ti (17)
结合式(17)和式(13)可以求得db。因此,上述的参数更新过程为:Combining equations (17) and (13) can find db. Therefore, the above parameter update process is:
Figure PCTCN2015077353-appb-000027
所以可以对仿射变换参数和b做如下更新:Xc=Xc+wtdXc,Yc=Yc+wtdYcθ=θ+wθdθ,s=s(1+wsds),b=b+wbdb式(18)
Figure PCTCN2015077353-appb-000027
Therefore, the affine transformation parameters and b can be updated as follows: X c = X c + w t dX c , Y c = Y c + w t dY c θ = θ + w θ dθ, s = s (1 + w s Ds), b=b+w b db (18)
上面的式子中wt,wθ,ws,Wb是用于控制参数变化的权值。这样就可以由式(5)和式(9)得到新的形状。当仿射变换的参数和b的变化不是很大或者迭代次数达到指定的阈值就结束该搜索过程。检测结果如图9所示。 In the above formula, w t , w θ , w s , W b are weights used to control parameter changes. Thus, a new shape can be obtained from the equations (5) and (9). The search process ends when the parameters of the affine transformation and the change in b are not very large or the number of iterations reaches a specified threshold. The test results are shown in Figure 9.
将检测到的五官的区域和脸颊的轮廓,贴合到一个已有的人脸3维模型上,使其能够自动的根据参数的设定,呈现不同的视角、光照和表情的变化。其具体实现方法如下:The detected facial features and cheeks are attached to an existing 3D model of the face, enabling them to automatically display different perspectives, illuminations, and expression changes based on parameter settings. The specific implementation method is as follows:
选用“BJUT-3D Face Database”三维人脸库,经重采样、平滑及坐标校正等预处理,选择100个男性和100个女性每人约60000个点和120000个三角片的数据作为稠密人脸样本集。然后通过手工交互选取每人60个三维特征点,作为稀疏对应的样本集,并使用这200人的平均模型作为一般模型。Select “BJUT-3D Face Database” 3D face database, pre-sampling, smoothing and coordinate correction, etc., and select data of about 60,000 points and 120,000 triangles for 100 males and 100 females as dense faces. Sample set. Then, 60 three-dimensional feature points per person are selected by manual interaction as a sparsely corresponding sample set, and the average model of 200 people is used as a general model.
重建分以下四个步骤,如图10所示:The reconstruction is divided into the following four steps, as shown in Figure 10:
a.由ASM模板检测人脸特征点。采用改进的ASM算法。自动提取其60个特征点;a. The face feature points are detected by the ASM template. Adopt the improved ASM algorithm. Automatically extract 60 feature points;
b.利用稀疏的形变模型获取特征点深度信息。利用先验三维人脸统计知识,将三维特征点样本集通过平面投影和线性组合来最优逼近照片的二维特征点,从而获得照片特征点对应的三维坐标。b. Obtain feature point depth information using a sparse deformation model. Using the a priori 3D face statistical knowledge, the 3D feature point sample set is optimally approximated by the planar projection and linear combination to obtain the 3D feature points corresponding to the photo feature points.
c.根据三维特征点的位移将一般人脸模型变形得到特定三维人脸。选择薄板样条插值算法(TPS),具体可参见:BOOKSTEINFL.Principlewarps:thin-platesplines and the decomposition of deformation(载于IEEETranson PAMI198911(6):567-585),将原始模型弹性变形为特定人脸模型。c. Deform the general face model to obtain a specific three-dimensional face according to the displacement of the three-dimensional feature points. Select the thin plate spline interpolation algorithm (TPS). For details, see: BOOKSTEINFL.Principlewarps: thin-platesplines and the decomposition of deformation (in IEEE Transon PAMI198911(6): 567-585), which elastically transforms the original model into a specific face model. .
d.通过纹理映射重建模型的颜色信息。将照片纹理作仿射变换后正交投影到三维模型表面。d. Reconstruct the color information of the model through texture mapping. The photo texture is affine transformed and then orthogonally projected onto the surface of the 3D model.
进一步地,将原人物模型调整替换到目标人物所在的所述图像中之后,还可以包括:Further, after the original character model adjustment is replaced with the image in which the target character is located, the method may further include:
对替换后的原人物添加道具,所述道具包括眼镜、帽子、衣服、背包、以及鞋子。Add items to the replaced original character, including glasses, hats, clothes, backpacks, and shoes.
具体的,当上述自动编辑系统将用户指定的第一人物替换到第二人物的图片或图像序列上后,还可以进一步对替换后的第一人物添加道具。其道具 可以是眼镜、帽子、衣服、背包等。Specifically, after the automatic editing system replaces the first character specified by the user with the picture or image sequence of the second character, the item may be further added to the replaced first character. Its props It can be glasses, hats, clothes, backpacks, etc.
进一步地,将原人物模型调整替换到目标人物所在的所述图像中,还可以包括:Further, replacing the original character model adjustment with the image in which the target character is located may further include:
根据ASM检测的二维特征点,纹理映射的全部特征点落在面部区域以内进行调整。According to the two-dimensional feature points detected by ASM, all the feature points of the texture map fall within the face area and are adjusted.
进一步地,所述纹理映射的全部特征点,还可以包括:Further, all the feature points of the texture mapping may further include:
使用的特征点经过肤色模型校正。The feature points used are corrected by the skin color model.
举例来说,本发明在模型重建过程中使用ASM检测的二维特征点,而纹理映射使用的特征点需要经过基于肤色模型的校正,使全部特征点落在脸部区域以内,从而避免纹理映射时侧面纹理的缺失。For example, the present invention uses ASM to detect two-dimensional feature points in the model reconstruction process, and the feature points used in texture mapping need to be corrected based on the skin color model, so that all feature points fall within the face region, thereby avoiding texture mapping. The absence of the side texture.
1)肤色点判定1) Skin color point determination
以YUV和YIQ空间为基础并加入Gamma矫正减少光照对图像质量影响的方法,具体可参见:CHEN Lu,YANG Jie所著的《Automatic 3D face model reconstruction using》,来进行肤色信息的检测。Based on the YUV and YIQ space and adding Gamma correction to reduce the effect of illumination on image quality, see: CHEN Lu, YANG Jie's "Automatic 3D face model reconstruction using" to detect skin color information.
在YUV空间中,U和V是平面上两个相互正交的矢量,色度信号(即U与V之和)是一个二维矢量,称之为色度信号矢量,并且每一种颜色对应一个色度信号矢量,它的饱和度由模值Ch表示,色调由相位角θ表示:In the YUV space, U and V are two mutually orthogonal vectors on the plane, and the chrominance signal (ie, the sum of U and V) is a two-dimensional vector called a chrominance signal vector, and each color corresponds to A chrominance signal vector whose saturation is represented by the modulus Ch and the hue is represented by the phase angle θ:
Figure PCTCN2015077353-appb-000028
式(19)
Figure PCTCN2015077353-appb-000028
Formula (19)
Figure PCTCN2015077353-appb-000029
式(20)
Figure PCTCN2015077353-appb-000029
Formula (20)
把彩色图像的像素P由RGB空间变换到YUV空间,如果满足条件θp∈[105,150],则P是肤色点。在YIQ空间中,I分量代表从桔黄到蓝绿的色调,I值越小,包含的黄色越多,蓝绿色越少。通过实验和统计分析可确定肤色在YIQ空间内的I值在[20,90]变化。分别对R、G、B三个分量作Gamma矫正,校正后的值分别记为Rgamma、Ggamma、Bgamm: The pixel P of the color image is transformed from the RGB space to the YUV space, and if the condition θ p ∈ [105, 150] is satisfied, P is the skin color point. In the YIQ space, the I component represents a hue from orange to blue-green, and the smaller the I value, the more yellow it contains, and the less blue-green. Through experimental and statistical analysis, it can be determined that the I value of skin color in the YIQ space changes in [20, 90]. The gamma correction is performed on the three components R, G and B respectively. The corrected values are recorded as Rgamma, Ggamma, Bgamm:
U=-0.147×Rgamma-0.289×Ggamma+0.436×Bgamma式(21)U=-0.147×R gamma -0.289×G gamma +0.436×B gamma (21)
V=-0.615×Rgamma-0.515×Ggamma-0.100×Bgamma式(22)V=-0.615×R gamma -0.515×G gamma -0.100×B gamma (22)
Figure PCTCN2015077353-appb-000030
式(23)
Figure PCTCN2015077353-appb-000030
Equation (23)
I=0.596×Rgamma-0.274×Ggamma-0.322×Bgamma式(24)I=0.596×R gamma -0.274×G gamma -0.322×B gamma (24)
根据求得的和值,判定该像素点为肤色点。若满足Based on the obtained sum value, it is determined that the pixel is a skin color point. If satisfied
Figure PCTCN2015077353-appb-000031
式(25)
Figure PCTCN2015077353-appb-000031
Equation (25)
则判定该像素点为肤色点。Then, the pixel is determined to be a skin color point.
2)校正特征点2) Correction feature points
由于ASM模板采用对称模板,对于不完全正面的人脸特征提取会出现一边侧面特征点的出界,进而对后面的纹理重建造成侧面信息缺失。对侧面特征点进行肤色点判定,若不是肤色点,则落在脸部以外,将该点向脸中心缩进,直到所有侧面特征点都成为肤色点。Since the ASM template adopts a symmetric template, the face feature points of one side will be out of bounds for the face feature extraction which is not completely positive, and the side information is missing for the subsequent texture reconstruction. The skin color point is determined for the side feature points, and if it is not the skin color point, it falls outside the face, and the point is indented toward the center of the face until all the side feature points become skin color points.
3)使用校正后特征点进行纹理映射3) Texture mapping using corrected feature points
由于模型重建必须使用对称的特征点,仍然使用校正前的二维特征点来计算三维特征点,最后得到模型。贴图时使用校正后的特征点来映射三维特征点,这样有效地避免了侧面的纹理缺失。Since the model reconstruction must use symmetric feature points, the two-dimensional feature points before correction are still used to calculate the three-dimensional feature points, and finally the model is obtained. The map features the corrected feature points to map the 3D feature points, which effectively avoids the side texture loss.
由该模型所生成的不同姿态、光照和表情的三维人脸模型也有较好的真实感。如图11所示,原始输入的人脸图像和生成的人脸三维模型,为合成丰富的人脸表情,基于面部运动编码系统(Facial Action Coding System,FACS)设立44个基本动作单元(Action Unit,AU),每个AU可以控制一个或几个人脸特征点在三维空间的位移。将不同的AU组合,可产生喜怒哀乐等各种表情。使用TPS对三维特征点进行插值变形,实现表情变化,如图12所示,模拟的表情实例。 The three-dimensional face model of different poses, illuminations and expressions generated by the model also has a good sense of reality. As shown in FIG. 11, the original input face image and the generated face three-dimensional model are rich synthetic facial expressions, and 44 basic action units (Action Units) are established based on the Facial Action Coding System (FACS). , AU), each AU can control the displacement of one or several face feature points in three dimensions. Combining different AUs can produce various expressions such as emotions and sorrows. The TPS is used to interpolate and deform the three-dimensional feature points to realize the expression change, as shown in Fig. 12, the simulated expression examples.
通过检测到的五官的区域和脸颊的轮廓,推断第二人物的相关特性信息。这些特性信息包括视角、光照和表情的变化等。如图13所示,具体实施方案如下。The relevant characteristic information of the second person is inferred from the detected area of the facial features and the outline of the cheek. These characteristics include changes in perspective, lighting, and expressions. As shown in Fig. 13, the specific embodiment is as follows.
当图片或图像序列中的人物的特征点已经确认,我们可以很容易的找到已经事先定义好的AU单元(如上文所述)。这些AU可以很细化地描述一个脸部的表情,本算法只要确定了特征点的具体位置和AU的具体形态就确定了人物脸部的表情。When the feature points of the characters in the picture or image sequence have been confirmed, we can easily find the AU units that have been defined in advance (as described above). These AUs can describe the expression of a face in a very detailed way. The algorithm determines the expression of the face of the person as long as the specific position of the feature point and the specific shape of the AU are determined.
至于通过二维图像中的人脸特征点估计人物头部的姿态,本算法利用的是POSIT方法。As for estimating the pose of the person's head through the face feature points in the two-dimensional image, the algorithm utilizes the POSIT method.
1、基本思想:算法分两部分1, the basic idea: the algorithm is divided into two parts
(1)带有比例系数的正交投影变换SOP(Standard Operation Procedure,标准作业程序),根据线性方程组求出旋转矩阵和平移向量;(1) SOP (Standard Operation Procedure) with proportional coefficient, and find the rotation matrix and translation vector according to the linear equations;
(2)由得出的旋转矩阵和平移向量系数,更新比例系数(Scale Factor),再由比例系数更新原有的点,进行迭代。(2) From the obtained rotation matrix and translation vector coefficients, update the scale factor (Scale Factor), and then update the original point by the scale factor, and iterate.
2、算法过程:2, the algorithm process:
(1)假设旋转矩阵
Figure PCTCN2015077353-appb-000032
和平移向量
Figure PCTCN2015077353-appb-000033
f是焦距;在透视投影变换中
Figure PCTCN2015077353-appb-000034
而在SOP中,
Figure PCTCN2015077353-appb-000035
其中比例因子是
Figure PCTCN2015077353-appb-000036
(1) Assume rotation matrix
Figure PCTCN2015077353-appb-000032
And translation vector
Figure PCTCN2015077353-appb-000033
f is the focal length; in the perspective projection transformation
Figure PCTCN2015077353-appb-000034
And in the SOP,
Figure PCTCN2015077353-appb-000035
Where the scale factor is
Figure PCTCN2015077353-appb-000036
(2)作基本的透视投影变换,将3D点a=(ax,ay,az)T透视投影到图像平面上得到齐次坐标m=(wx,wy,w)T,变换过程为
Figure PCTCN2015077353-appb-000037
因为m是齐次 坐标,所以等式右边除以Tz,不会受影响,则得到
Figure PCTCN2015077353-appb-000038
其中s=f/Tz,即得到
Figure PCTCN2015077353-appb-000039
其中
Figure PCTCN2015077353-appb-000040
(2) For the basic perspective projection transformation, the 3D point a=(a x , a y , a z ) T is projected onto the image plane to obtain the homogeneous coordinates m=(wx, wy, w) T , and the transformation process is
Figure PCTCN2015077353-appb-000037
Since m is a homogeneous coordinate, the right side of the equation is divided by T z and will not be affected.
Figure PCTCN2015077353-appb-000038
Where s=f/T z is obtained
Figure PCTCN2015077353-appb-000039
among them
Figure PCTCN2015077353-appb-000040
(3)现在变换过程为
Figure PCTCN2015077353-appb-000041
即为方程组
(3) Now the transformation process is
Figure PCTCN2015077353-appb-000041
Equation group
Figure PCTCN2015077353-appb-000042
w初始值为1;
Figure PCTCN2015077353-appb-000042
w initial value is 1;
(4)令K1=(sR11 sR12 sR13 sTx)T,K2=(sR21 sR22 sR23 sTy)T
Figure PCTCN2015077353-appb-000043
A为(n+1)×4矩阵,
Figure PCTCN2015077353-appb-000044
然后初始方程组变成
Figure PCTCN2015077353-appb-000045
应用最小二乘法,得到解
Figure PCTCN2015077353-appb-000046
(4) Let K 1 = (sR 11 sR 12 sR 13 sT x ) T , K 2 = (sR 21 sR 22 sR 23 sT y ) T ,
Figure PCTCN2015077353-appb-000043
A is a (n+1)×4 matrix,
Figure PCTCN2015077353-appb-000044
Then the initial equations become
Figure PCTCN2015077353-appb-000045
Apply the least squares method to get the solution
Figure PCTCN2015077353-appb-000046
(5)至少有4个不共面的2D-3D点对,求出K1,K2后,将其除以已知的定值s,可得到R1,R2,Tx,Ty,然后得到R3=R1×R2,并且将R1,R2,R3归一化为单位向量;(5) At least 4 non-coplanar 2D-3D point pairs. After K1 and K2 are obtained, divide it by the known fixed value s to obtain R1, R2, Tx, Ty, and then get R3=R1. ×R2, and normalize R1, R2, R3 into unit vectors;
(6)然后更新
Figure PCTCN2015077353-appb-000047
因为对不同的2D-3D点对,s=f/Tz是定值,f是焦距,是已知的定值参数,Tz也是已知的定值参数,可以看做是所有的3D点Z坐标的平均值;对不同的3D点,a不同,所以w也就不同,这样将原来的2D点变为(wx,wy)T
(6) then update
Figure PCTCN2015077353-appb-000047
Because for different 2D-3D point pairs, s=f/T z is a fixed value, f is the focal length, which is a known fixed value parameter, and T z is also a known fixed value parameter, which can be regarded as all 3D points. The average of the Z coordinates; for different 3D points, a is different, so w is different, so that the original 2D point becomes (wx, wy) T ;
(7)再从步骤(2)开始,由原有的3D点和更新后的2D点,用最小二乘法解方程组,得到新的K1,K2;再更新w,更新2D点坐标; (7) Starting from step (2), the original 3D point and the updated 2D point are solved by the least squares method to obtain a new K1, K2; then w is updated, and the 2D point coordinates are updated;
3、求解过程:3. Solution process:
(1)给出摄像机的初始位置:焦距f,图像坐标中心,即(cx,cy),图像范围,即合理的2D坐标值范围。(1) Give the initial position of the camera: focal length f, image coordinate center, ie (c x , c y ), image range, ie a reasonable range of 2D coordinate values.
(2)共有8个未知数,至少需要4个2D-3D点对;(2) There are 8 unknowns, at least 4 2D-3D point pairs are required;
(3)第一个2D-3D点对必须是(0,0)-(0,0,0);(3) The first 2D-3D point pair must be (0,0)-(0,0,0);
(3)算法执行停止条件为:限制迭代次数,设置每次2D点的变动值大小(精确度)阈值。(3) The algorithm execution stop condition is: limit the number of iterations, and set the threshold value (accuracy) threshold value of the 2D point each time.
实施中,在第一人物的模型中根据所述特征信息调整第一人物的脸部的显示时,所述特征信息可以为以下参数之一或者其组合:第二人物的脸部的三维3D姿态、第二人物的脸部的基本动作单元AU的状态、第二人物的脸部的轮廓的长宽的比例、第二人物的脸部的特征点周围的皮肤的亮暗程度。In an implementation, when the display of the face of the first person is adjusted according to the feature information in the model of the first person, the feature information may be one of the following parameters or a combination thereof: a three-dimensional 3D gesture of the face of the second person The degree of the basic motion unit AU of the face of the second person, the ratio of the length and width of the contour of the face of the second person, and the degree of lightness and darkness of the skin around the feature point of the face of the second person.
具体的,将第一人物模型进一步替换第二人物,可以如下:Specifically, the first character model is further replaced by the second character, which can be as follows:
a.根据第二人物的相关特征信息,调整第一人物的模型,使其与第二人物的相关特性相仿;如要可以分为:a. according to the relevant feature information of the second character, adjust the model of the first character to be similar to the related characteristics of the second character;
a.1根据估算出来的第二人物的3D人脸的姿态,调整第一人物3D模型的姿态;A.1 adjusting the posture of the first character 3D model according to the estimated posture of the 3D face of the second character;
a.2根据估计出来的第二人物的AU的状态,调整第一人物3D模型的表情;A.2 adjusting the expression of the first character 3D model according to the estimated state of the AU of the second character;
a.3根据第二人物的脸的轮廓,主要是长宽的比例,调整第一人物3D模型的脸型;A.3 adjusting the face shape of the first character 3D model according to the contour of the face of the second character, mainly the ratio of the length to the width;
a.4根据第二人物脸部的所有特征点周围的皮肤的亮暗程度,调整第一人物相应特征点周围的脸部的亮暗。A.4 Adjusting the brightness of the face around the corresponding feature point of the first character according to the brightness of the skin around all the feature points of the second person's face.
b.通过检测到的五官的区域和脸颊的轮廓,将每帧图像中的第二人物的脸部区域抹去;b. erasing the face area of the second person in each frame image by detecting the contour of the facial features and the contour of the cheek;
c.在每帧图像中,将调整好的第一人物的模型放置在第二人物的脸部区域; c. in each frame image, placing the adjusted model of the first person in the face area of the second person;
基于同一发明构思,本发明实施例中还提供了一种图像处理系统,由于该系统所解决问题的原理与一种图像处理方法相似,因此这些系统的实施可以参见方法的实施,重复之处不再赘述。Based on the same inventive concept, an image processing system is further provided in the embodiment of the present invention. Since the principle of the problem solved by the system is similar to an image processing method, the implementation of these systems can be referred to the implementation of the method, and the repetition is not Let me repeat.
图14为图像处理系统结构示意图,如图14所示,可以包括:14 is a schematic structural diagram of an image processing system, as shown in FIG. 14, which may include:
模型模拟模块1401,用于根据至少一张包含第一人物的图像,模拟出第一人物的模型;a model simulation module 1401, configured to simulate a model of the first character according to at least one image including the first person;
目标图像确定模块1402,用于确定包含第二人物的目标图像;a target image determining module 1402, configured to determine a target image that includes the second person;
特征信息确定模块1403,用于确定在目标图像中显示第二人物的特征信息;The feature information determining module 1403 is configured to determine that the feature information of the second person is displayed in the target image;
调整显示模块1404,用于在第一人物的模型中根据所述特征信息调整第一人物的显示;Adjusting the display module 1404, for adjusting the display of the first character according to the feature information in the model of the first character;
人物替换模块1405,用于在目标图像中,将第二人物替换为显示调整后的第一人物。The character replacement module 1405 is configured to replace the second character with the adjusted first character in the target image.
实施中,模型模拟模块1401可以包括:In an implementation, the model simulation module 1401 can include:
第一检测单元,用于检测出第一人物的脸部的区域;a first detecting unit, configured to detect an area of a face of the first person;
第一确定单元,用于在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;a first determining unit, configured to determine an outline of the facial features and a cheek in the detected area of the face;
贴合单元,用于将检测到的五官的区域和脸颊的轮廓,贴合到已有的人脸3D模型上后获得模拟出的第一人物的脸部的模型。The fitting unit is configured to fit the detected facial features and the outline of the cheek to the existing human face 3D model to obtain a model of the simulated first person's face.
实施中,目标图像确定模块1402可以包括:In an implementation, the target image determining module 1402 may include:
第二检测单元,用于检测出第二人物的脸部的区域;a second detecting unit, configured to detect an area of a face of the second person;
第二确定单元,用于在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;a second determining unit, configured to determine an outline of the facial features and a cheek in the detected area of the face;
特征单元,用于根据检测到的五官的区域和脸颊的轮廓,确定在目标图像中显示第二人物的脸部的特征信息。 And a feature unit configured to determine feature information of displaying a face of the second person in the target image according to the detected area of the facial features and the outline of the cheek.
实施中,特征信息确定模块1403进一步用于根据第一人物的脸部的区域与第二人物的脸部的区域将第二人物的脸部替换为显示调整后的第一人物的脸部。In the implementation, the feature information determining module 1403 is further configured to replace the face of the second person with the face of the second person to display the face of the adjusted first person according to the area of the face of the first person and the area of the face of the second person.
实施中,调整显示模块1404进一步用于在第一人物的模型中根据以下参数之一或者其组合的所述特征信息调整第一人物的脸部的显示:第二人物的脸部的3D姿态、第二人物的脸部的AU的状态、第二人物的脸部的轮廓的长宽的比例、第二人物的脸部的特征点周围的皮肤的亮暗程度。In an implementation, the adjustment display module 1404 is further configured to adjust, in the model of the first character, the display of the face of the first person according to the feature information of one of the following parameters or a combination thereof: a 3D posture of the face of the second person, The degree of the AU of the face of the second person, the aspect ratio of the outline of the face of the second person, and the degree of lightness and darkness of the skin around the feature point of the face of the second person.
实施中,调整显示模块1404进一步用于在检测出的脸部的区域里,采用人脸识别算法确定五官的区域和脸颊的轮廓。In an implementation, the adjustment display module 1404 is further configured to determine a contour of the facial features and the cheeks using a face recognition algorithm in the detected area of the face.
实施中,进一步可以包括:In implementation, further can include:
道具添加模块,用于在目标图像中,将第二人物替换为显示调整后的第一人物之后,为目标图像中的第一人物添加图像。The item adding module is configured to add an image to the first person in the target image after replacing the second character with the adjusted first character in the target image.
在本发明实施例提供的技术方案中,模拟出原人物的模型,综合考虑目标人物的特征信息,将原人物模型调整替换到目标人物所在的图像中。解决了图像模拟替换时,人物图像与所替换的拍摄视角图像不相符,及人物表情不可改变的问题。可以应用在很多场景如:友情、爱情、亲子、卡拉ok换脸时,身临其境的将一个人物带到另一个人物所在的环境中;还可以虚拟出一个人物来替代另一个人物来做一些事情;还可以替换掉照片中不喜欢的人物,换成自己喜欢的人物。In the technical solution provided by the embodiment of the present invention, the model of the original character is simulated, and the feature information of the target person is comprehensively considered, and the original character model is adjusted and replaced with the image of the target person. When the image simulation is replaced, the image of the person does not match the image of the captured angle of view, and the expression of the character cannot be changed. Can be applied in many scenes such as: friendship, love, parent-child, karaoke face change, immersive to bring a character to the environment where another character is located; can also virtualize a character to replace another character to do Some things; you can also replace characters you don't like in photos and switch to characters you like.
采用在本发明实施例提供的技术方案,用户只用一幅图片即可,就可以替换或编辑任意照片或视频中的任意人物的面部;替换后,第一人物的面部的拍摄角度可以根据目标人物的拍摄角度的变化而变化;替换后,第一人物的面部的表情可以根据目标人物的表情的变化而变化;还可以身临其境的将第一人物带到目标人物所在的世界里。According to the technical solution provided by the embodiment of the present invention, the user can replace or edit the face of any person in any photo or video by using only one picture; after the replacement, the shooting angle of the face of the first person can be based on the target. The change of the shooting angle of the character changes; after the replacement, the expression of the face of the first character can be changed according to the change of the expression of the target character; and the first character can be brought to the world where the target character is located.
本领域内的技术人员应明白,本发明的实施例可提供为方法、系统、或 计算机程序产品。因此,本发明可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本发明可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art will appreciate that embodiments of the present invention may be provided as a method, system, or Computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment, or a combination of software and hardware. Moreover, the invention can take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) including computer usable program code.
本发明是参照根据本发明实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present invention has been described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (system), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flowchart illustrations and/or FIG. These computer program instructions can be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing device to produce a machine for the execution of instructions for execution by a processor of a computer or other programmable data processing device. Means for implementing the functions specified in one or more of the flow or in a block or blocks of the flow chart.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。The computer program instructions can also be stored in a computer readable memory that can direct a computer or other programmable data processing device to operate in a particular manner, such that the instructions stored in the computer readable memory produce an article of manufacture comprising the instruction device. The apparatus implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded onto a computer or other programmable data processing device such that a series of operational steps are performed on a computer or other programmable device to produce computer-implemented processing for execution on a computer or other programmable device. The instructions provide steps for implementing the functions specified in one or more of the flow or in a block or blocks of a flow diagram.
尽管已描述了本发明的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例作出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本发明范围的所有变更和修改。 While the preferred embodiment of the invention has been described, it will be understood that Therefore, the appended claims are intended to be interpreted as including the preferred embodiments and the modifications and

Claims (14)

  1. 一种图像处理方法,其特征在于,包括如下步骤:An image processing method includes the following steps:
    根据至少一张包含第一人物的图像,模拟出第一人物的模型;Simulating a model of the first character based on at least one image containing the first person;
    确定包含第二人物的目标图像;Determining a target image containing the second person;
    确定在目标图像中显示第二人物的特征信息;Determining to display feature information of the second person in the target image;
    在第一人物的模型中根据所述特征信息调整第一人物的显示;Adjusting the display of the first character according to the feature information in the model of the first character;
    在目标图像中,将第二人物替换为显示调整后的第一人物。In the target image, the second character is replaced with the adjusted first character.
  2. 如权利要求1所述的方法,其特征在于,在根据至少一张包含第一人物的图像,模拟出第一人物的脸部的模型时,包括:The method according to claim 1, wherein when simulating the model of the face of the first person based on the at least one image including the first person, the method comprises:
    检测出第一人物的脸部的区域;Detecting the area of the face of the first person;
    在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;Determining the contours of the facial features and cheeks in the area of the detected face;
    将检测到的五官的区域和脸颊的轮廓,贴合到已有的人脸三维3D模型上后获得模拟出的第一人物的脸部的模型。The model of the face of the first character is obtained by fitting the detected area of the facial features and the contour of the cheek to the existing three-dimensional 3D model of the face.
  3. 如权利要求1或2所述的方法,其特征在于,在确定在目标图像中显示第二人物的脸部的特征信息时,包括:The method according to claim 1 or 2, wherein when determining the feature information of the face of the second person in the target image, the method comprises:
    检测出第二人物的脸部的区域;Detecting the area of the face of the second person;
    在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;Determining the contours of the facial features and cheeks in the area of the detected face;
    根据检测到的五官的区域和脸颊的轮廓,确定在目标图像中显示第二人物的脸部的特征信息。Based on the detected area of the facial features and the outline of the cheek, it is determined that the feature information of the face of the second person is displayed in the target image.
  4. 如权利要求3所述的方法,其特征在于,将第二人物替换为显示调整后的第一人物,是根据第一人物的脸部的区域与第二人物的脸部的区域将第二人物的脸部替换为显示调整后的第一人物的脸部。The method according to claim 3, wherein the replacing the second character with the display of the adjusted first character is based on the area of the face of the first person and the area of the face of the second person The face is replaced with the face of the adjusted first character.
  5. 如权利要求4所述的方法,其特征在于,在第一人物的模型中根据所述特征信息调整第一人物的脸部的显示时,所述特征信息为以下参数之一或者其组合:第二人物的脸部的3D姿态、第二人物的脸部的基本动作单元AU的状态、第二人物的脸部的轮廓的长宽的比例、第二人物的脸部的特征点周 围的皮肤的亮暗程度。The method according to claim 4, wherein, in the model of the first person, when the display of the face of the first person is adjusted according to the feature information, the feature information is one of the following parameters or a combination thereof: The 3D posture of the face of the two persons, the state of the basic action unit AU of the face of the second person, the ratio of the length and width of the outline of the face of the second person, and the feature point of the face of the second person The degree of darkness of the surrounding skin.
  6. 如权利要求3所述的方法,其特征在于,在检测出的脸部的区域里,采用人脸识别算法确定五官的区域和脸颊的轮廓。The method according to claim 3, wherein in the area of the detected face, a face recognition algorithm is used to determine the contour of the facial features and the cheeks.
  7. 如权利要求1至6任一所述的方法,其特征在于,在目标图像中,将第二人物替换为显示调整后的第一人物之后,进一步包括:The method according to any one of claims 1 to 6, wherein after the second character is replaced with the adjusted first character in the target image, the method further comprises:
    为目标图像中的第一人物添加图像。Add an image to the first person in the target image.
  8. 一种图像处理系统,其特征在于,包括:An image processing system, comprising:
    模型模拟模块,用于根据至少一张包含第一人物的图像,模拟出第一人物的模型;a model simulation module, configured to simulate a model of the first character according to at least one image including the first character;
    目标图像确定模块,用于确定包含第二人物的目标图像;a target image determining module, configured to determine a target image that includes the second person;
    特征信息确定模块,用于确定在目标图像中显示第二人物的特征信息;a feature information determining module, configured to determine that feature information of the second person is displayed in the target image;
    调整显示模块,用于在第一人物的模型中根据所述特征信息调整第一人物的显示;Adjusting a display module, configured to adjust a display of the first character according to the feature information in a model of the first character;
    人物替换模块,用于在目标图像中,将第二人物替换为显示调整后的第一人物。a character replacement module for replacing the second character with the adjusted first character in the target image.
  9. 如权利要求8所述的系统,其特征在于,所述模型模拟模块包括:The system of claim 8 wherein said model simulation module comprises:
    第一检测单元,用于检测出第一人物的脸部的区域;a first detecting unit, configured to detect an area of a face of the first person;
    第一确定单元,用于在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓;a first determining unit, configured to determine an outline of the facial features and a cheek in the detected area of the face;
    贴合单元,用于将检测到的五官的区域和脸颊的轮廓,贴合到已有的人脸3D模型上后获得模拟出的第一人物的脸部的模型。The fitting unit is configured to fit the detected facial features and the outline of the cheek to the existing human face 3D model to obtain a model of the simulated first person's face.
  10. 如权利要求8或9所述的系统,其特征在于,所述目标图像确定模块包括:The system of claim 8 or 9, wherein the target image determining module comprises:
    第二检测单元,用于检测出第二人物的脸部的区域;a second detecting unit, configured to detect an area of a face of the second person;
    第二确定单元,用于在检测出的脸部的区域里,确定五官的区域和脸颊的轮廓; a second determining unit, configured to determine an outline of the facial features and a cheek in the detected area of the face;
    特征单元,用于根据检测到的五官的区域和脸颊的轮廓,确定在目标图像中显示第二人物的脸部的特征信息。And a feature unit configured to determine feature information of displaying a face of the second person in the target image according to the detected area of the facial features and the outline of the cheek.
  11. 如权利要求10所述的系统,其特征在于,所述特征信息确定模块进一步用于根据第一人物的脸部的区域与第二人物的脸部的区域将第二人物的脸部替换为显示调整后的第一人物的脸部。The system according to claim 10, wherein said feature information determining module is further configured to replace the face of the second person with the display according to the area of the face of the first person and the area of the face of the second person Adjust the face of the first character.
  12. 如权利要求11所述的系统,其特征在于,所述调整显示模块进一步用于在第一人物的模型中根据以下参数之一或者其组合的所述特征信息调整第一人物的脸部的显示:第二人物的脸部的3D姿态、第二人物的脸部的AU的状态、第二人物的脸部的轮廓的长宽的比例、第二人物的脸部的特征点周围的皮肤的亮暗程度。The system according to claim 11, wherein the adjustment display module is further configured to adjust a display of a face of the first person according to one of the following parameters or the combination of the feature information in a model of the first person : the 3D posture of the face of the second person, the state of the AU of the face of the second person, the ratio of the length and width of the outline of the face of the second person, and the brightness of the skin around the feature point of the face of the second person Darkness.
  13. 如权利要求10所述的系统,其特征在于,所述调整显示模块进一步用于在检测出的脸部的区域里,采用人脸识别算法确定五官的区域和脸颊的轮廓。The system of claim 10 wherein said adjustment display module is further operative to determine a contour of the facial features and cheeks using a face recognition algorithm in the region of the detected face.
  14. 如权利要求8至13任一所述的系统,其特征在于,进一步包括:The system of any of claims 8 to 13, further comprising:
    道具添加模块,用于在目标图像中,将第二人物替换为显示调整后的第一人物之后,为目标图像中的第一人物添加图像。 The item adding module is configured to add an image to the first person in the target image after replacing the second character with the adjusted first character in the target image.
PCT/CN2015/077353 2014-07-23 2015-04-24 Image processing method and system WO2016011834A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410352939.4A CN104123749A (en) 2014-07-23 2014-07-23 Picture processing method and system
CN2014103529394 2014-07-23

Publications (1)

Publication Number Publication Date
WO2016011834A1 true WO2016011834A1 (en) 2016-01-28

Family

ID=51769146

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/077353 WO2016011834A1 (en) 2014-07-23 2015-04-24 Image processing method and system

Country Status (2)

Country Link
CN (1) CN104123749A (en)
WO (1) WO2016011834A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107644455A (en) * 2017-10-12 2018-01-30 北京旷视科技有限公司 Face image synthesis method and apparatus
CN109410298A (en) * 2018-11-02 2019-03-01 北京恒信彩虹科技有限公司 A kind of production method and expression shape change method of dummy model
CN109448069A (en) * 2018-10-30 2019-03-08 维沃移动通信有限公司 A kind of template generation method and mobile terminal
CN109461117A (en) * 2018-10-30 2019-03-12 维沃移动通信有限公司 A kind of image processing method and mobile terminal
CN109685044A (en) * 2019-02-18 2019-04-26 上海德拓信息技术股份有限公司 Recognition of face search method based on k means clustering algorithm
CN109788312A (en) * 2019-01-28 2019-05-21 北京易捷胜科技有限公司 The replacement method of personage in a kind of video
CN110503599A (en) * 2019-08-16 2019-11-26 珠海天燕科技有限公司 Image processing method and device
CN111508050A (en) * 2020-04-16 2020-08-07 北京世纪好未来教育科技有限公司 Image processing method and device, electronic equipment and computer storage medium
CN112101073A (en) * 2019-06-18 2020-12-18 北京陌陌信息技术有限公司 Face image processing method, device, equipment and computer storage medium
CN112330529A (en) * 2020-11-03 2021-02-05 上海镱可思多媒体科技有限公司 Dlid-based face aging method, system and terminal
CN112949360A (en) * 2019-12-11 2021-06-11 广州市久邦数码科技有限公司 Video face changing method and device
CN113569789A (en) * 2019-07-30 2021-10-29 北京市商汤科技开发有限公司 Image processing method and device, processor, electronic device and storage medium
CN113963127A (en) * 2021-12-22 2022-01-21 深圳爱莫科技有限公司 Simulation engine-based model automatic generation method and processing equipment
CN113963425A (en) * 2021-12-22 2022-01-21 北京的卢深视科技有限公司 Testing method and device of human face living body detection system and storage medium

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104123749A (en) * 2014-07-23 2014-10-29 邢小月 Picture processing method and system
CN104378620B (en) * 2014-11-24 2017-01-18 联想(北京)有限公司 Image processing method and electronic device
CN104376589A (en) * 2014-12-04 2015-02-25 青岛华通国有资本运营(集团)有限责任公司 Method for replacing movie and TV play figures
CN105118082B (en) * 2015-07-30 2019-05-28 科大讯飞股份有限公司 Individualized video generation method and system
CN105069745A (en) * 2015-08-14 2015-11-18 济南中景电子科技有限公司 face-changing system based on common image sensor and enhanced augmented reality technology and method
CN105488489A (en) * 2015-12-17 2016-04-13 掌赢信息科技(上海)有限公司 Short video message transmitting method, electronic device and system
CN105577517A (en) * 2015-12-17 2016-05-11 掌赢信息科技(上海)有限公司 Sending method of short video message and electronic device
CN106022221B (en) * 2016-05-09 2021-11-30 腾讯科技(深圳)有限公司 Image processing method and system
CN106127139B (en) * 2016-06-21 2019-06-25 东北大学 A kind of dynamic identifying method of MOOC course middle school student's facial expression
CN106054641A (en) * 2016-06-29 2016-10-26 Tcl集团股份有限公司 Method, apparatus, and system for turning on intelligent household appliance control interface
CN107945102A (en) * 2017-10-23 2018-04-20 深圳市朗形网络科技有限公司 A kind of picture synthetic method and device
CN109978754A (en) * 2017-12-28 2019-07-05 广东欧珀移动通信有限公司 Image processing method, device, storage medium and electronic equipment
CN109977739A (en) * 2017-12-28 2019-07-05 广东欧珀移动通信有限公司 Image processing method, device, storage medium and electronic equipment
CN108334821B (en) * 2018-01-18 2020-12-18 联想(北京)有限公司 Image processing method and electronic equipment
CN110096925B (en) * 2018-01-30 2021-05-14 普天信息技术有限公司 Enhancement method, acquisition method and device of facial expression image
CN108256497A (en) * 2018-02-01 2018-07-06 北京中税网控股股份有限公司 A kind of method of video image processing and device
US10755112B2 (en) * 2018-03-13 2020-08-25 Toyota Research Institute, Inc. Systems and methods for reducing data storage in machine learning
CN108537880A (en) * 2018-03-30 2018-09-14 盎锐(上海)信息科技有限公司 Data capture method with image comparing function and model generating means
CN108510581A (en) * 2018-03-30 2018-09-07 盎锐(上海)信息科技有限公司 Data capture method and model generating means
CN113658298A (en) * 2018-05-02 2021-11-16 北京市商汤科技开发有限公司 Method and device for generating special-effect program file package and special effect
CN108682030B (en) * 2018-05-21 2022-04-26 北京微播视界科技有限公司 Face replacement method and device and computer equipment
CN108897856A (en) * 2018-06-29 2018-11-27 联想(北京)有限公司 A kind of information processing method and electronic equipment
CN109271931A (en) * 2018-09-14 2019-01-25 辽宁奇辉电子系统工程有限公司 It is a kind of that gesture real-time identifying system is pointed sword at based on edge analysis
CN111274602B (en) * 2020-01-15 2022-11-18 腾讯科技(深圳)有限公司 Image characteristic information replacement method, device, equipment and medium
CN112990123B (en) * 2021-04-26 2021-08-13 北京世纪好未来教育科技有限公司 Image processing method, apparatus, computer device and medium
CN113962845B (en) * 2021-08-25 2023-08-29 北京百度网讯科技有限公司 Image processing method, image processing apparatus, electronic device, and storage medium
CN113766339B (en) * 2021-09-07 2023-03-14 网易(杭州)网络有限公司 Bullet screen display method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1522425A (en) * 2001-07-03 2004-08-18 �ʼҷ����ֵ������޹�˾ Method and apparatus for interleaving a user image in an original image
US20090252435A1 (en) * 2008-04-04 2009-10-08 Microsoft Corporation Cartoon personalization
CN101930618A (en) * 2010-08-20 2010-12-29 李浩民 Method for producing individual two-dimensional anime
CN104123749A (en) * 2014-07-23 2014-10-29 邢小月 Picture processing method and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1522425A (en) * 2001-07-03 2004-08-18 �ʼҷ����ֵ������޹�˾ Method and apparatus for interleaving a user image in an original image
US20090252435A1 (en) * 2008-04-04 2009-10-08 Microsoft Corporation Cartoon personalization
CN101930618A (en) * 2010-08-20 2010-12-29 李浩民 Method for producing individual two-dimensional anime
CN104123749A (en) * 2014-07-23 2014-10-29 邢小月 Picture processing method and system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LI, CHUANXUE.: "Automatically Replacing Faces in Images and Videos", ELECTRONIC JOURNALS OF CHINESE MASTER'S THESES FULL-TEXT DATABASE, 31 July 2012 (2012-07-31), pages 1 - 64 *

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107644455A (en) * 2017-10-12 2018-01-30 北京旷视科技有限公司 Face image synthesis method and apparatus
CN109461117B (en) * 2018-10-30 2023-11-24 维沃移动通信有限公司 Image processing method and mobile terminal
CN109448069B (en) * 2018-10-30 2023-07-18 维沃移动通信有限公司 Template generation method and mobile terminal
CN109461117A (en) * 2018-10-30 2019-03-12 维沃移动通信有限公司 A kind of image processing method and mobile terminal
CN109448069A (en) * 2018-10-30 2019-03-08 维沃移动通信有限公司 A kind of template generation method and mobile terminal
CN109410298A (en) * 2018-11-02 2019-03-01 北京恒信彩虹科技有限公司 A kind of production method and expression shape change method of dummy model
CN109410298B (en) * 2018-11-02 2023-11-17 北京恒信彩虹科技有限公司 Virtual model manufacturing method and expression changing method
CN109788312A (en) * 2019-01-28 2019-05-21 北京易捷胜科技有限公司 The replacement method of personage in a kind of video
CN109788312B (en) * 2019-01-28 2022-10-21 北京易捷胜科技有限公司 Method for replacing people in video
CN109685044A (en) * 2019-02-18 2019-04-26 上海德拓信息技术股份有限公司 Recognition of face search method based on k means clustering algorithm
CN109685044B (en) * 2019-02-18 2023-06-06 上海德拓信息技术股份有限公司 Face recognition retrieval method based on k-means clustering algorithm
CN112101073A (en) * 2019-06-18 2020-12-18 北京陌陌信息技术有限公司 Face image processing method, device, equipment and computer storage medium
CN112101073B (en) * 2019-06-18 2023-12-19 北京陌陌信息技术有限公司 Face image processing method, device, equipment and computer storage medium
CN113569789A (en) * 2019-07-30 2021-10-29 北京市商汤科技开发有限公司 Image processing method and device, processor, electronic device and storage medium
CN113569789B (en) * 2019-07-30 2024-04-16 北京市商汤科技开发有限公司 Image processing method and device, processor, electronic equipment and storage medium
CN110503599B (en) * 2019-08-16 2022-12-13 郑州阿帕斯科技有限公司 Image processing method and device
CN110503599A (en) * 2019-08-16 2019-11-26 珠海天燕科技有限公司 Image processing method and device
CN112949360A (en) * 2019-12-11 2021-06-11 广州市久邦数码科技有限公司 Video face changing method and device
CN111508050B (en) * 2020-04-16 2022-05-13 北京世纪好未来教育科技有限公司 Image processing method and device, electronic equipment and computer storage medium
CN111508050A (en) * 2020-04-16 2020-08-07 北京世纪好未来教育科技有限公司 Image processing method and device, electronic equipment and computer storage medium
CN112330529A (en) * 2020-11-03 2021-02-05 上海镱可思多媒体科技有限公司 Dlid-based face aging method, system and terminal
CN113963425B (en) * 2021-12-22 2022-03-25 北京的卢深视科技有限公司 Testing method and device of human face living body detection system and storage medium
CN113963127B (en) * 2021-12-22 2022-03-15 深圳爱莫科技有限公司 Simulation engine-based model automatic generation method and processing equipment
CN113963425A (en) * 2021-12-22 2022-01-21 北京的卢深视科技有限公司 Testing method and device of human face living body detection system and storage medium
CN113963127A (en) * 2021-12-22 2022-01-21 深圳爱莫科技有限公司 Simulation engine-based model automatic generation method and processing equipment

Also Published As

Publication number Publication date
CN104123749A (en) 2014-10-29

Similar Documents

Publication Publication Date Title
WO2016011834A1 (en) Image processing method and system
JP7200139B2 (en) Virtual face makeup removal, fast face detection and landmark tracking
Rhodin et al. General automatic human shape and motion capture using volumetric contour cues
US11450075B2 (en) Virtually trying cloths on realistic body model of user
US11494915B2 (en) Image processing system, image processing method, and program
US7876931B2 (en) Face recognition system and method
US20170372505A1 (en) Systems and Methods for Generating Computer Ready Animation Models of a Human Head from Captured Data Images
US10755145B2 (en) 3D spatial transformer network
Park et al. Articulated pose estimation with tiny synthetic videos
Malassiotis et al. Real-time hand posture recognition using range data
WO2022095721A1 (en) Parameter estimation model training method and apparatus, and device and storage medium
JP6207210B2 (en) Information processing apparatus and method
KR20140033088A (en) Generation of avatar reflecting player appearance
JP2009020761A (en) Image processing apparatus and method thereof
US20200065559A1 (en) Generating a video using a video and user image or video
KR20230085931A (en) Method and system for extracting color from face images
Yu et al. Improving person detection using synthetic training data
Jiang et al. independent hand gesture recognition with Kinect
US11361467B2 (en) Pose selection and animation of characters using video data and training techniques
KR20050039765A (en) Method for extracting face outline and face feature points
Micilotta Detection and tracking of humans for visual interaction
Ming et al. 3D face reconstruction using a single 2D face image
Mavzuna MODELING OF TEXT RECOGNITION IN IMAGES
Liu et al. Human performance capture using multiple handheld kinects
Aleksandrova et al. Approach for Creating a 3D Model of a Face from its 2D Image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15825342

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15825342

Country of ref document: EP

Kind code of ref document: A1