EP2076886A1 - Verfahren und vorrichtung für die virtuelle simulation einer videobildsequenz - Google Patents

Verfahren und vorrichtung für die virtuelle simulation einer videobildsequenz

Info

Publication number
EP2076886A1
EP2076886A1 EP07858653A EP07858653A EP2076886A1 EP 2076886 A1 EP2076886 A1 EP 2076886A1 EP 07858653 A EP07858653 A EP 07858653A EP 07858653 A EP07858653 A EP 07858653A EP 2076886 A1 EP2076886 A1 EP 2076886A1
Authority
EP
European Patent Office
Prior art keywords
face
image
sequence
digital
head
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP07858653A
Other languages
English (en)
French (fr)
Inventor
Jean-Marc Robin
Christophe Blanc
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to EP11007429A priority Critical patent/EP2450852A1/de
Publication of EP2076886A1 publication Critical patent/EP2076886A1/de
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/001Texturing; Colouring; Generation of texture or colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis

Definitions

  • the present invention relates to a method and a device for automatically simulating and processing in real time one or more aesthetic images of a real objective, for example a face and / or a head of a moving character in a scene, by detecting and monitoring its characteristic features.
  • the features of the face participate in the act of communication between human beings. Nevertheless, it should be noted that the visualization of the characteristic features of the face is a support for communication only if these features are extracted with sufficient precision. In the opposite case, the information resulting from an analysis that is too brief constitutes more an inconvenience than an aid, particularly for high-level industrial applications for the aesthetic embellishment of a face / head, for example.
  • an online user on the Internet from their own PC or Mac can use with a delay of more than 24 hours, and after posting on the Internet their color portrait photograph in JPEG format and under a number of constraints.
  • an ASP web server platform (“Application Service Provider") which provides users of a third-party Web site with contour contouring functions, color detection hair, carnation and eyes, which are obtained according to statistical and manual methods performed by the intervention of human technical teams.
  • the photograph can also be retouched using different layer techniques.
  • each element can be placed on a different layer, the final result being obtained by the superposition of all the layers, which allows to obtain the final retouched photograph for its exploitation.
  • the retouched photograph can then be exploited locally in a dedicated Microsoft® X Active Type application.
  • This technology and the set of tools developed by Microsoft® make it possible to program components that can make the content of a Web page interact with applications executable on the computer of the Internet user type PC, in particular under a computer system. Windows® operation. Another equivalent technique is to use the Java® application.
  • WO 01/75796 discloses a system for virtually transforming a still image, such as a photo of a face.
  • the present invention aims to provide a method and an image processing device do not reproduce the disadvantages of the prior art.
  • the present invention is intended in particular to provide such a method and device for processing video image sequences, in particular for moving subjects.
  • the present invention also aims to provide such a method and device, which is simple, reliable and inexpensive to make and use.
  • the present invention therefore relates to a method and a high-performance digital device, allowing in a video stream composed of a succession of images, for example from 25 to 30 images per second, component the illusion of movement, to obtain by computer and in real time, an automatic and precise extraction of the contours of all the characteristic features of the face, the hair, the color of the eyes, the complexion and the hair and taking into account certain occlusions.
  • the set leads to an aesthetic and individualized virtual simulation of an initial objective, for example a face / head of a character moving in a scene, in a video stream by a robust and real-time or delayed processing or " play-back Marin This simulation can comprise an encoding then a transformation by a new reading of the sequence.
  • the subject of the present invention is therefore an automatic method for virtual simulation of an individualized video image sequence for each user, which can be produced from a real video image sequence of a moving face / head, comprising, for example, an acquisition and initialization phase: the detection and analysis of the shapes and / or contours and / or dynamic components of an image of the face / head of the actual video sequence, the extraction of characteristic points of the face head, such as the corners of the eyes and mouth, using predefined parametric models; during a phase of evolution: the definition of specific parametric models from said extracted characteristic points, serving as initial points of initiation, the deformation of said specific models to adapt to the contours of the features present on the analyzed face, the detection and analysis of cutaneous structure of one or more regions of the face / head; and during a follow-up and transformation phase: the modification of the characteristic features of the other images of the video sequence, the modification of the colors of the cutaneous structure, said modifications being performed according to criteria provided for in at least one database and / or according to decision criteria of at least one expert
  • the detection and analysis step for determining region / contour and temporal spatial information, is performed by maximizing the luminance gradient and / or chrominance fluxes.
  • said modifications are obtained by translations of the neighborhoods of the characteristic points of the preceding image in the following image, affine models, including a deformation matrix, which can be used when the neighborhood of the characteristic points can also undergo a deformation.
  • the tracking phase uses an algorithm to follow characteristic points from one image to another.
  • said algorithm uses only the neighborhood of characteristic points.
  • the characteristic points are recaled using a simplified version of the active contours, and / or by deforming the curves of a model obtained in the previous image.
  • the method comprises a step of modeling the closed and / or open mouth by means of a plurality of characteristic points connected by a plurality of cubic curves.
  • the present invention also relates to a device for implementing the method described above, comprising a computer system, a light source, an electronic message management system, at least one database, local or remote on the networks digital, such as the Internet, and / or at least one expert system of order 0+ or 1, for obtaining and transforming a sequence of digital real images into a virtual image sequence, preferably at the speed of 25 frames per second, said virtual image sequence being transformed according to decision criteria of at least one expert system of order 0+ or 1.
  • said computer system is based on a microprocessor type CPU ("Central Processing Unit”) mono, dual, quad cores and higher, or conventional multicore processors, types pentium, athlon or higher or type SPU ("Streaming Processor Unit”) , equipped with a main core and up to eight specific cores, arranged in a cabin, a console, a self-service device, a pocket or mobile device, a digital television, a local server, or deported over digital networks, such as the Internet, at least one digital video camera, at least one screen, at least one printer and / or a connection to digital networks, such as the Internet, in wherein the computer system providing the image processing comprises a computer having a hard disk, preferably of a capacity of at least 500K bytes, and / or a digital storage memory, one or more media, such as CD-ROM, DVD, Multimedia Card®, Memory Stick®, MicroDrive®, XD Card®, SmartMedia®, SD Card®, Compact Flash® Type 1 and 2, USB stick, modem or wired or radio
  • a printer After displaying the virtual image sequence on a screen, a printer proceeds locally, or remotely, printing, preferably color, of at least one photograph selected from all or part of the sequence of virtual images.
  • the image processing module to carry out the steps of acquisition, detection, transformation and monitoring, is integrated in one or more processors specialized in the processing of signals of the DSP ("Digital Signal Processor” or "DSP") type. digital signal processor).
  • DSP Digital Signal Processor
  • DSP digital signal processor
  • the present invention therefore relates to a method and a device for simulating, by automatic processing and in all the environmental conditions and poses of the moving subject, an image or a sequence of aesthetic images in a video stream, starting from a or several images of a real objective, for example a face / head of a moving character in a scene, where one extracts from the image and / or from the sequence of real images, captured preferably by a Color digital video camera, at the real-time rate, the contours of the dynamic components of the face / head of the moving character, to produce relevant parameters to synchronize the tools of virtual aesthetic transformations, for example in the regions of the eyes, eyebrows , mouth and neighborhoods, in multi-criteria function provided in at least one local and / or remote database on digital networks, such as the Internet, and / or according to decision criteria previously defined in the knowledge base of at least one expert system of order 0+ or 1.
  • the computer system implemented may be installed in a cabin or console or a self-service device or a pocket or mobile device or a digital television or a local server or remote on digital networks, including the Internet, or any form of possible devices to come.
  • the computer may include a computer or a microprocessor for processing type CPU ("Central Processing Unit” or CPU), mono, dual, quad cores and higher or multi core processors classic types pentium, athlon or higher , or SPU type ("Streaming Processor Unit"), equipped with a main core and up to eight or more specific cores, with a hard disk of at least 500K bytes and / or a digital storage memory, one or more media of the CD-ROM, DVD, Multimedia Card®, Memory Stick®, MicroDrive®, XD Card®, SmartMedia®, SD Card®, Compact Flash® types 1 and 2, USB key or other, all types of modems or modules for wired or wireless connections to digital networks, such as the Internet, one or more connection modules for the local environment such as Bluetooth®, infrared, wifi®, wimax® and forthcoming , a color camera fixed video or type CCD and superior digital television, a discrete light source or not, all types of screens, preferably colors, current and future, all types of printers monochrome or current
  • CPU
  • the vision system defined above can be composed of a discreet daylight or white light-emitting diode, a mono CCD or higher camera, a graphics card, a flat and touch-sensitive color screen. a receipt-type printer or A4 and A3 paper color or higher.
  • the assembly can then be fully integrated in a lightweight "panel PC" ultra light whose size is given by the dimensions of all types of flat screen, preferably color.
  • the processing is local and all technical updates or content such as maintenance can be done via a wired connection or radio frequency to digital networks including the Internet.
  • the system operates by the visual servoing of the device via its camera, automatically leaning on the face of the user by the use of a module for detecting and tracking a face / head of a character. From a position that we say balance, when a simulation is desired by a user, the system will stop when it has detected its purpose, for example the image of a face / head.
  • the vision system can automatically adjust the camera's illumination and zoom in order to have an image size and a facial image code on the screen of almost constant optimal quality.
  • control system of the process functions can be a terminal, such as an alphanumeric keyboard, a mouse or any other means.
  • the camera depending on the terminal, can be connected by all connections or all types of digital networks to an editing system, preferably with the outputs on color screen and / or paper assembled in a single device next to the user.
  • the processing and calculation part can be managed by one or more local servers, or deported on digital networks, including the Internet, and equipped with at least one microprocessor, for example 32 or 64 bit CPU ("Central Processing Unit") type. mono, dual, quad-core and higher or multi-core processors classic types pentium, athlon, or type SPU ("Streaming Process Unit") or a main core and eight specific cores type cell, and all types of electronic or magnetic memories.
  • CPU Central Processing Unit
  • the capture of the color images is advantageously obtained in real time by means of all types of digital video cameras, preferably color, such as a digital color video camera mono or CCD, or a device to load coupling and superior, a complementary color CMOS (metal-oxide semiconductor) video camera, or similar, for example, a webcam, in order to provide, in real time, a valuable aesthetic simulation by high quality detection, components geometric and dynamic facial features and appropriate image processing.
  • CMOS metal-oxide semiconductor
  • the processing can be done locally or on a remote server and according to the speed of calculation, in real time or assimilated as such, or in playback mode.
  • the whole treatment can be done without too many lighting constraints and poses for each character present in the image, considering an uncertain background fixed or mobile, and a number of occlusions.
  • the experiments demonstrate that the method implemented according to the present invention remains robust and sufficiently precise during the phase of extraction and evolution of the contours of the permanent features of the face / head namely namely: the eyes, eyebrows, lips, hair and other morphological elements, according to the work of aesthetic transformation sought, etc.
  • the method advantageously comprises three synchronized phases: 1.
  • An acquisition and initialization phase the shapes and contours of the face / head are analyzed and detected in a digitized video image corresponding to the first image of the image. 'a sequence. Dots characteristics and areas of interest of the face / head are extracted, corners of eyes and mouth for example, and serve as initial anchors for each of the predefined adapted parametric models.
  • each model is deformed in order to coincide with the contours of the features present on the analyzed face. This deformation is done by maximizing a luminance gradient flux and / or chrominance along the contours defined by each curve of the model.
  • the definition of models makes it possible to naturally introduce a regularization constraint on the contours sought. Nevertheless, the chosen models remain flexible enough to allow a realistic extraction of the contours of the eyes, the eyebrows and the mouth.
  • a phase of monitoring and transformation Tracking makes segmentation more robust and faster in subsequent frames of the video clip.
  • the transformation leads to the modification of the fine characteristic areas of the face / head followed in the video sequence according to multiple criteria provided for in the database and / or as the case may be according to decision criteria of an expert system of order 0 + or 1.
  • a rendering phase offering on a screen and / or paper and / or via a server on all digital networks, the results of the transformation phase for the entire video sequence.
  • the video processing system will coordinate several successive operations.
  • the detected face / head corresponds to an area of interest in the image.
  • the method makes it possible to overcome illumination variations by using a filter adapted to the behavior of the retina, for the region of interest.
  • the system then proceeds, for the region of interest thus filtered, to the extraction of the characteristic features of the face, preferably using suitable parametric models, namely the irises, eyes, eyebrows, lips, the contour of the face and the helmet of the hair.
  • the initial positioning of each model on the image to be processed takes place after the automatic extraction of characteristic points on the face.
  • a maximum luminance gradient point tracking process can be used to detect the corner of the eyes.
  • Two Bézier curves one of which is curved towards its end to naturally follow the drawing of the lower contours of the eye, models chosen for the upper and lower contours of the eye, can be initialized by the two corners of the eyes. and the lowest point of the circle detected for the iris for the lower contour, and by the two corners of the eyes and the center of the circle detected for the iris for the upper contour.
  • the proposed model for the modeling of the lips is advantageously at least composed of five independent curves, each of them naturally describing part of the outer lip contour and at least two curves for the inner contours.
  • the characteristic points of the mouth in order to initialize the model can be analyzed by using a discriminant information combining luminance and chrominance as well as the convergence of an active contour type that makes it possible to overcome the parameters of the parameters of the contour. as well as its high dependence on the initial position.
  • the modeling of the contour of the face advantageously uses eight characteristic points situated on this contour. These eight points initialize an outline modeled by deformable ellipse quarters, according to the position of the face in a temporal dimension.
  • the hair helmet can be segmented from the detection of the contour of the face by associating the filtering of the background of the image with the use of active contours. Characteristic points located on the contour of the hair are thus detected. Between each of these points, the model used can be a cubic polynomial curve.
  • each desired contour is a set of points of maximum luminance gradient.
  • the selected curves will preferably be those that maximize the normalized luminance gradient flux across the contour.
  • the tracking step allows the segmentation in the following images of the video sequence.
  • the results obtained in previous images provide additional information that can make segmentation more robust and faster.
  • the precise tracking procedure uses an algorithm that makes it possible to follow characteristic points from one image to another. This differential method, using only the neighborhood of points, brings a significant time saving compared to a direct extraction technique.
  • the characteristic points can be recalculated by using a simplified version of the active contours, and / or by deforming the curves of a model obtained in the previous image.
  • the transformation step may lead to the modification of the fine characteristic areas of the face / head followed in the video sequence according to multiple criteria provided in the database (s) and / or as the case may be according to decision criteria of at least one expert system of order 0+ or 1.
  • the present invention may offer the user different looks, palettes, present in the database for viewing on his face.
  • the system can search, from anthropometric ratios calculated by an expert system of order 0+ or 1, the characteristic zones, for example the cheekbones, to transform .
  • the expert system can define makeup procedures that are dependent on the shape of the face, round or elongate or square or triangular or oval, and certain features, eyes apart or eyes close together or eyes equal, nose size, etc. These rules can be communicated to the transformation module for a realistic and dependent simulation of each face to transform.
  • the method also carries out during this phase a classification of faces such man, woman, child, teenager in particular.
  • the rendering phase offers on a screen and / or on paper and / or via a server on all digital networks, the results of the transformation phase for the entire video sequence and / or for part of this sequence
  • FIG. 1 is a block diagram of a virtual image simulation system according to an advantageous embodiment of the present invention
  • FIG. 2 is a block diagram illustrating the phase of extraction of the faces / heads of characters and the characteristic zones according to an advantageous embodiment of the present invention
  • FIG. 3 represents the block diagram of the retinal filtering
  • FIG. 4 is a drawing of one of the parametric models adapted to the tracking of moving lips
  • FIG. 5 represents the result of the automatic extraction of the characteristic areas of the face from a video sequence presenting a single character with the head moving in front of the camera lens along the axes of orientations X, Y and Z symbolized in this same figure, namely the contour of the face, the iris, the eyes, the mouth, the eyebrows, and the headgear of the hair;
  • FIG. 1 represents an example of a detection system and real-time automatic tracking of the characteristic features of a real objective, such as a face / head of a character moving in a scene, with the possibility of virtual simulation of images and comprising an image acquisition and initialization module 1, a tracking and transformation module 2 and a reproduction module 3. Each module will be described in detail below.
  • the image acquisition and initialization module 1 is implemented from all types of digital color video cameras, such as a color digital video camera CCD mono or charge coupled device and superior, a video camera complementary color CMOS (metal-oxide-semiconductor), or the like.
  • CMOS metal-oxide-semiconductor
  • the sequence of images taken by the acquisition module is analyzed in order to detect the zones and characteristic points of the face / head.
  • This analysis is implemented in the form of a microprocessor of the 32 or 64-bit CPU type, SPU, or a main core and eight specific cores type cell, mono, double, quad cores and higher or multi core processors classic types pentium, athlon , a personal computer or a digital signal processor.
  • the zones and characteristic points of the face / head of the moving character in a scene thus extracted and coupled to the flow of images are sent to the tracking and transformation module which according to the multiple criteria provided in one or more database (s) of data or, depending on the case, according to decision criteria of one or more expert system (s) 21, refers to the rendering module 3 its results: a video sequence with, for example, the masked face.
  • the rendering module offers, according to the present invention, the results on any type of screen (cathodic, LCD, plasma or the like) and / or on any paper format and / or via a server on all digital networks, for example the Internet.
  • Figure 2 shows a block diagram illustrating the extraction phase of the face / head of the character and characteristic areas according to the present invention.
  • the software for processing the video sequence clocked at the acquisition speed of the digital video sensor will coordinate several successive operations according to the invention.
  • the software for processing the video sequence clocked at the acquisition speed of the digital video sensor will coordinate several successive operations according to the invention.
  • he proceeds to locate the face / head of the character in a scene.
  • the typical chrominance information associated with the skin is considered.
  • a pretreatment phase 12 of this region of interest makes it possible to overcome illumination variations by using adapted filtering inspired by the behavior of the retina.
  • This filtering makes it possible, by performing a succession of filterings and adaptive compressions, to perform a local smoothing of the variations of illumination.
  • I 1n be the initial image and Ii the result of its filtering
  • the image X 0 makes it possible to define the compression function C by the relation:
  • Figure 3 gives the block diagram of the retinal filtering, the output of this filtering is noted I out .
  • the luminance variations will be greatly reduced.
  • the automatic extraction of the contours of the permanent features of the face namely the contour of the face, whose homogeneity is taken into account, the irises, the eyes, the eyebrows, the lips, the hair helmet, follows in a second time .
  • a specific parametric model (cubic polynomial curves, Bezier curves, circle, etc.) capable of accounting for all possible deformations is defined.
  • the iris we search for the semicircle that maximizes the normalized luminance gradient flux in each right and left quarter of the rectangle encompassing the face since the contour of the iris is the border between a dark zone, the iris, and a clear area, the white of the eye.
  • the method of maximizing the standardized gradient flow has the advantage of being very fast, without parameter adjustment, and it leads unambiguously to the selection of the right semicircle since the standardized gradient flow always has a very marked peak corresponding to the correct position for the desired semi-circle.
  • Characteristic points of the face are extracted (corners of the eyes and mouth for example) and serve as initial anchors 13 for each of the other models.
  • Bezier curves one of which is curved towards its end, models chosen for the upper and lower contours of the eye, are initialized by the two corners of the eyes, detected by a process of tracking points of maximum luminance gradient, and the lowest point of the circle detected for the iris for the lower contour and the two corners of the eyes and the center of the circle detected for the iris for the upper contour.
  • the two inner and outer corners of each eyebrow are advantageously extracted.
  • the search area of these points is reduced to the area of the image above the detected iris.
  • the abscissae of the inner and outer corners we search the abscissa of the points for which there is a change of sign or cancellation of the derivative of the horizontal projection of the valley image along the lines.
  • the two inner and outer corners and the center of its two corners serve as initial control points for the Bezier curve associated with each eyebrow. Since this method is subject to noise, the points thus detected are readjusted during the deformation phase of the model associated with the eyebrows.
  • the proposed model for lip modeling can be composed of five independent cubic curves, each of which describes a part of outside lip contour.
  • Figure 4 shows a drawing of this model for a closed mouth. Unlike most models proposed in the literature, this original model is sufficiently deformable to faithfully represent the specificities of very different lips. Between Q2 and Q4, Cupid's bow is described by a broken line while the other portions of the outline are described by cubic polynomial curves. In addition, it requires having a zero derivative at point Q2, Q4 and Q6. For example, the cubic between Qi and Q 2 must have a null derivative in Q 2 .
  • Extraction of the characteristic points Q 1 , Q 2 , Q 3 , Q 4 , Q 5 , Q 6 from the mouth in order to initialize the model is done by using a discriminant information combining luminance and chrominance as well as the convergence of an active contour type that makes it possible to dispense with the settings of the parameters of the active contour as well as its high dependence on the initial position. The same goes for the internal labial contours where two curves allow to perfectly marry the inner contours.
  • Detecting the inner contour is more difficult when the mouth is open, because of non-linear variations in appearance inside the mouth. Indeed, during a conversation, the area between the lips can take different configuration: teeth, oral cavity, gums and tongue.
  • the parametric model for the inner contour when the mouth is open, can be composed of four cubic.
  • "Cupid's bow inside” is less pronounced than for a closed mouth; thus two cubic is enough to precisely extract the upper inner contour of the lips.
  • the model With four cubic, the model is flexible and overcomes the problem of the segmentation of the inner contour for asymmetrical mouths.
  • jumping snakes Two active contours called jumping snakes can be used to adjust the model; one for the upper contour and one for the lower contour.
  • the convergence of a jumping snake is a succession of phases of growth and jumping.
  • the "snake” is initialized from a germ, then it grows adding points to the left and right of the germ. Each new point is found by maximizing a gradient flow through the segment formed by the current point to be added and the previous point. Finally, the seed jumps to a new position closer to the desired outline. The growth and jump processes are repeated until the jump amplitude is below a certain threshold.
  • the initialization of the two "snakes” begins with the search for two points on the upper and lower contours, and belonging to the vertical passing through Q 3 in Figure 4.
  • the difficulty of the task lies in the fact that there can have different areas between the lips, which may have characteristics (color, texture or luminance) similar or completely different than those of the lips, when the mouth is open.
  • the final inner contour can be given by four cubic.
  • the two cubes for the upper contour can be calculated by the least squares method.
  • the two cubic of the lower contour can also be calculated by the least squares method.
  • the modeling of the contour of the face preferably uses eight characteristic points situated on this contour a priori since a face can present very long hair which completely cover the forehead and possibly the eyebrows and the eyes: two points at the level of the eyes, two points at the level of the eyebrows, two points at the mouth, a point at the chin and a point at the forehead, which are extracted from a thresholding in the V plane of the HSV representation of the image. These eight points initialize an outline modeled by quarter ellipses.
  • the helmet of the hair can be segmented from the detection of the contour of the face by associating the filtering of the bottom of the image with the use of active contours. Characteristic points located on the contour of the hair are thus detected. Between each of these points, the model used can be a cubic polynomial curve. It is possible that the automatic extraction of one or more points fails, in this case the point or points can be very easily replaced manually to correctly replace the model or models and approach their evolution phase.
  • each model is deformed to coincide with the contours of the features present on the analyzed face. This deformation is done by maximizing a luminance gradient flux and / or chrominance, along the contours defined by each curve of the model.
  • FIG. 5 represents the result of the automatic extraction of the characteristic zones of the face, namely the contour of the face, the irises, the eyes, the mouth, the eyebrows, and the helmet of the hair, which respectively form anthropometric modules of the face, according to an aspect of the present invention.
  • the software tracks the face / head and facial features in the video sequence.
  • the results obtained in the previous images provide additional information that can make the segmentation more robust and faster.
  • the precise tracking procedure uses an algorithm that makes it possible to follow characteristic points from one image to another.
  • This differential method using only the neighborhood of points, brings a significant time saving compared to a direct extraction technique.
  • This method is based on the apparent motion stress equation derived from a Taylor development of the equation below:
  • the method advantageously uses a registration of the characteristic points by using a simplified version of the active contours and / or by deforming the curves of the model obtained in the image. previous. Finally, the final contours are extracted. For this, the shape of the characteristic zones in the previous image as well as the characteristic points are used to calculate the optimal curves constituting the different models.
  • the tools for recognizing and monitoring the anthropometric areas of the face in the image communicate all the data they have extracted during the transformation phase.
  • the module will then determine the treatments to be performed. These will be determined by the theme or themes that the user will have chosen. For example, if it is a make-up operation, the characteristic areas of the face, defined according to the extraction results and according to the function chosen by the user (look / palette), are automatically modified in the sequence of consecutive images according to harmonic and personalized choices. For example, for a round face, the process blurs the sides of the face in a darker tone.
  • the user can also choose zone by zone (lips, eyes, cheekbones, face, etc.) the color he wishes to apply. These colors will be in harmony with the characteristics of the face.
  • the expert system determines a range of available colors, correlated with those of a range available in its database or databases, according to the data from the initialization phase and evolution.
  • the tool will be able to make a coloring proposal in harmony with the face for example, but also offer a selection of colors, from a range, in perfect harmony with the face.
  • the colors complemented by their original textures are analyzed, calculated and defined in their particular context (lipsticks or glosses or powders in particular).
  • the tools will then apply depending on the texture of the area (lip, cheek, hair, etc.) the color corresponding to the makeup, but also in a transparent manner the effect of the cosmetic product, that is to say that it will reproduce its real appearance, for example its brilliance, its powdered or glittery appearance (glittery lipstick in figure 6, especially its matte appearance)
  • This operation takes into account the context of the sequence of consecutive images in each of their respective areas (lighting, brightness, shadows, reflections, etc.), which will allow the help of algorithmic tools to calculate their textures, to define them in their real aspect, as they would be reproduced in reality .
  • the quality and realistic properties of the consecutive image sequence are substantially improved.
  • some features of the face are improved. For example, face wrinkles, crow's feet wrinkles, dark circles, lion wrinkles, nasolabial folds, bitter creases, perioral wrinkles, freckles, acne and rosacea are strongly blurred.
  • aesthetic treatments such as face whitening, tanning, teeth whitening, eyelid lifting, thickening of the lips, the light rectification of the oval of the face, the rectification of the shape of the chin and / or the nose, the raising and raising of the cheekbones are simulated automatically for a face appearing in a video sequence.
  • the invention also makes it possible to recognize visemes that describe the different configurations, or different pronounced phonemes, of a talking mouth. It thus makes it possible to determine the personality and the character of a person, examined from the morphological observation of his / her face / head, such as, for example, the presence of the folds of bitterness, the size and the spacing of the eyes, the size and shape of the nose, the lobe of the ears, the database corresponding to the observation of the faces being then supplemented by the techniques implemented by the morpho psychologists, the psychiatrists, the profilers and the anatomists in the domain considered.
  • RGB or red, green, blue completed indications drawings, thresholds and coordinates constituting the realization of a "Look" or the natural visualization of a lipstick in a palette for example can be made and recorded in the form of a simple file consisting of a low-weight alphanumeric string broadcast on all digital media or download from a server on digital networks such as the Internet.
  • This file can be used for the artistic update of the database or the expert system in a flexible and fast way or be used immediately by the user from a simple download from a web page for example.
  • the database associated with the expert system is enriched with specific rules relating to the application of the invention, for example cosmetics and / or dermatology, plastic surgery and / or aesthetic medicine, ophthalmology, hairdressers and / or hairdressers, facial biometrics, etc.
  • the treatment is independent of the content which allows a use of the process on an industrial scale and a very strong propagation of its use with a strong increase in yield.
  • Figure 6 shows the before / after result of a makeup simulation (look), accessories (color lenses, piercing), and hair coloring for an image extracted from a video sequence acquired by a color video camera.
  • the rendering module translates, according to the present invention, the display of the sequence of transformed images on any type of color screen and / or then by printing one or more images simulated on n ' any paper format and / or via a server on any digital network.
  • the restitution phase results in an aesthetic proposition characterized by the transformation of the initial video sequence into a a new virtual video sequence on which the desired aesthetic modifications appear in perfect harmony.
  • a makeup complete with accessories and a hair color and the references and sales prices of the corresponding products in one or more brands.
  • a static image chosen by the user from the video sequence can then be edited locally, on a color printer of needles type, inkjet, solid inkjet, laser, or thermal sublimation, in A4 format or any other format available technically.
  • the content of its information formulates a beauty prescription, taking the initial image and the transformed image, technical and scientific advice, professional tips, facial features (shape, color, etc.), product photography, personal color palette in harmony with the transformed facial features, a color garment board compared to our colors etc.
  • the results can be similarly edited on high definition delocalized printers from an Internet server which will send them to the user's postal address.
  • the new image and / or the new video sequence completed or not information can be sent by the email function and using the command "Insert attachment" to one or more correspondents with an email address type email.
  • the same is true with a mobile phone device having an MMS mode, email or future.
  • this system can receive a large number of applications by completing the expert system (s) and / or the local database (s) or remote (s) by scientific data. and specific techniques.
  • the invention can find an application for image processing in two or three dimensions.
  • a 3D application we can build a 3D modeling of the face, to precisely apply 3D makeup.
  • the 3D reconstruction of the face, from a static image of the face or a flow of images of faces, is made using conventional algorithms and procedures, such as the analysis of shadows, texture , movement, the use of generic 3D models of the face or the use of a stereo scopic system.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)
  • Processing Or Creating Images (AREA)
  • Image Analysis (AREA)
EP07858653A 2006-10-24 2007-10-23 Verfahren und vorrichtung für die virtuelle simulation einer videobildsequenz Withdrawn EP2076886A1 (de)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP11007429A EP2450852A1 (de) 2006-10-24 2007-10-23 Verfahren und Vorrichtung zur virtuellen Simulation eines Bildes

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FR0654483A FR2907569B1 (fr) 2006-10-24 2006-10-24 Procede et dispositif de simulation virtuelle d'une sequence d'images video.
PCT/FR2007/052234 WO2008050062A1 (fr) 2006-10-24 2007-10-23 Procédé et dispositif de simulation virtuelle d'une séquence d'images vidéo.

Publications (1)

Publication Number Publication Date
EP2076886A1 true EP2076886A1 (de) 2009-07-08

Family

ID=37964796

Family Applications (2)

Application Number Title Priority Date Filing Date
EP11007429A Withdrawn EP2450852A1 (de) 2006-10-24 2007-10-23 Verfahren und Vorrichtung zur virtuellen Simulation eines Bildes
EP07858653A Withdrawn EP2076886A1 (de) 2006-10-24 2007-10-23 Verfahren und vorrichtung für die virtuelle simulation einer videobildsequenz

Family Applications Before (1)

Application Number Title Priority Date Filing Date
EP11007429A Withdrawn EP2450852A1 (de) 2006-10-24 2007-10-23 Verfahren und Vorrichtung zur virtuellen Simulation eines Bildes

Country Status (8)

Country Link
US (1) US20100189357A1 (de)
EP (2) EP2450852A1 (de)
JP (1) JP2010507854A (de)
KR (1) KR20090098798A (de)
BR (1) BRPI0718306A2 (de)
CA (1) CA2667526A1 (de)
FR (1) FR2907569B1 (de)
WO (1) WO2008050062A1 (de)

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101903908A (zh) * 2007-12-18 2010-12-01 皇家飞利浦电子股份有限公司 基于特征的2d/3d图像配准
JP5221436B2 (ja) * 2009-04-02 2013-06-26 トヨタ自動車株式会社 顔特徴点検出装置及びプログラム
KR20120055598A (ko) 2009-08-04 2012-05-31 베잘리스 기준 이미지에 따라 타켓 이미지를 수정하는 이미지 처리방법 및 이미지 처리장치
CN102792337B (zh) * 2010-01-12 2016-01-20 新加坡南洋理工大学 用于生成数字图片的方法和装置
US10108852B2 (en) * 2010-06-07 2018-10-23 Affectiva, Inc. Facial analysis to detect asymmetric expressions
EP2619761B1 (de) 2010-09-22 2018-06-20 NDS Limited Anreicherung von digitalfotos
JP4893862B1 (ja) * 2011-03-11 2012-03-07 オムロン株式会社 画像処理装置、および画像処理方法
JP4893863B1 (ja) * 2011-03-11 2012-03-07 オムロン株式会社 画像処理装置、および画像処理方法
US20130271451A1 (en) 2011-08-09 2013-10-17 Xiaofeng Tong Parameterized 3d face generation
US9460462B1 (en) * 2012-05-22 2016-10-04 Image Metrics Limited Monetization using video-based simulation of cosmetic products
US9449412B1 (en) * 2012-05-22 2016-09-20 Image Metrics Limited Adaptive, calibrated simulation of cosmetic products on consumer devices
US20140168204A1 (en) * 2012-12-13 2014-06-19 Microsoft Corporation Model based video projection
US9390481B2 (en) * 2013-05-05 2016-07-12 Google Inc. Enhancing content appearance
CN103632165B (zh) 2013-11-28 2017-07-04 小米科技有限责任公司 一种图像处理的方法、装置及终端设备
US9704038B2 (en) * 2015-01-07 2017-07-11 Microsoft Technology Licensing, Llc Eye tracking
JP6497162B2 (ja) * 2015-03-26 2019-04-10 オムロン株式会社 画像処理装置および画像処理方法
JP6266736B1 (ja) 2016-12-07 2018-01-24 株式会社コロプラ 仮想空間を介して通信するための方法、当該方法をコンピュータに実行させるためのプログラム、および当該プログラムを実行するための情報処理装置
CN108280883B (zh) * 2018-02-07 2021-05-04 北京市商汤科技开发有限公司 变形特效程序文件包的生成及变形特效生成方法与装置
CN111814520A (zh) * 2019-04-12 2020-10-23 虹软科技股份有限公司 肤质检测方法、肤质等级分类方法及肤质检测装置
KR102285084B1 (ko) * 2019-12-24 2021-08-03 주식회사 텔레칩스 이종의 멀티 cpu를 운용하는 시스템-온-칩 및 그 동작 방법
US20220230471A1 (en) * 2021-01-15 2022-07-21 Chih-Wei Li Artificial Intelligence-Assisted Evaluation Method for Aesthetic Medicine and Evaluation System Using Same
JP2024537064A (ja) * 2021-09-30 2024-10-10 ロレアル 拡張現実化粧デザインフィルタ
FR3130423B1 (fr) * 2021-12-15 2024-01-12 Oreal Filtres de dessin cosmétique en réalité augmentée

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2728982A1 (fr) * 1994-12-29 1996-07-05 Jean Marc Robin Reconnaissance automatique des traits caracteristiques du visage et simulation d'une image esthetique d'un objectif (visage) reel
FR2783949B1 (fr) * 1998-09-30 2001-06-08 Lucette Robin Systeme teleinformatique et numerique pour la transformation d'une image, notamment de l'image d'un visage humain
JP4226730B2 (ja) * 1999-01-28 2009-02-18 株式会社東芝 物体領域情報生成方法及び物体領域情報生成装置並びに映像情報処理方法及び情報処理装置
AU3662600A (en) 2000-03-30 2001-10-15 Lucette Robin Digital remote data processing system for transforming an image, in particular an image of the human face
CN1313979C (zh) * 2002-05-03 2007-05-02 三星电子株式会社 产生三维漫画的装置和方法
JP3993029B2 (ja) * 2002-06-24 2007-10-17 デジタルファッション株式会社 化粧シミュレーション装置、化粧シミュレーション方法、化粧シミュレーションプログラム及びそのプログラムを記録した記録媒体
US7612794B2 (en) * 2005-05-25 2009-11-03 Microsoft Corp. System and method for applying digital make-up in video conferencing

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
CAPLIER A ET AL: "Accurate and Quasi-Automatic Lip Tracking", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 14, no. 5, 1 May 2004 (2004-05-01), pages 706 - 715, XP011112294, ISSN: 1051-8215, DOI: 10.1109/TCSVT.2004.826754 *
HAMMAL Z ET AL: "Parametric models for facial features segmentation", SIGNAL PROCESSING, ELSEVIER SCIENCE PUBLISHERS B.V. AMSTERDAM, NL, vol. 86, no. 2, 1 February 2006 (2006-02-01), pages 399 - 413, XP024997795, ISSN: 0165-1684, [retrieved on 20060201], DOI: 10.1016/J.SIGPRO.2005.06.006 *
NICOLAS EVENO: "Segmentation des lèvres par un modèle déformable analytique - PhD Thesis", 14 November 2003 (2003-11-14), GRENOBLE, FRANCE, pages 1 - 156, XP055064217, Retrieved from the Internet <URL:http://www-clips.imag.fr/geod/User/nicolas.eveno/Doc/PhD_EVENO.pdf> [retrieved on 20130527] *
See also references of WO2008050062A1 *
Z. HAMMAL ET AL: "Eyes and eyebrows parametric models for automatic segmentation", 6TH IEEE SOUTHWEST SYMPOSIUM ON IMAGE ANALYSIS AND INTERPRETATION, 2004., 1 January 2004 (2004-01-01), pages 138 - 141, XP055064258, ISBN: 978-0-78-038387-6, DOI: 10.1109/IAI.2004.1300961 *
ZAKIA HAMMAL: "Facial Features Segmentation, Analysis and Recognition of Facial Expressions using the Transferable Belief Model, PhD Thesis", 29 June 2006 (2006-06-29), pages 1 - 233, XP055064252, Retrieved from the Internet <URL:http://theses.eurasip.org/media/theses/documents/hammal-zakia-facial-features-segmentation-analysis-and-recognition-of-facial-expressions-by-the-transferable-belief-model.pdf> [retrieved on 20130527] *

Also Published As

Publication number Publication date
BRPI0718306A2 (pt) 2013-11-12
US20100189357A1 (en) 2010-07-29
CA2667526A1 (fr) 2008-05-02
EP2450852A1 (de) 2012-05-09
JP2010507854A (ja) 2010-03-11
FR2907569A1 (fr) 2008-04-25
FR2907569B1 (fr) 2009-05-29
KR20090098798A (ko) 2009-09-17
WO2008050062A1 (fr) 2008-05-02

Similar Documents

Publication Publication Date Title
EP2076886A1 (de) Verfahren und vorrichtung für die virtuelle simulation einer videobildsequenz
JP4449723B2 (ja) 画像処理装置、画像処理方法、およびプログラム
CA2579903C (en) System, method, and apparatus for generating a three-dimensional representation from one or more two-dimensional images
FR3053502A1 (fr) Systeme et procede de miroir de maquillage numerique
JP3984191B2 (ja) 仮想化粧装置及びその方法
US20110234581A1 (en) Methods and systems for three-dimensional rendering of a virtual augmented replica of a product image merged with a model image of a human-body feature
JP2002133446A (ja) 顔画像処理方法およびシステム
WO2014081394A1 (en) Method, apparatus and system for virtual clothes modelling
CN114821675B (zh) 对象的处理方法、系统和处理器
WO2021245273A1 (fr) Procédé et dispositif de reconstruction tridimensionnelle d&#39;un visage avec partie dentée à partir d&#39;une seule image
FR2728982A1 (fr) Reconnaissance automatique des traits caracteristiques du visage et simulation d&#39;une image esthetique d&#39;un objectif (visage) reel
FR2920938A1 (fr) Procede et dispositif de simulation virtuelle d&#39;une image
EP1643448A2 (de) Verfahren um das Erscheinungsbild mindestens eines Körperteils einer Person vorherzusagen
WO2001075796A1 (fr) Systeme teleinformatique et numerique pour la transformation d&#39;une image, notamment de l&#39;image d&#39;un visage humain
FR2968436A1 (fr) Methode de determination des mouvements d&#39;un objet a partir d&#39;un flux d&#39;images
Huang et al. A process for the semi-automated generation of life-sized, interactive 3D character models for holographic projection
Heravi Three-dimension facial de-ageing and ageing Modeling: extrinsic factors impact
Zhao 3D Human Face Reconstruction and 2D Appearance Synthesis
FR2783949A1 (fr) Systeme teleinformatique et numerique pour la transformation d&#39;une image, notamment de l&#39;image d&#39;un visage humain
Guo From Seeing Better to Understanding Better: Deep Learning for Modern Computer Vision Applications
FR3114895A1 (fr) Systèmes et procédés pour une classification améliorée des attributs faciaux et leurs applications
FR3112633A1 (fr) Vieillissement du visage contrôlable à haute résolution avec des GAN conditionnels à sensibilité spatiale
FR2920560A1 (fr) Procede de fabrication et d&#39;immersion automatique d&#39;avatar 3d a partir d&#39;une image 2d
Mehdi Developing a Computer System for the Generation of Unique Wrinkle Maps for Human Faces. Generating 2D Wrinkle Maps using Various Image Processing Techniques and the Design of 3D Facial Ageing System using 3D Modelling Tools.
Ghys Analysis, 3D reconstruction, & Animation of Faces

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20090430

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR

17Q First examination report despatched

Effective date: 20090916

DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20131016