WO2006003625A1 - Traitement video - Google Patents

Traitement video Download PDF

Info

Publication number
WO2006003625A1
WO2006003625A1 PCT/IB2005/052162 IB2005052162W WO2006003625A1 WO 2006003625 A1 WO2006003625 A1 WO 2006003625A1 IB 2005052162 W IB2005052162 W IB 2005052162W WO 2006003625 A1 WO2006003625 A1 WO 2006003625A1
Authority
WO
WIPO (PCT)
Prior art keywords
person
input signal
viewed
feature
image
Prior art date
Application number
PCT/IB2005/052162
Other languages
English (en)
Inventor
Richard P. Kleihorst
Hasan Ebrahimmalek
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Publication of WO2006003625A1 publication Critical patent/WO2006003625A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/802D [Two Dimensional] animation, e.g. using sprites

Definitions

  • the present invention relates to a video processing apparatus and method, and in particular, to a video processing apparatus and method involving cartoonizing.
  • Video communication is increasingly being used in numerous applications such as video telephones, video conferencing, television-collaboration, shared virtual table environments, and so on.
  • face detection and recognition is being actively researched in order to enhance the services provided by applications such as video conferencing.
  • face detection is used in video conferencing systems to create a virtual conference room, whereby participants of the meeting are seated around a virtual table.
  • Numerous approaches have been used to assist in face detection, including techniques such as feature invariant approaches, appearance-based approaches, and wavelength analysis.
  • Feature extraction methods utilize various properties of the face and skin to isolate and extract data, such as "eye" data.
  • Popular methods include skin color segmentation, principle component analysis, eigenspace modeling, histogram analysis and texture analysis.
  • face detection and skin detection methods are currently used in applications such as creating virtual video conferencing systems, or face recognition systems for security applications.
  • the aim of the present invention is to provide a video processing apparatus and method that utilizes information received from sources such as face and/or skin detection for cartoon applications.
  • an apparatus for cartoonizing an image signal having an object of interest comprises detecting means for detecting a feature of the object, and receiving means for receiving an input signal, the input signal relating to a characteristic of the object.
  • the apparatus further comprises image processing means that is configured to automatically adapt the image signal based on the received input signal and/or the detected feature.
  • the invention has the advantage of being able to automatically adapt an image signal based on an input signal and/or detected feature of the object being viewed.
  • a method of cartoonizing an image signal having an object of interest comprises the steps of detecting a feature of the object, and receiving an input signal relating to a characteristic of the object.
  • the image signal is automatically adapted based on the received input signal and/or the detected feature.
  • Fig. 1 shows a first embodiment of the present invention
  • Fig. 2 shows a second embodiment of the present invention
  • Fig. 3 shows a third embodiment of the present invention.
  • a first embodiment of the invention in which a skin detection unit 1 receives an image signal from a sensor or camera 3.
  • a video processor 5 receives the image signal from the camera 3, and a skin detection signal from the skin detection unit 1.
  • the video processor 5 processes the image data to produce an output video signal 9 for display on a display means (not shown).
  • the video processor 5 is configured to change the skin color based on an input signal 7.
  • the input 7 relates to a characteristic of the image being viewed.
  • the input signal 7 may relate to an emotional characteristic of the person being viewed.
  • the emotion of the person being viewed can be detected from the tone of voice of that person (e.g. the average pitch of the voice).
  • the emotion can be detected by means of a separate infrared camera (not shown), which detects heat from the face of the person being viewed.
  • the video processor 5 is configured to automatically adapt the image signal accordingly.
  • the skin color of the person is changed according to the emotion of the person.
  • the skin color of the person could be changed to red when an angry tone is detected, or grey when a calm tone is detected.
  • the skin color could be changed to red when the infrared camera detects an increase in heat dissipation, or grey when less heat is detected.
  • a user can configure the system such that the adaptation carried out by the video processor 5 is programmable.
  • the user can configure a settings table stored in a memory, to select the input condition that triggers an adaptation by the video processor 5, and a corresponding output condition for each input signal 5.
  • the settings table maps a received input signal with an adaptation process to be performed by the video processor 5.
  • the adaptation carried out by the video processor 5 may comprise other forms of video processing, for example facial texturing may also be applied.
  • the image signal can be automatically changed in accordance with an input signal relating to a characteristic of the person being viewed.
  • the cartoonizing involves a form of emotional conditioning.
  • Fig. 2 shows a second embodiment of the invention.
  • a feature extraction unit 21 is provided for detecting a feature in the object being viewed by a sensor or camera 23.
  • the feature extraction unit 21 may be configured to detect a feature in the face of a person being viewed.
  • a video processor 25 receives the image signal from the camera 23, and a feature extraction signal from the feature extraction unit 21.
  • the feature may be, for example, a left eye, a right eye, a left cheek, a right cheek, a chin, a left ear, a right ear, the top of the head, a left eyebrow, a right eyebrow, a beard, a nose or a mouth, etc.
  • the video processor 25 is configured to alter or adapt the image signal, by superimposing a secondary feature onto the image.
  • the secondary feature is preferably positioned in a predetermined relationship to the feature originally detected, for example on or next to the feature originally detected.
  • the object being superimposed may be, for example, sunglasses, a hat, a beard, a tattoo, or any other feature chosen by a user.
  • a user may select in advance which secondary feature is to be automatically superimposed onto the image signal, for example by configuring a settings table to map detected features with secondary features.
  • the system could be configured to automatically superimpose a pair of spectacles onto the eyes of the person being viewed, or a beard onto the person's chin.
  • the second embodiment can also be configured to automatically superimpose a secondary feature according to an emotional characteristic of the person being viewed.
  • the emotion of the person could be determined from the voice of the person, or heat detected from a separate infrared camera.
  • a set of horns could be placed on the head of the person being viewed, or smoke arranged to appear from the person's ears or forehead.
  • the background of a scene could be automatically changed according to the emotional characteristic of the person being viewed.
  • Fig. 3 shows a further embodiment of the invention, comprising a visual light camera 33 and an infrared (IR) or near infrared (nIR) camera 34.
  • a face and/or skin detection unit 31 receives the signals from the visual light camera 33 and the IR camera 34, and based on the two received signals, an improved face/skin tone detection is carried out.
  • a video processor 35 receives an image signal from the visual light camera 33, plus a face/skin detection signal from the face/skin detection unit 31.
  • the video processor 35 is configured to change the skin color based on an input signal 37.
  • the input signal 37 relates to a characteristic of the image being viewed.
  • the input signal 37 may relate to an emotional characteristic of the person being viewed.
  • the emotion of the person being viewed can be detected from the tone of voice of that person (e.g. the average pitch of the voice).
  • the emotion can be detected using the infrared camera 34, which detects heat from the face of the person being viewed.
  • the video processor 35 is configured to adapt the image signal accordingly.
  • the skin color of the person is changed according to the emotion of the person. For example, the skin color of the person could be changed to red when an angry tone is detected, or grey when a calm tone is detected. Likewise, the skin color could be changed to red when the infrared camera detects an increase in heat dissipation, or grey when less heat is detected.
  • a cartoon apparatus that automatically adapts an image signal in accordance with an input signal relating to a characteristic of an image being viewed, and/or a feature detected in the image being viewed.
  • any reference signs placed between parentheses shall not be construed as limiting the claim.
  • the word "a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
  • the invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the device claim enumerating several means, several ofthese means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

Cette invention concerne un processeur vidéo (5) qui reçoit un signal d'image d'une caméra (3), et un signal de détection de peau émis par une unité de détection de peau (1). Le processeur vidéo (5) traite les données d'image et produit un signal de sortie vidéo (9) pour affichage sur écran (non montré). Le processeur vidéo (5) est conçu pour adapter automatiquement le signal d'image, par exemple en vue d'un changement de couleur de la peau, à partir d'un signal d'entrée (7). Le signal d'entrée (7) se rapporte, par exemple, à une caractéristique émotionnelle de la personne observée. L'émotion ressentie par la personne peut être détectée à partir du ton de la voix de cette personne (tel que la hauteur moyenne de la voix), ou au moyen d'une caméra à infrarouges distincte détectant la chaleur dégagée par le visage de la personne observée.. Le processeur vidéo (5) est conçu pour adapter le signal d'image en fonction du signal d'entrée (7) représentant une caractéristique de l'objet observé. A titre d'exemple, on modifie la couleur de la peau de la personne en fonction de son état émotif. On peut ainsi faire virer la couleur de la peau au rouge en cas de détection d'un accent de colère, ou au gris en cas de détection d'un ton posé.
PCT/IB2005/052162 2004-07-02 2005-06-29 Traitement video WO2006003625A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP04103125.3 2004-07-02
EP04103125 2004-07-02

Publications (1)

Publication Number Publication Date
WO2006003625A1 true WO2006003625A1 (fr) 2006-01-12

Family

ID=34972268

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2005/052162 WO2006003625A1 (fr) 2004-07-02 2005-06-29 Traitement video

Country Status (2)

Country Link
TW (1) TW200617804A (fr)
WO (1) WO2006003625A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104410819A (zh) * 2014-11-24 2015-03-11 苏州福丰科技有限公司 一种用于人耳检测的dsp装置

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5426460A (en) * 1993-12-17 1995-06-20 At&T Corp. Virtual multimedia service for mass market connectivity
US5710590A (en) * 1994-04-15 1998-01-20 Hitachi, Ltd. Image signal encoding and communicating apparatus using means for extracting particular portions of an object image
WO2001077976A2 (fr) * 2000-03-28 2001-10-18 Eyeweb, Inc. Segmentation electronique visant a autoriser des achats en ligne d'articles portables et de services cosmetiques
US20030117485A1 (en) * 2001-12-20 2003-06-26 Yoshiyuki Mochizuki Virtual television phone apparatus

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5426460A (en) * 1993-12-17 1995-06-20 At&T Corp. Virtual multimedia service for mass market connectivity
US5710590A (en) * 1994-04-15 1998-01-20 Hitachi, Ltd. Image signal encoding and communicating apparatus using means for extracting particular portions of an object image
WO2001077976A2 (fr) * 2000-03-28 2001-10-18 Eyeweb, Inc. Segmentation electronique visant a autoriser des achats en ligne d'articles portables et de services cosmetiques
US20030117485A1 (en) * 2001-12-20 2003-06-26 Yoshiyuki Mochizuki Virtual television phone apparatus

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KSHIRSAGAR S ET AL IEEE COMPUTER SOCIETY: "Personalized face and speech communication over the internet", PROCEEDINGS IEEE 2001 VIRTUAL REALITY. (VR). YOKOHAMA, JAPAN, MARCH 13, 13 March 2001 (2001-03-13), pages 37 - 44, XP010535482, ISBN: 0-7695-0948-7 *
N. MAGNENAT-THALMANN ET AL.: "face to virtual face", PROCEEDINGS OF THE IEEE, vol. 86, no. 5, May 1998 (1998-05-01), usa, pages 870 - 883, XP002347421 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104410819A (zh) * 2014-11-24 2015-03-11 苏州福丰科技有限公司 一种用于人耳检测的dsp装置

Also Published As

Publication number Publication date
TW200617804A (en) 2006-06-01

Similar Documents

Publication Publication Date Title
US11288504B2 (en) Iris liveness detection for mobile devices
US9792490B2 (en) Systems and methods for enhancement of facial expressions
US9443307B2 (en) Processing of images of a subject individual
EP3555799B1 (fr) Procédé de sélection de trames utilisées dans un traitement de visage
KR20190038594A (ko) 얼굴 인식-기반 인증
JP7092108B2 (ja) 情報処理装置、情報処理方法、及びプログラム
US20060110014A1 (en) Expression invariant face recognition
CN104170374A (zh) 在视频会议期间修改参与者的外观
KR20050007427A (ko) 절반의 얼굴 이미지를 이용하는 얼굴-인식
JP2008236141A (ja) 画像処理装置、画像処理方法
CN105960801B (zh) 增强视频会议
US7023454B1 (en) Method and apparatus for creating a virtual video of an object
CN110276308A (zh) 图像处理方法及装置
JP7068710B2 (ja) 情報処理装置、情報処理方法、情報処理プログラム
KR101919090B1 (ko) 3차원 깊이정보 및 적외선정보에 기반하여 생체여부의 확인을 행하는 얼굴인식 장치 및 방법
CN110309693A (zh) 多层次状态侦测系统与方法
WO2006003625A1 (fr) Traitement video
US20230046710A1 (en) Extracting information about people from sensor signals
TWI717008B (zh) 生物特徵資料庫的建立方法、臉部識別方法與系統
CN110765969A (zh) 一种拍照方法及系统
KR102439216B1 (ko) 인공지능 딥러닝 모델을 이용한 마스크 착용 얼굴 인식 방법 및 서버
JP3245447U (ja) 顔認識システム
WO2022185596A1 (fr) Système d'estimation, procédé d'estimation et programme
KR20180114605A (ko) 안면인식방법
Shikata et al. Development of vision based meeing support system for hearing impaired

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

122 Ep: pct application non-entry in european phase