WO2006003625A1 - Traitement video - Google Patents
Traitement video Download PDFInfo
- Publication number
- WO2006003625A1 WO2006003625A1 PCT/IB2005/052162 IB2005052162W WO2006003625A1 WO 2006003625 A1 WO2006003625 A1 WO 2006003625A1 IB 2005052162 W IB2005052162 W IB 2005052162W WO 2006003625 A1 WO2006003625 A1 WO 2006003625A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- person
- input signal
- viewed
- feature
- image
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/80—2D [Two Dimensional] animation, e.g. using sprites
Definitions
- the present invention relates to a video processing apparatus and method, and in particular, to a video processing apparatus and method involving cartoonizing.
- Video communication is increasingly being used in numerous applications such as video telephones, video conferencing, television-collaboration, shared virtual table environments, and so on.
- face detection and recognition is being actively researched in order to enhance the services provided by applications such as video conferencing.
- face detection is used in video conferencing systems to create a virtual conference room, whereby participants of the meeting are seated around a virtual table.
- Numerous approaches have been used to assist in face detection, including techniques such as feature invariant approaches, appearance-based approaches, and wavelength analysis.
- Feature extraction methods utilize various properties of the face and skin to isolate and extract data, such as "eye" data.
- Popular methods include skin color segmentation, principle component analysis, eigenspace modeling, histogram analysis and texture analysis.
- face detection and skin detection methods are currently used in applications such as creating virtual video conferencing systems, or face recognition systems for security applications.
- the aim of the present invention is to provide a video processing apparatus and method that utilizes information received from sources such as face and/or skin detection for cartoon applications.
- an apparatus for cartoonizing an image signal having an object of interest comprises detecting means for detecting a feature of the object, and receiving means for receiving an input signal, the input signal relating to a characteristic of the object.
- the apparatus further comprises image processing means that is configured to automatically adapt the image signal based on the received input signal and/or the detected feature.
- the invention has the advantage of being able to automatically adapt an image signal based on an input signal and/or detected feature of the object being viewed.
- a method of cartoonizing an image signal having an object of interest comprises the steps of detecting a feature of the object, and receiving an input signal relating to a characteristic of the object.
- the image signal is automatically adapted based on the received input signal and/or the detected feature.
- Fig. 1 shows a first embodiment of the present invention
- Fig. 2 shows a second embodiment of the present invention
- Fig. 3 shows a third embodiment of the present invention.
- a first embodiment of the invention in which a skin detection unit 1 receives an image signal from a sensor or camera 3.
- a video processor 5 receives the image signal from the camera 3, and a skin detection signal from the skin detection unit 1.
- the video processor 5 processes the image data to produce an output video signal 9 for display on a display means (not shown).
- the video processor 5 is configured to change the skin color based on an input signal 7.
- the input 7 relates to a characteristic of the image being viewed.
- the input signal 7 may relate to an emotional characteristic of the person being viewed.
- the emotion of the person being viewed can be detected from the tone of voice of that person (e.g. the average pitch of the voice).
- the emotion can be detected by means of a separate infrared camera (not shown), which detects heat from the face of the person being viewed.
- the video processor 5 is configured to automatically adapt the image signal accordingly.
- the skin color of the person is changed according to the emotion of the person.
- the skin color of the person could be changed to red when an angry tone is detected, or grey when a calm tone is detected.
- the skin color could be changed to red when the infrared camera detects an increase in heat dissipation, or grey when less heat is detected.
- a user can configure the system such that the adaptation carried out by the video processor 5 is programmable.
- the user can configure a settings table stored in a memory, to select the input condition that triggers an adaptation by the video processor 5, and a corresponding output condition for each input signal 5.
- the settings table maps a received input signal with an adaptation process to be performed by the video processor 5.
- the adaptation carried out by the video processor 5 may comprise other forms of video processing, for example facial texturing may also be applied.
- the image signal can be automatically changed in accordance with an input signal relating to a characteristic of the person being viewed.
- the cartoonizing involves a form of emotional conditioning.
- Fig. 2 shows a second embodiment of the invention.
- a feature extraction unit 21 is provided for detecting a feature in the object being viewed by a sensor or camera 23.
- the feature extraction unit 21 may be configured to detect a feature in the face of a person being viewed.
- a video processor 25 receives the image signal from the camera 23, and a feature extraction signal from the feature extraction unit 21.
- the feature may be, for example, a left eye, a right eye, a left cheek, a right cheek, a chin, a left ear, a right ear, the top of the head, a left eyebrow, a right eyebrow, a beard, a nose or a mouth, etc.
- the video processor 25 is configured to alter or adapt the image signal, by superimposing a secondary feature onto the image.
- the secondary feature is preferably positioned in a predetermined relationship to the feature originally detected, for example on or next to the feature originally detected.
- the object being superimposed may be, for example, sunglasses, a hat, a beard, a tattoo, or any other feature chosen by a user.
- a user may select in advance which secondary feature is to be automatically superimposed onto the image signal, for example by configuring a settings table to map detected features with secondary features.
- the system could be configured to automatically superimpose a pair of spectacles onto the eyes of the person being viewed, or a beard onto the person's chin.
- the second embodiment can also be configured to automatically superimpose a secondary feature according to an emotional characteristic of the person being viewed.
- the emotion of the person could be determined from the voice of the person, or heat detected from a separate infrared camera.
- a set of horns could be placed on the head of the person being viewed, or smoke arranged to appear from the person's ears or forehead.
- the background of a scene could be automatically changed according to the emotional characteristic of the person being viewed.
- Fig. 3 shows a further embodiment of the invention, comprising a visual light camera 33 and an infrared (IR) or near infrared (nIR) camera 34.
- a face and/or skin detection unit 31 receives the signals from the visual light camera 33 and the IR camera 34, and based on the two received signals, an improved face/skin tone detection is carried out.
- a video processor 35 receives an image signal from the visual light camera 33, plus a face/skin detection signal from the face/skin detection unit 31.
- the video processor 35 is configured to change the skin color based on an input signal 37.
- the input signal 37 relates to a characteristic of the image being viewed.
- the input signal 37 may relate to an emotional characteristic of the person being viewed.
- the emotion of the person being viewed can be detected from the tone of voice of that person (e.g. the average pitch of the voice).
- the emotion can be detected using the infrared camera 34, which detects heat from the face of the person being viewed.
- the video processor 35 is configured to adapt the image signal accordingly.
- the skin color of the person is changed according to the emotion of the person. For example, the skin color of the person could be changed to red when an angry tone is detected, or grey when a calm tone is detected. Likewise, the skin color could be changed to red when the infrared camera detects an increase in heat dissipation, or grey when less heat is detected.
- a cartoon apparatus that automatically adapts an image signal in accordance with an input signal relating to a characteristic of an image being viewed, and/or a feature detected in the image being viewed.
- any reference signs placed between parentheses shall not be construed as limiting the claim.
- the word "a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
- the invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the device claim enumerating several means, several ofthese means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
- Processing Or Creating Images (AREA)
Abstract
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP04103125.3 | 2004-07-02 | ||
EP04103125 | 2004-07-02 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006003625A1 true WO2006003625A1 (fr) | 2006-01-12 |
Family
ID=34972268
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2005/052162 WO2006003625A1 (fr) | 2004-07-02 | 2005-06-29 | Traitement video |
Country Status (2)
Country | Link |
---|---|
TW (1) | TW200617804A (fr) |
WO (1) | WO2006003625A1 (fr) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104410819A (zh) * | 2014-11-24 | 2015-03-11 | 苏州福丰科技有限公司 | 一种用于人耳检测的dsp装置 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5426460A (en) * | 1993-12-17 | 1995-06-20 | At&T Corp. | Virtual multimedia service for mass market connectivity |
US5710590A (en) * | 1994-04-15 | 1998-01-20 | Hitachi, Ltd. | Image signal encoding and communicating apparatus using means for extracting particular portions of an object image |
WO2001077976A2 (fr) * | 2000-03-28 | 2001-10-18 | Eyeweb, Inc. | Segmentation electronique visant a autoriser des achats en ligne d'articles portables et de services cosmetiques |
US20030117485A1 (en) * | 2001-12-20 | 2003-06-26 | Yoshiyuki Mochizuki | Virtual television phone apparatus |
-
2005
- 2005-06-29 WO PCT/IB2005/052162 patent/WO2006003625A1/fr active Application Filing
- 2005-06-30 TW TW094122203A patent/TW200617804A/zh unknown
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5426460A (en) * | 1993-12-17 | 1995-06-20 | At&T Corp. | Virtual multimedia service for mass market connectivity |
US5710590A (en) * | 1994-04-15 | 1998-01-20 | Hitachi, Ltd. | Image signal encoding and communicating apparatus using means for extracting particular portions of an object image |
WO2001077976A2 (fr) * | 2000-03-28 | 2001-10-18 | Eyeweb, Inc. | Segmentation electronique visant a autoriser des achats en ligne d'articles portables et de services cosmetiques |
US20030117485A1 (en) * | 2001-12-20 | 2003-06-26 | Yoshiyuki Mochizuki | Virtual television phone apparatus |
Non-Patent Citations (2)
Title |
---|
KSHIRSAGAR S ET AL IEEE COMPUTER SOCIETY: "Personalized face and speech communication over the internet", PROCEEDINGS IEEE 2001 VIRTUAL REALITY. (VR). YOKOHAMA, JAPAN, MARCH 13, 13 March 2001 (2001-03-13), pages 37 - 44, XP010535482, ISBN: 0-7695-0948-7 * |
N. MAGNENAT-THALMANN ET AL.: "face to virtual face", PROCEEDINGS OF THE IEEE, vol. 86, no. 5, May 1998 (1998-05-01), usa, pages 870 - 883, XP002347421 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104410819A (zh) * | 2014-11-24 | 2015-03-11 | 苏州福丰科技有限公司 | 一种用于人耳检测的dsp装置 |
Also Published As
Publication number | Publication date |
---|---|
TW200617804A (en) | 2006-06-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11288504B2 (en) | Iris liveness detection for mobile devices | |
US9792490B2 (en) | Systems and methods for enhancement of facial expressions | |
US9443307B2 (en) | Processing of images of a subject individual | |
EP3555799B1 (fr) | Procédé de sélection de trames utilisées dans un traitement de visage | |
KR20190038594A (ko) | 얼굴 인식-기반 인증 | |
JP7092108B2 (ja) | 情報処理装置、情報処理方法、及びプログラム | |
US20060110014A1 (en) | Expression invariant face recognition | |
CN104170374A (zh) | 在视频会议期间修改参与者的外观 | |
KR20050007427A (ko) | 절반의 얼굴 이미지를 이용하는 얼굴-인식 | |
JP2008236141A (ja) | 画像処理装置、画像処理方法 | |
CN105960801B (zh) | 增强视频会议 | |
US7023454B1 (en) | Method and apparatus for creating a virtual video of an object | |
CN110276308A (zh) | 图像处理方法及装置 | |
JP7068710B2 (ja) | 情報処理装置、情報処理方法、情報処理プログラム | |
KR101919090B1 (ko) | 3차원 깊이정보 및 적외선정보에 기반하여 생체여부의 확인을 행하는 얼굴인식 장치 및 방법 | |
CN110309693A (zh) | 多层次状态侦测系统与方法 | |
WO2006003625A1 (fr) | Traitement video | |
US20230046710A1 (en) | Extracting information about people from sensor signals | |
TWI717008B (zh) | 生物特徵資料庫的建立方法、臉部識別方法與系統 | |
CN110765969A (zh) | 一种拍照方法及系统 | |
KR102439216B1 (ko) | 인공지능 딥러닝 모델을 이용한 마스크 착용 얼굴 인식 방법 및 서버 | |
JP3245447U (ja) | 顔認識システム | |
WO2022185596A1 (fr) | Système d'estimation, procédé d'estimation et programme | |
KR20180114605A (ko) | 안면인식방법 | |
Shikata et al. | Development of vision based meeing support system for hearing impaired |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Country of ref document: DE |
|
122 | Ep: pct application non-entry in european phase |