TWI318108B - A real-time face detection under complex backgrounds - Google Patents

A real-time face detection under complex backgrounds Download PDF

Info

Publication number
TWI318108B
TWI318108B TW94142001A TW94142001A TWI318108B TW I318108 B TWI318108 B TW I318108B TW 94142001 A TW94142001 A TW 94142001A TW 94142001 A TW94142001 A TW 94142001A TW I318108 B TWI318108 B TW I318108B
Authority
TW
Taiwan
Prior art keywords
face
block
sub
wavelet
blocks
Prior art date
Application number
TW94142001A
Other languages
Chinese (zh)
Other versions
TW200719871A (en
Inventor
Jing Wein Wang
Original Assignee
Univ Nat Kaohsiung Applied Sci
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Univ Nat Kaohsiung Applied Sci filed Critical Univ Nat Kaohsiung Applied Sci
Priority to TW94142001A priority Critical patent/TWI318108B/en
Publication of TW200719871A publication Critical patent/TW200719871A/en
Application granted granted Critical
Publication of TWI318108B publication Critical patent/TWI318108B/en

Links

Landscapes

  • Image Analysis (AREA)
  • Image Processing (AREA)

Description

1318108 九、發明說明: 【發明所屬之技術領域】 =發明係有關於一種適用於複雜背景的即時人 ΐ正確,尤其是指—種利用三步驟人臉偵測演算法,^以 人:正確位置與大小,並進-步找出限睛和 置之適用於複雜背景的即時人臉_方法 【先前技術】 按,隨著有線及無線網際網路應用日漸普及、 易化時代的來臨’傳統的保密與認證 = =:證!:錄,使用者除需記憶多組密以 外’早純密碼認證更有可能被盜用或入侵的产 双J而,利用個人獨特的生物特徵所辨識之「生理= 碼」’像是藉由人臉、指紋、個人簽名、與虹膜辨識 „複製或遭竊的特性,可以真正有效解決安全 問題。在所有生物認證技術中,人臉辨識是其 辨識者最自然及最友善的方法,且其為一非侵犯性系統 被辨識者無需要做特定的辨識程序動作’即可完成辨識。 因此較不會引起侵犯人權之異議,可以有效的達到過濾嫌 犯的目的。況且人臉也是人和人之間最明顯差異、同時亦 是隨時隨地具備的特徵。 現有方法對人臉偵測技術性能的探討相當多,但大多 數是靜態頭及肩(Head-and-Shou 1 der )人臉資料庫的測試 分析,如我國專利420939提出以連續晝面來判斷時間前 後像素差,進行動作機率與膚色完整度的對應;我國專利 505892 出以膚色、移動及輪廊資訊來找出人臉可此區 1318108 域’在匯入新舊人臉記錄器比對後以進行人臉追蹤;我國 專利550517提出強化暗色像素與其周圍亮色像素的對比 度並增加暗色像素與其周圍亮色像素的亮度差異,以债測 影像中之可能眼睛片段(Eye-Analogue Segment);我國專 利569148提出對輸入影像執行色彩分割(Color Segment),再利用人臉區塊中與膚色相異的特徵子區塊來 定位人臉特徵;美國專利6, 661,907提出應用色彩分割影 像内容,再從中搜尋最具有膚色的區塊;美國專利 6, 792, 135提出利用邊緣密度(Edge Density)及具關聯性 的樣本模組(Template Module)來搜尋人臉區塊;美國專 利6, 816, 611在偵測出膚色區塊後,利用基因演算法搜尋 具拋物線形狀的人臉區塊。 综合以上得知’不分中外膚色仍是人臉彳貞測技術的主 流;雖然這些方法各擅其長,然而在動態的實況(Live)人 臉偵測時則需面對更實際的問題:(1)色彩模型的選擇與 參數設定是否能維持膚色在窗戶邊、日光燈下、有陰影、 人臉受遮蔽、及似膚色的背景或衣著時的穩定性;(2)人 臉看成非人臉(False Reject),而把非人臉看成人臉 (False Acceptance)的錯誤率;(3)及時偵測的速度。已 有的實況人臉偵測論文很少,在SCI期刊的檢索中,僅能 找到有連續影像資料串(Video Sequence)檔案輸入的測試 、報告。就實況人臉偵測的實驗設計而言,這些文獻所能提 供的性能分析頗為有限。因為有關人數、視角、背景、表 情、物距、及光源變化等條件,並未經歷廣泛且深入的大 量數據探究,部份取像條件仍是設限的情況。 另一方面’在商用的套裝軟體方面如知名的 13181081318108 IX. Description of the invention: [Technical field to which the invention belongs] = The invention relates to an instant human being that is suitable for complex backgrounds, in particular, a three-step face detection algorithm, ^ people: correct position And size, and step-by-step to find out the limit and the real face for complex backgrounds _ method [previous technology] Press, with the increasing popularity of wired and wireless Internet applications, the era of easy access, 'traditional secrecy And authentication = =: certificate!: Record, in addition to the need to memorize multiple sets of secrets, 'early pure password authentication is more likely to be stolen or invaded by the production of double J, using the unique biometric characteristics of the individual identified by the "physiological = code "Like the characteristics of face, fingerprint, personal signature, and iris recognition" copying or stolen, it can really solve the security problem. Among all biometric authentication technologies, face recognition is the most natural and most identifiable. A friendly method, and it is a non-invasive system that is identifiable without the need to perform a specific identification procedure action to complete the identification. Therefore, it does not cause objections to human rights violations. Effectively, the purpose of filtering suspects is achieved. Moreover, the face is also the most obvious difference between people and people, and it is also a feature that can be used anytime and anywhere. The existing methods have a lot of discussion on the performance of face detection technology, but most of them are static heads. Head-and-Shou 1 der face database analysis, such as our patent 420939 proposed to use continuous face to judge the pixel difference before and after the time, the action probability and skin color integrity; Chinese patent 505892 Skin color, movement and wheel information to find the face can be in this area 1318108 domain 'after importing new and old face recorders for face tracking; China Patent 550517 proposes to strengthen the contrast between dark pixels and bright pixels around them Increasing the difference in brightness between the dark pixels and the bright pixels around them, to measure the possible eye segments in the image (Eye-Analogue Segment); Chinese Patent 569148 proposes to perform color segmentation on the input image, and then use the face segment and Feature sub-blocks with different skin colors to locate facial features; U.S. Patent 6,661,907 proposes to apply color segmentation image content, and then Search for the most skin-colored blocks; US Patent 6,792, 135 proposes the use of Edge Density and associated Template Modules to search for face blocks; US Patent 6,816,611 After detecting the skin color block, the gene algorithm is used to search for the face block with parabolic shape. The above is known that 'no matter the color of Chinese and foreign skin is still the mainstream of face speculation technology; although these methods are good at each other However, in dynamic live (Live) face detection, you need to face more practical problems: (1) whether the color model selection and parameter setting can maintain the skin color under the window, under the fluorescent light, with shadow, face is affected Shading, and the background color of the skin or the stability of clothing; (2) the face of the face as a non-face (False Reject), and the non-human face of the adult face (False Acceptance) error rate; (3) timely detection Speed measured. There are very few live face detection papers. In the SCI journal search, only tests and reports with continuous video sequence file input can be found. For the experimental design of live face detection, the performance analysis provided by these documents is quite limited. Because of the conditions, the angle of view, the background, the situation, the object distance, and the change of the light source, there is no extensive and in-depth study of the large amount of data, and some of the image acquisition conditions are still limited. On the other hand, in the case of commercial software packages, such as the well-known 1318108

Identix、BioID等公司,雖然有人臉偵測辨識的 SDK(Software Development Kit)推出,如 Identix 公司 的FacelT SDK ’以協助研發人員開發應用軟體。但是在實 際使用時,對於其中函式的呼叫參數無法知其所以然、函 式功能也不能做修改’當遇到False Acceptance或False Reject時更是苦無對策。另一方面,SDK高昂的價格更是 令人卻步’因而有必要發展自有的技術以利國家經濟發 展。 【發明内容】 本發明為適用於複雜背景的即時人臉偵測方法,此演 算法主要有三大步驟:首先,經修正的瓜〇彩色模型被用 來提供人臉可能區塊的線索,在這些區塊中,小波轉換特 徵技術被運用以迅速地移除大多數的非人臉區塊;之後, 支援向量機演算法負責分辨由空間模板匹配法所篩選出 的可能人臉候選區塊,藉由臉部各器官間所提供的有效鑑 別資訊,得以判定人臉區塊的真偽並取得正確位置資訊; 最後,應用小波係數分散度的方法以進一步萃取出眼睛和 嘴巴的坐標,再依據眼、嘴三角形並參酌人體形態學以橢 圓框出人臉區塊。 |[實施方式】 本發明係採用先做眼睛及嘴巴萃取、再進行人臉區域 搜尋、及人臉驗証的三階段策略;如第一圖所示為本發明 所設計的系統功能架構流程圖。系統主要的功能模組共有 以下四個方塊··僅用於實況測試時彩色影像使用的膚色區 域偵測、小波表示法、人臉候選區塊選取、及人臉區塊辨 識輸出。主要步驟詳述如下: 1318108 (A) 動態實況測試··利用膚色來檢測出輸入影像中可能為 人臉的區塊。對於輸入影像的每一個像素/(x,〆),求 取其在左、C、方之參數設定值分別為L 〇、〇. 95、j 〇 時的彩色模型轉換後之值,並以線性切割之方法 取出影像中呈現膚色之區塊。 (B) 小波表示法(Discrete Wavelet Frames,DWF): 在小波分析中’尺度函數办)(Sca!ing Function)用來 分解出輸入信號的低頻分量,而小波函數^(々(Wavelet Function)則用來分解出輸入信號的高頻分量。兩函數 間的關聯性可由以下算式(Two-Scale Difference Equation)得知 &lt;Κχ、= 4ϊγ^φ、2χ-θ, (1) k ⑻兴2jc-A:)。 (2) k 其中 罐)=〈也+1,,,也,,〉, (3) 洲=〈〜,,么,,〉, (4) 分別為低通及高通正交鏡面濾波器(Quadrature Mirror Filters , QMF), g(k)=(-i)kh(i-k) (5) 同時,低通濾波器須滿足限制條件 » (6) 為力的之轉換(/-transform)。在hd(z)=i的初期條 件下,低通及高通濾波器可以遞迴的形式來表示不同 寬度的滤波器 1318108Identix, BioID and other companies, although the SDK (Software Development Kit) for face detection, such as Identix's FacelT SDK ‘to help developers develop application software. However, in actual use, the call parameters of the function are not known, and the function of the function cannot be modified. When it encounters False Acceptance or False Reject, it is even more difficult. On the other hand, the high price of the SDK is even more daunting. It is therefore necessary to develop its own technology for the economic development of the country. SUMMARY OF THE INVENTION The present invention is an instant face detection method suitable for complex backgrounds. The algorithm mainly has three major steps: First, the modified color model of the melon is used to provide clues about possible blocks of the face. In the block, the wavelet transform feature technique is applied to quickly remove most non-human face blocks; after that, the support vector machine algorithm is responsible for distinguishing possible face candidate blocks selected by the spatial template matching method. From the effective identification information provided between the organs of the face, it is possible to determine the authenticity of the face block and obtain the correct position information. Finally, the wavelet coefficient dispersion method is applied to further extract the coordinates of the eyes and the mouth, and then according to the eye. The mouth triangle and the human body morphology are used to frame the human face with an oval frame. [Embodiment] The present invention adopts a three-stage strategy of first performing eye and mouth extraction, performing face region searching, and face verification; as shown in the first figure, a flowchart of a system functional architecture designed by the present invention is shown. The main function modules of the system have the following four blocks: · Skin color area detection, wavelet representation, face candidate block selection, and face block recognition output for color images only for live test. The main steps are detailed below: 1318108 (A) Dynamic Live Test··Use skin color to detect blocks in the input image that may be human faces. For each pixel /(x,〆) of the input image, the value of the color model after the left, C, and square parameter settings are L 〇, 〇. 95, j 分别, and the value is linear. The method of cutting removes the block of the skin color in the image. (B) Discrete Wavelet Frames (DWF): In the wavelet analysis, the Sca!ing Function is used to decompose the low-frequency components of the input signal, and the wavelet function ^(Wavelet Function) It is used to decompose the high-frequency component of the input signal. The correlation between the two functions can be known by the following formula (Two-Scale Difference Equation) &lt;Κχ, = 4ϊγ^φ, 2χ-θ, (1) k (8) Xing 2jc- A:). (2) k where can)=<also +1,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,, Mirror Filters, QMF), g(k)=(-i)kh(ik) (5) At the same time, the low-pass filter must satisfy the constraint » (6) as the force conversion (/-transform). In the initial condition of hd(z)=i, the low-pass and high-pass filters can be recursively represented by filters of different widths. 1318108

Hs(z) = H(zv)hs-1(z) *Gs{z)=G(z2,)Hs-i(z) (5=0, 1, ···, S-\sZ) &gt; 在信號處理領域則寫成 &gt;為*~ (9) 8s+i=[g]n-*hs 1 (1〇) [•L說明信號點數插補因子為5(通常取2),故信號 在第51階(Level)的離散小波轉換即可以列式為.尤 /(χ)=2ΧΑ*(χ) ‘、 ⑺ (δ) (11) 而近似係數Cs,,(Approximation Coefficient)與細微係 數心(Detail Coefficient)的關係式則為 , cs,t='Zcs&lt;kh(k-2t) &gt; k (12) (13) 將上述算式經稍加修改後,即可得小波轉換的不抽樣 表示法如下 :{hs(x-t),f(x)) ’ (⑷ = (^(^-0,/W) (15) 小波表示法產生較小波轉換多餘的資訊,在圖訊識別 的研究範嚀内業已獲得正面的肯定,因而我們將把這 項技術應用在這個申請案裡冀以獲得令人滿意的成 果。第二圖為輸入影像的小波水平分量表示法之輸出 結果,由於眼睛及嘴巴均呈現水平狀分佈,以致於不 太需要其他次頻帶的小波分量。 在動態貫況測試時僅轉換人臉候選區塊的部份成Hs(z) = H(zv)hs-1(z) *Gs{z)=G(z2,)Hs-i(z) (5=0, 1, ···, S-\sZ) &gt; In the field of signal processing, write &gt; is *~ (9) 8s+i=[g]n-*hs 1 (1〇) [•L indicates that the signal factor interpolation factor is 5 (usually 2), so the signal Discrete wavelet transform at the 51st order can be classified as . / / χ = 2 ΧΑ * (χ) ', (7) (δ) (11) and the approximation coefficient Cs, (Approximation Coefficient) and fine coefficients The relationship of Detail Coefficient is cs, t='Zcs&lt;kh(k-2t) &gt; k (12) (13) After the above formula is slightly modified, the wavelet transform can be obtained without sampling. The notation is as follows: {hs(xt),f(x)) ' ((4) = (^(^-0,/W) (15) Wavelet notation produces less information for smaller-wave conversion, in the study of image recognition Fan Yi has been positively recognized, so we will apply this technology to this application to achieve satisfactory results. The second picture shows the output of the wavelet horizontal component representation of the input image due to the eye and The mouths are horizontally distributed so that the wavelet components of other sub-bands are less needed. Only convert part of the candidate block of the face into

nDWFnDWF

1DWF 9 1318108 為小波係數,在靜態測試時則轉換整張影像成為小波 係數;在(A)中可能的人臉區塊裏,以小波視窗來過濾 出可能的人臉候選區塊,並進一步整併各種視窗尺寸 及相鄰位置所得到的不同尺寸大小人臉候選區塊;而 視窗整併原則係以各種不同大小波視窗濾波器人臉區 塊濾波所得之結果為參考依據,小波視窗在相鄰像素 位置間偵測出人臉候選區塊次數愈多者,其聚落所形 成之面積也相對愈大,面積相對較大者將被保留作為 整併後進一步判定是否為人臉之候選區塊,在這裡前 2/3大之區塊將會被保留’其他較小的區塊則視為非人 臉區塊予以刪除。 (C)人臉候選區塊: 臉部特徵偵測(Facial Features Detection with Varying Size Windows):由於人體姿態改變往往會造 成人臉呈現平面旋轉或深度旋轉的情形;另外,不設 限的環境(unconstrained environments)及光線更會 造成人臉出現方向性的變化。在心理學的領域裏,已 把這種執行多人臉&quot;ί貞測(Multiface Detection)時所遇 到人臉變異性很大的問題定義為與視界相關(View Dependent)的研究。為解決以上的問題,在這個功能 模組中,設計了一個多視角(Multiview)的特徵視窗如 第二圖所示,以搜尋眼睛及嘴巴的可能位置。第二圖 的特徵視窗主要係根據人臉T字型部位眼睛、鼻子、 嘴巴與臉頰週遭區域之小波係數強弱不同的人臉特性 而設計。當人臉呈現1/2側偏(Half Profile),2/3側 偏(Two-Thirds Profile),臉部遭到手、裝飾物品、 10 1318108 頭髮、鬍鬚所造成的遮蔽,或因表情變化所造成嘴形 的各式各樣外貌如大開、微開、合嘴等均需納入考量。 細合以上要點並根據各種情境的觀察可得知,眼睛區 塊的亮度通常較鼻樑及臉頰區塊暗許多,嘴巴區塊的 壳度也總是較週遭上下左右區塊暗許多。所以分別計 算視窗中各黑色及灰色方塊裡的小波係數絕對值之 和’比較區塊間彼此的大小關係便可以判定何處為候 選區塊。位於第二圖下方的判定規則(ReqUisi f 〇r W Facial Candidate),係依據黑色區塊裡的小波係數絕 對值和需大於周圍灰色子區塊裡的小波係數取絕對值 和的原則構思而成。在完成視窗的架構後,接著要進 行人臉候選區塊的搜哥。為了適應人臉的大小在不同 的取像距離下所造成的尺寸變化,這一個多視界視窗 的起始尺寸是從最小的24x24像素逐次增大5/4倍, 例如30x30、37x37、46x46·..,一直到最大的尺寸設為 起始大小的10x5/4倍。緣於照片大小不一,故當輸入 • 影像的長或寬小於增大中的視窗時,視窗大小便不可 以再變更,即最後的尺寸是以輸入影像尺寸為上限。 在第二圖中每一白色點代表符合特徵視窗判定規則的 像素,當相鄰像素都通過該規則時就會形成較大面積 的白色區塊。白色區塊面積愈大者即意味著有愈多的 人臉候選區塊重疊在一起,但實際上除了人群擁擠時 的遮蔽情況外,人臉間並不會彼此完全重疊。因此我 們有必要整併大幅度重疊的區塊,以找出人臉候選區 塊。另一方面,由於人臉候選區塊的搜尋是在各個不 同尺寸視窗下逐像素比對判定,故在每—種尺寸視窗 11 1318108 搜尋完畢後皆需進行整併作業。 使用支援向量機來辨識人臉(Face Classification by Using SVM):在這個功能模組裏, 辨識人臉/非人臉的工作將在空間領域長執行。在人臉 候選區塊,如第三圖所示,辨識方面最重要的部份就 是要先暸解所使用的分類器其複雜度與強韌性。複雜 度影響執行效率,而強韌性則顧及系統穩定度。在人 臉辨識的過程裏會遇到的錯誤類型有錯誤接受(False Accept ’ FA)與錯誤拒絕(False Reject ’ FR)兩種。若 把這兩種錯誤的相互關聯變化趨勢纟會圖,即可得出決 定系統性能的接受操作曲線(Receiver Operating Curve,ROC),系統最佳操作點則可根據實際應用需求 來設定。由於錯誤接受率與錯誤拒絕率兩變數形成所 謂的零和狀況(Zero Sum Situation),所以欲降低錯 誤接受率彺往得犧牲錯誤拒絕率。為消除人臉候選區 塊因尺寸大小不一造成後續處理步驟的困擾,尺寸規 格化(Normalization)是必要的,而這尺寸大小在對空 間領域(Spatial Domain)的分析中業已經設定為24x24 像素。當然,參數值的抉擇必須考量人臉辨識系統是 否能忍受過大或過小的人臉候選區塊可能有抽樣不足 (Under Sampling)或過度抽樣(Over Sampling)的問 題。為減輕光線對人臉候選區域的影響,等化灰階分 佈(Histogram Equal ization)是很有效的一個前處理 過程。與小波領域(Wavelet Domain)很類似的一點, 就是在空間領域内眼睛區塊也通常較鼻樑上端部位 暗,所以跟第二圖特徵視窗類似的黑-白-黑長條型空 12 間視窗也被用以搜尋人臉區塊中的T字部位中心坐 標,以做為人臉/非人臉辨識所需之基準點 (Landmark)。若此空間視窗中間位置的灰色子區塊裡 的灰階平均值同時大於左、右兩個黑色子區塊裡的灰 階平均值,則灰色子區塊的中心坐標位置便定為基準 點。可以定出基準點的人臉候選區塊很有可能就是真 正的人臉,然無法定出基準點的人臉候選區塊將被移 除。如第二圖所示,在找到T字部位之基準點後以黑 點標示’藉著此基準點之辅助得以重新修正人臉候選 區塊的面積俾去除臉部以外多餘的部份,之後人臉候 選區塊由規格化24x24像素變更為19x19像素。面積 修正後的人臉候選區塊在等化灰階分佈的處理後逕饋 入支援向量機做人臉/非人臉的辨識,經判定為人臉的 區塊參酌基準點並依未規格化前之原始區塊大小以 1:1矩形框覆蓋標示輸出結果。 以不同尺寸之特徵視窗來搜尋符合人臉器官對光 線明暗不同反應的候選區塊;依人臉候選區塊取出相 對應的原始輸入影像中的人臉區塊,首次自原始輸入 影像中取出之人臉候選區塊正規化大小值為24χ24像 素’第二次自原始輪入影像中取出之人臉候選區塊正 規化大小值19x19像素,經正規化及像素直方分佈等 化處理後’以空間領域黑_白_黑樣板取出人臉Τ字部 位的t心點,藉由Τ字部位的中心點,可將人臉候選 區塊大小重新修正’以去除多餘的背景部份,重新經 像素直方分佈等化處理後,饋人支援向量機做人臉/# 人臉之判讀’支援向量機所用之基底函數為多項式基 13 1318108 底,^[貞測照片中的人臉時準位值設為-0 · 4,實況侦測 人臉時準位值設為0。 另,請參閱第四圖所示,尚可由下列步輝做更進一步 的精確人臉器官定位與人臉切割: (C1)由於眉毛、眼睛、與嘴巴均呈現長條形,在水平 分量小波次頻帶的分佈尤其顯著,為有效的應用這項資 訊,已偵測到的輸入人臉區塊藉正規化過程調整為40x60 像素,以利小波係數分散度(Wavelet Entropy)計算,算 式如下 _1 N-11DWF 9 1318108 is the wavelet coefficient. In the static test, the whole image is converted into wavelet coefficient. In the possible face block in (A), the wavelet window is used to filter out possible face candidate blocks, and further And different sizes of face candidate blocks obtained by various window sizes and adjacent positions; and the window merging principle is based on the results obtained by various size and size window filter face block filtering, and the wavelet window is in phase. The more times the face candidate block is detected between adjacent pixel positions, the larger the area formed by the settlement, and the larger the area will be retained as a candidate block for further determination of the face. Here, the first 2/3 block will be reserved. 'Other smaller blocks are treated as non-face blocks to be deleted. (C) Facial Features Detection with Varying Size Windows: The change of the human body poses a situation in which the face is rotated or rotated in depth; in addition, there is no limit to the environment ( Unconstrained environments) and light can cause directional changes in the face. In the field of psychology, the problem of large face variability when performing multiface detection has been defined as a study of View Dependent. In order to solve the above problems, in this function module, a multiview feature window is designed as shown in the second figure to search for possible positions of eyes and mouth. The feature window of the second figure is mainly designed according to the characteristics of the face with different wavelet coefficients of the eyes, nose, mouth and cheeks around the T-shaped part of the face. When the face presents 1/2 Half Profile, 2/3 Side-Shirds Profile, the face is covered by hands, decorative items, 10 1318108 hair, beard, or caused by changes in expression The various shapes of the mouth shape, such as wide open, slightly open, and closed mouth, should be taken into consideration. Combining the above points and observing the various situations, it is known that the brightness of the eye area is usually much darker than the bridge of the nose and cheeks, and the shell of the mouth block is always much darker than the surrounding upper and lower blocks. Therefore, the sum of the absolute values of the wavelet coefficients in the black and gray squares in the window is calculated separately. Comparing the size relationships between the blocks can determine where the candidate blocks are. The decision rule (ReqUisi f 〇r W Facial Candidate) located at the bottom of the second figure is based on the principle that the absolute value of the wavelet coefficient in the black block and the wavelet coefficient in the surrounding gray sub-block are taken as the absolute value sum. . After completing the architecture of the window, it is then necessary to search for the face candidate block. In order to adapt to the size change caused by the size of the face at different imaging distances, the starting size of this multi-view window is increased by 5/4 times from the smallest 24x24 pixels, for example, 30x30, 37x37, 46x46·. . Until the maximum size is set to 10x5/4 times the starting size. Due to the different sizes of the photos, when the length or width of the image is smaller than the window in the increase, the window size cannot be changed. The final size is the upper limit of the input image size. In the second figure, each white dot represents a pixel that conforms to the feature window decision rule, and a white block of a larger area is formed when the adjacent pixels pass the rule. The larger the area of the white block, the more face candidates are overlapped, but in reality, the faces do not completely overlap each other except for the crowded situation when the crowd is crowded. Therefore, it is necessary for us to integrate blocks that overlap greatly to find candidate blocks for faces. On the other hand, since the search for the candidate blocks of the face is determined pixel-by-pixel in each different size window, the consolidation operation is required after each search for the size window 11 1318108. Face Classification by Using SVM: In this function module, the recognition of face/non-face work will be performed in the space field. In the face candidate block, as shown in the third figure, the most important part of the identification is to understand the complexity and toughness of the classifier used. Complexity affects execution efficiency, while toughness takes into account system stability. There are two types of errors that are encountered during face recognition: False Accept ( FA) and False Reject (FR). If you change the relationship between the two errors, you can get the Receiver Operating Curve (ROC) to determine the system performance. The optimal operating point of the system can be set according to the actual application requirements. Since the error acceptance rate and the error rejection rate form a so-called Zero Sum Situation, it is desirable to reduce the error acceptance rate and sacrifice the false rejection rate. In order to eliminate the problem that the candidate blocks of the face are troubled by subsequent processing steps due to different sizes, normalization is necessary, and this size has been set to 24x24 pixels in the analysis of the Spatial Domain. . Of course, the choice of parameter values must consider whether the face recognition system can tolerate excessive or too small face candidate blocks may have problems with Under Sampling or Over Sampling. In order to reduce the influence of light on the candidate region of the face, Histogram Equalization is a very effective pre-processing process. Similar to the Wavelet Domain, the eye block is usually darker than the upper end of the bridge of the nose, so the black-white-black strip-shaped 12-window window similar to the second feature window is also It is used to search the center coordinates of the T-shaped part in the face block as the reference point (Landmark) required for face/non-face recognition. If the grayscale average value in the gray sub-block in the middle of the space window is greater than the grayscale average in the left and right black sub-blocks, the center coordinate position of the gray sub-block is determined as the reference point. The face candidate block in which the reference point can be determined is likely to be a true face, and the face candidate block in which the reference point cannot be determined will be removed. As shown in the second figure, after finding the reference point of the T-shaped part, the black dot indicates that the area of the candidate block of the face can be re-corrected by the aid of the reference point, and the extra part outside the face is removed. The face candidate block is changed from normalized 24x24 pixels to 19x19 pixels. The area-corrected face candidate block is fed into the support vector machine to perform face/non-face recognition after the equalization gray scale distribution is processed, and the block determined to be the face is determined by the reference point and is not normalized. The original block size is covered by a 1:1 rectangular frame to indicate the output. The feature blocks of different sizes are used to search for candidate blocks that meet the different responses of the face organs to the light and shade; the face blocks in the corresponding original input images are taken out according to the face candidate block, and are taken out from the original input image for the first time. The normalized size of the face candidate block is 24χ24 pixels. The normalized size of the face candidate block taken out from the original round-in image is 19x19 pixels. After normalization and pixel histogram distribution, the space is used. The field black _ white _ black template takes out the t-heart point of the face part of the face. By the center point of the Τ word part, the size of the face candidate block can be re-corrected to remove the excess background part and re-pixel-rectangular After the distribution equalization process, the base function used by the support vector machine to make the face/# face interpretation 'support vector machine is the polynomial base 13 1318108 bottom, ^[the face value of the face in the photo is set to - 0 · 4, the level value is set to 0 when the face is detected live. In addition, please refer to the fourth figure, the following step can be used to make further accurate face organ positioning and face cutting: (C1) because the eyebrows, eyes, and mouth are elongated, in the horizontal component wavelet The distribution of the frequency band is particularly significant. For the effective application of this information, the detected input face block is adjusted to 40x60 pixels by the normalization process to calculate the wavelet coefficient dispersion (Wavelet Entropy), and the calculation is as follows: _1 N -1

Wavelet_entropy = — (x, y)|l〇g|i/i!S (x, y)| ( 1 6 ) (C2)於步驟(Cl)中,#=3,人/7為水平分量小波次 頻帶,使用3x3分散度視窗搜尋人臉區塊内的每—點後, 區塊内所有的小波係數值皆由此視窗的平均分散度來取 代。 (C3)經由二值化(Binary Thresholding)過程,可以 很有效的找到人臉區塊内具有較大水平分量的係數分佈 子區塊;反之,水平分量較小的係數分佈子區塊將在此遭 到排除,經由18x18的區塊樣板比對(c〇rreiati〇n Coefficient) ’面積夠大的子區塊方才列入人臉器官之考 量。 (C4)根據T字部位基準點以確立左右眼睛的兩個子區 塊’再參酌等腰眼嘴三角形之幾何結構以搜尋嘴巴子區 塊,若由於嘴巴受到遮蔽或因人臉深度旋轉造成射攝影機 無法明確取得嘴形時’人臉偵測輸出將僅提供眼睛的坐 標、同時不標示嘴巴的位置。 14 1318108 ⑽在這個步驟裡,使用點集合(㈤心駆卿心) 的方法來取得人臉器官子區塊的m⑽注意的是 此坐標大賴㈣落在_是目卩讀與嘴^心的位置;由 步驟(C4)所得之子區塊的中心坐標是展開子區塊切割 (Region Segmentation)過程的起始位置’應用像素4—相 連(4-connectivity)的子區塊成長(Region Growing)原理 後即可更新各子區塊的面積範圍,以定出子區塊的正確邊 界。在重新計算各子區塊的中心位置後,就能夠取出各子 區塊的重心坐標並加以標示。一般而言,眼睛子區塊在子 區塊成長過程時往往會納入眉毛部位或因配戴眼睛之故 而將鏡框包括進來,所以在取出眼睛子區塊的重心坐標標 示時需往下調整約1/3眼睛子區塊個的距離做為位置補 償。最後,依據眼嘴三角形的重心坐標及人臉縱橫比6:4 以橢圓形框出人臉區塊,即是人臉切割的結果。 綜上所述,本發明實施例確能達到所預期之使用功 效,又其所揭露之具體構造,不僅未曾見諸於同類產品 中,亦未曾公開於申請前,誠已完全符合專利法之規定與 要求,爰依法提出發明專利之申請,懇請惠予審查,並賜 准專利,則實感德便。 15 1318108 【圖式簡單說明】 第一圖:本發明之系統功能架構流程圖 第二圖:本發明之人臉定位功能方塊流程圖 第三圖:本發明之支援向量機用於人臉區塊辨識示意圖 第四圖:本發明之人臉器官定位與人臉切割示意圖 【主要元件符號說明】Wavelet_entropy = — (x, y)|l〇g|i/i!S (x, y)| ( 1 6 ) (C2) In step (Cl), #=3, person/7 is the horizontal component wavelet In the frequency band, after searching for each point in the face block using the 3x3 dispersion window, all wavelet coefficient values in the block are replaced by the average dispersion of the window. (C3) Through the Binary Thresholding process, the coefficient distribution sub-blocks with larger horizontal components in the face block can be effectively found; otherwise, the coefficient distribution sub-blocks with smaller horizontal components will be located here. Excluded, through the 18x18 block model comparison (c〇rreiati〇n Coefficient) 'area large enough to be included in the face organs consideration. (C4) According to the T-point reference point to establish the two sub-blocks of the left and right eyes', then consider the geometry of the isosceles eye triangle to search for the mouth sub-block, if the camera is blocked due to the mouth being blocked or due to the depth of the face rotation When the mouth shape cannot be clearly obtained, the 'face detection output will only provide the coordinates of the eye while not indicating the position of the mouth. 14 1318108 (10) In this step, use the method of point set ((5) 駆心駆心) to get the m(10) of the face organ sub-block. Note that this coordinate is too big (4) to fall in _ is the target reading and mouth ^ heart Position; the center coordinate of the sub-block obtained by the step (C4) is the starting position of the expansion sub-region cutting process. The principle of applying the pixel growing--------- The area range of each sub-block can then be updated to determine the correct boundary of the sub-block. After recalculating the center position of each sub-block, the coordinates of the center of gravity of each sub-block can be taken out and marked. In general, the eye sub-blocks are often included in the eyebrows during the growth process of the sub-blocks or include the frame because of the wearing of the eyes. Therefore, when the coordinates of the center of gravity of the sub-blocks are taken out, it is necessary to adjust downwards by about 1 The distance of the /3 eye sub-block is used as the position compensation. Finally, according to the center of gravity coordinates of the eye triangle and the aspect ratio of the face 6:4, the face block is ovalized, which is the result of face cutting. In summary, the embodiments of the present invention can achieve the expected use efficiency, and the specific structure disclosed therein has not been seen in similar products, nor has it been disclosed before the application, and has completely complied with the provisions of the Patent Law. And the request, the application for the invention of a patent in accordance with the law, please forgive the review, and grant the patent, it is really sensible. 15 1318108 [Simplified description of the drawings] First: Flow chart of the functional structure of the system of the present invention. Second figure: Flow chart of the face positioning function block of the present invention. Third figure: The support vector machine of the present invention is used for the face block. Identification diagram fourth diagram: schematic diagram of face organ positioning and face cutting of the present invention [main component symbol description]

1616

Claims (1)

1318108 十、申請專利範圍: L/種適用於複雜背景的即時人臉偵測方法’主要包括下 列步驟: 動態實況測試:利用膚色線索來快速的找出可能是人臉 的區域; 小波表示法:在動態實況測試時僅轉換人臉候選區塊的 部份成為小波係數,在靜態測試時則轉換 • 整張影像成為小波係數; 人臉候選區塊:以不同尺寸之特徵視窗來搜尋符合人臉 _ 器官對光線明暗不同反應的候選區塊。 2. —種適用於複雜背景的即時人臉偵測方法,主要包括下 列步驟: (A)利用膚色來檢測出輸入影像中可能為人臉的區塊: 對於輸入影像的每一個像素/(&gt;,7),求取其在 彩色模型轉換後之值,並以線性切割之方法取出影 像中呈現膚色之區塊; 鲁 (B)在(A)中可能的人臉區塊裏,以小波視窗來過濾出可 能的人臉候選區塊,並進一步整併各種視窗尺寸及 相鄰位置所得到的不同尺寸大小人臉候選區塊;_ (C)依人臉候選區塊取出相對應的原始輸入影像中的人 臉區塊,經正規化及像素直方分佈等化處理後,以 空間領域黑-白-黑樣板取出人臉T字部位的中心 點,藉由T字部位的中心點,可將人臉候選區塊大 小重新修正’以去除多餘的背景部份’重新經像素 直方分佈等化處理後,饋入支援向量機做人臉/非人 17 1318108 臉之判讀。 3. 如申請專利範圍第2項所述適用於複雜背景的即時人臉 偵測方法,其中,於步驊(A)中尤、C、万之參數設定值 分別為 1. 0、0. 95、1. 0。 4. 如申請專利範圍第2項戶斤述適用於複雜背景的即時人臉 偵測方法,其中,於步驟(B)中視窗整併原則係以各種 不同大小波視窗濾、波器人臉區塊渡波所得之結果為參 考依據,小波視窗在相鄰像素位置間4貞測出人臉候選區 塊次數愈多者,其聚落所形成之面積也相對愈大,面積 相對較大者將被保留作為整併後進一步判定是否為人 臉之候選區塊,在這裡前2/3大之區塊將會被保留,其 他較小的區塊則視為非人臉區塊予以刪除。 5·如申請專利範圍第2項所述適用於複雜背景的即時人臉 偵測方法,其中,於步驟(C)中首次自原始輸入影像中 取出之人臉候選區塊正規化大小值為24x24像素,第二 次自原始輸入影像中取出之人臉候選區塊正規化大小 值19x19像素。 6.如申請專利範圍第2項所述適用於複雜背景的即時人臉 偵測方法,其中,於步驟(c)中支援向量機所用之基底 函數為多項式基底,偵測照片中的人臉時準位值設為 4,實況偵測人臉時準位值設為〇。 … λ如申請專圍第2項所述適用於複雜背景 偵測方法’其中,於步驟(c)尚可由下列步驟做隹 的精確人臉器官定位與人臉切割: 尺進一步 (C1)由於眉毛、眼睛、與嘴巴均呈現長條形 量小波次頻帶的分佈尤其顯著,為有致7平分 竭用這項 18 ^ 1318108 資訊,已偵測到的輸入人臉區塊藉正規化過程調整 為40x60像素,以利小波係數分散度(Wavelet Entropy)計算,算式如下 Wavelet _ entropy = —Ej^(^y)|i〇g|^u,y)| (C2)於步驟(Cl)中’ #=3,為水平分量小波次頻 帶,使用3x3分散度視窗搜尋人臉區塊内的每一點 後,區塊内所有的小波係數值皆由此視窗的平均分 散度來取代; (C3)經由二值化(Binary Thresholding)過程,可以很 有效的找到人臉區塊内具有較大水平分量的係數 分佈子區塊;反之,水平分量較小的係數分佈子區 塊將在此遭到排除,經由18x18的區塊樣板比對 (Correlation Coefficient),面積夠大的子區塊 方才列入人臉器官之考量; (C4)根據T字部位基準點以轉立左右眼睛的兩個子區 塊,再參酌等腰眼嘴三角形之幾何結構以搜尋嘴巴 子區塊,若由於嘴巴受到遮蔽或因人臉深度旋轉造 成射攝影機無法明確取得嘴形時,人臉偵測輸出將 僅提供眼睛的坐標、同時不標示嘴巴的位置; (C5)在這個步驟裡,使用點集合(Point Aggregati〇n) 的方法來取得人臉器官子區塊的重心位置,值得注 思的疋此坐標大都數都會落在剛好是眼球與嘴巴 中心的位置;由步驟(C4)所得之子區塊的中心坐標 疋展開子區塊切割(Region Segmentation)過程的 起始位置’應用像素4-相連(4-connectivity)的子 19 1318108 區塊成長(Region Growing)原理後即可更新各子區 塊的面積範圍,以定出子區塊的正確邊界;在重新 計算各子區塊的中,讀置後’就能夠取出各子 的重心坐標並加«示;—般而言,眼睛 框r進來’所以在取出眼』 重心坐標及人⑽眼嘴三角形的 汉縱枳比6:4以橢圓形框出人臉區 塊,即疋人臉切割的結果。 201318108 X. Patent application scope: L/A kind of instant face detection method suitable for complex backgrounds mainly includes the following steps: Dynamic live test: use skin color cues to quickly find out areas that may be faces; Wavelet notation: In the dynamic live test, only the part of the face candidate block is converted into a wavelet coefficient, and in the static test, the entire image becomes a wavelet coefficient; the face candidate block: the feature window of different sizes is used to search for the face. _ Candidate blocks for different responses of organs to light and dark. 2. An instant face detection method suitable for complex backgrounds, which mainly comprises the following steps: (A) Using the skin color to detect a block that may be a human face in the input image: For each pixel of the input image / (&gt ;, 7), find the value after the color model conversion, and take out the block of the skin color in the image by linear cutting; Lu (B) in the possible face block in (A), with wavelet The window filters out possible face candidate blocks, and further refines the different size and size face candidate blocks obtained by various window sizes and adjacent positions; _ (C) extracts the corresponding original according to the face candidate block After inputting the face block in the image, after normalization and pixel histogram distribution, the center point of the T-shaped part of the face is taken out by the black-white-black model of the space field, and the center point of the T-shaped part can be The face candidate block size is re-corrected 'to remove the extra background part' and then equalized by the pixel histogram distribution, and then fed into the support vector machine to make the face/non-human 17 1318108 face interpretation. 3. In the case of the instant face detection method for complex backgrounds as described in item 2 of the patent application scope, the parameter setting values of the parameters in the step (A), C, and 10,000 are respectively 1. 0, 0.95. , 1. 0. 4. For example, the second paragraph of the patent application scope describes the instant face detection method applicable to complex backgrounds. Among them, in the step (B), the window consolidation principle is based on various sizes of window filters and wave face areas. The result obtained by the block wave is the reference basis. The more the number of face candidate blocks detected by the wavelet window between adjacent pixel positions, the larger the area formed by the settlement, the larger the area will be retained. As a candidate block for further determining whether it is a face, the first 2/3 block will be reserved, and the other smaller blocks will be deleted as non-face blocks. 5. The instant face detection method applicable to complex backgrounds as described in the second application of the patent application scope, wherein the face candidate block normalized size value taken out from the original input image for the first time in step (C) is 24x24. Pixel, the face candidate block normalized size value taken from the original input image for the second time is 19x19 pixels. 6. The instant face detection method for complex backgrounds as described in claim 2, wherein the base function used in the support vector machine in step (c) is a polynomial base, and when detecting a face in a photo The level value is set to 4, and the level value is set to 〇 when the face is detected live. ... λ as applied for the complex background detection method as described in item 2 of the application, in which step (c) can be performed by the following steps: accurate face organ positioning and face cutting: Ruler further (C1) due to eyebrows The distribution of the small-wavelet sub-bands, eyes, and mouth are particularly significant. In order to use the 18^1318108 information, the detected input face block is adjusted to 40x60 pixels by the normalization process. , Calculated by Wavelet Entropy, the formula is as follows Wavelet _ entropy = -Ej^(^y)|i〇g|^u,y)| (C2) in step (Cl) ' #=3 For the horizontal component wavelet sub-band, after searching each point in the face block using the 3x3 dispersion window, all wavelet coefficient values in the block are replaced by the average dispersion of the window; (C3) via binarization (Binary Thresholding) process, it is very effective to find the coefficient distribution sub-blocks with large horizontal components in the face block; otherwise, the coefficient distribution sub-blocks with smaller horizontal components will be excluded here, via 18x18 Block template comparison (Correlatio n Coefficient), the sub-blocks with large enough area are included in the face organ consideration; (C4) according to the T-point reference point to turn the two sub-blocks of the left and right eyes, and then consider the geometric structure of the isosceles eye triangle In order to search for the mouth block, if the camera is unable to clearly obtain the mouth shape due to the obscuration of the mouth or the deep rotation of the face, the face detection output will only provide the coordinates of the eye and not the position of the mouth; (C5) In this step, the point set (Point Aggregati〇n) method is used to obtain the position of the center of gravity of the face organ sub-block. It is worth noting that most of the coordinates will fall in the center of the eyeball and the mouth; (C4) The central coordinate of the obtained sub-block 疋 The starting position of the Region Segmentation process is applied to the principle of the Region Growing principle after applying the pixel 4-connectivity sub- 19 1318108 The area range of each sub-block can be updated to determine the correct boundary of the sub-block; in recalculating each sub-block, after reading, it is possible to take out the weight of each sub-block. Coordinates and additions «In general, the eye frame r comes in 'so the eye is taken out'. The barycentric coordinates and the human (10) eye-triangular triangle are smaller than the 6:4 oval frame, which is the monk. The result of face cutting. 20
TW94142001A 2005-11-30 2005-11-30 A real-time face detection under complex backgrounds TWI318108B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW94142001A TWI318108B (en) 2005-11-30 2005-11-30 A real-time face detection under complex backgrounds

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW94142001A TWI318108B (en) 2005-11-30 2005-11-30 A real-time face detection under complex backgrounds

Publications (2)

Publication Number Publication Date
TW200719871A TW200719871A (en) 2007-06-01
TWI318108B true TWI318108B (en) 2009-12-11

Family

ID=45073458

Family Applications (1)

Application Number Title Priority Date Filing Date
TW94142001A TWI318108B (en) 2005-11-30 2005-11-30 A real-time face detection under complex backgrounds

Country Status (1)

Country Link
TW (1) TWI318108B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI470563B (en) * 2011-04-11 2015-01-21 Intel Corp Method of detecting attributes in an image, processing system to perform image analysis processing, and computer-readable storage medium

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4479756B2 (en) 2007-07-05 2010-06-09 ソニー株式会社 Image processing apparatus, image processing method, and computer program
JP4911165B2 (en) * 2008-12-12 2012-04-04 カシオ計算機株式会社 Imaging apparatus, face detection method, and program
TWI424359B (en) * 2009-12-03 2014-01-21 Chunghwa Telecom Co Ltd Two - stage Face Recognition System and Method
TWI402479B (en) 2009-12-15 2013-07-21 Ind Tech Res Inst Depth detection method and system using thereof
TWI488496B (en) * 2010-01-20 2015-06-11 Altek Corp Face capture method for image capture device
TWI419057B (en) * 2010-09-08 2013-12-11 Univ Nat Kaohsiung Applied Sci Method and system for detecting blur fingerprint images
TW201224955A (en) 2010-12-15 2012-06-16 Ind Tech Res Inst System and method for face detection using face region location and size predictions and computer program product thereof
CN103300815B (en) * 2012-03-15 2015-05-13 凹凸电子(武汉)有限公司 Eyeball focus determination method, device and system
TWI489396B (en) * 2013-03-01 2015-06-21 First Optotech Co Ltd Image structure analysis method
CN110490029B (en) * 2018-05-15 2022-04-15 瑞昱半导体股份有限公司 Image processing method capable of performing differentiation processing on face data
CN110008802B (en) 2018-12-04 2023-08-29 创新先进技术有限公司 Method and device for selecting target face from multiple faces and comparing face recognition
TWI775006B (en) 2019-11-01 2022-08-21 財團法人工業技術研究院 Imaginary face generation method and system, and face recognition method and system using the same

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI470563B (en) * 2011-04-11 2015-01-21 Intel Corp Method of detecting attributes in an image, processing system to perform image analysis processing, and computer-readable storage medium

Also Published As

Publication number Publication date
TW200719871A (en) 2007-06-01

Similar Documents

Publication Publication Date Title
TWI318108B (en) A real-time face detection under complex backgrounds
US10339362B2 (en) Systems and methods for performing fingerprint based user authentication using imagery captured using mobile devices
KR102561723B1 (en) System and method for performing fingerprint-based user authentication using images captured using a mobile device
CN105956578B (en) A kind of face verification method of identity-based certificate information
US20220215686A1 (en) Systems and methods for performing fingerprint based user authentication using imagery captured using mobile devices
JP6650946B2 (en) System and method for performing fingerprint-based user authentication using images captured with a mobile device
WO2020000908A1 (en) Method and device for face liveness detection
JP4767971B2 (en) Distance iris recognition system
Shah et al. Iris segmentation using geodesic active contours
Genovese et al. Touchless palmprint recognition systems
Jan Segmentation and localization schemes for non-ideal iris biometric systems
Das et al. A new efficient and adaptive sclera recognition system
CN111582197A (en) Living body based on near infrared and 3D camera shooting technology and face recognition system
JP2007188504A (en) Method for filtering pixel intensity in image
CN107273812B (en) A kind of living body iris method for anti-counterfeit for authentication
CN109255319A (en) For the recognition of face payment information method for anti-counterfeit of still photo
Szczepański et al. Pupil and iris detection algorithm for near-infrared capture devices
Amjed et al. Noncircular iris segmentation based on weighted adaptive hough transform using smartphone database
Méndez-Llanes et al. On the use of local fixations and quality measures for deep face recognition
CN112801034A (en) Finger vein recognition device
何晓光 et al. Illumination normalization with morphological quotient image
Goranin et al. Evolutionary Algorithms Application Analysis in Biometric Systems.
Harakannanavar et al. Performance evaluation of face recognition based on multiple feature descriptors using Euclidean distance classifier
Rossant et al. A robust iris identification system based on wavelet packet decomposition and local comparisons of the extracted signatures
Tandon et al. An efficient age-invariant face recognition

Legal Events

Date Code Title Description
MM4A Annulment or lapse of patent due to non-payment of fees