TWI450206B - Automatic image identification method - Google Patents

Automatic image identification method Download PDF

Info

Publication number
TWI450206B
TWI450206B TW101124625A TW101124625A TWI450206B TW I450206 B TWI450206 B TW I450206B TW 101124625 A TW101124625 A TW 101124625A TW 101124625 A TW101124625 A TW 101124625A TW I450206 B TWI450206 B TW I450206B
Authority
TW
Taiwan
Prior art keywords
image
model
category
processed
majority
Prior art date
Application number
TW101124625A
Other languages
Chinese (zh)
Other versions
TW201403496A (en
Inventor
Yu Chiang Frank Wang
Chih Fan Chen
Chia Po Wei
Original Assignee
Academia Sinica
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Academia Sinica filed Critical Academia Sinica
Priority to TW101124625A priority Critical patent/TWI450206B/en
Publication of TW201403496A publication Critical patent/TW201403496A/en
Application granted granted Critical
Publication of TWI450206B publication Critical patent/TWI450206B/en

Links

Landscapes

  • Collating Specific Patterns (AREA)
  • Image Analysis (AREA)

Description

影像自動辨識方法Automatic image recognition method

本發明是關於一種影像自動辨識方法,特別是關於一種以電腦系統自動建立人臉或類似影像的辨認用影像,藉以辨認人臉或類似影像來源的方法。The present invention relates to an automatic image recognition method, and more particularly to a method for automatically identifying a face or a similar image by a computer system, thereby identifying a face or a similar image source.

科技始終來自於人性,人永遠是科學研究中最重要的課題。因此在電腦視覺中,有關人臉的研究始終是個熱門的議題。目前可見的數位相機、手機等等科技產品都有額外處理與「人」相關資訊的功能。最廣為人知的功能像是人臉偵測、表情辨識等。不同於這些已商業化的功能,人臉辨識仍是個較難的議題。為什麼會有如此差異呢?主要原因是目前數位相機或智慧手機所提供得人臉辨識功能,所需要的資訊均可適用於所有人,而非針對特定的人。舉例來說,人臉一定是兩個眼睛、一個鼻子。擷取到這樣的資訊,系統即可判定該影像極有可能是人臉影像。但是「人臉的辨識」與前述「辨識是否人臉」的功能不同。通常需要先建立資料庫,並先讓系統知道需要辨識的人是誰,才能經由計算及比較得到相對應的答案。另外還需要找到屬於每個人的影像特徵,來幫助系統辨識。像剛才所述的兩個眼睛一個鼻子、這種屬於人臉共同特徵的資訊,無法在人臉辨識下發揮作用。因此在人臉辨識中,如何蒐集「有用」的資料及如何從這些資料中找出每個人所屬的「特徵」,乃是最重要且待解決的兩個問題。Science and technology always comes from human nature, and people are always the most important subject in scientific research. Therefore, in computer vision, research on face is always a hot topic. Currently available digital cameras, mobile phones and other technology products have additional functions to deal with "people" related information. The most widely known functions are face detection, expression recognition, and so on. Unlike these commercial features, face recognition is still a difficult issue. Why is there such a difference? The main reason is that face recognition is provided by digital cameras or smart phones, and the information needed can be applied to everyone, not to specific people. For example, a human face must be two eyes and a nose. By capturing such information, the system can determine that the image is most likely a human face image. However, "face recognition" is different from the above "identification of face". It is usually necessary to first establish a database and let the system know who the person needs to be identified in order to get a corresponding answer through calculation and comparison. In addition, you need to find the image features that belong to everyone to help the system identify. The two eyes, one nose, and the information that belong to the common features of the face, can't play a role in face recognition. Therefore, in face recognition, how to collect "useful" information and how to find out the "characteristics" of each person from these materials is the most important and problem to be solved.

過去已經有許多研究學者提出不同的技術來解決人臉辨識的問題。但之前在人臉辨識的研究上都有個很嚴格的假設,就是訓練人臉模型時的圖片需要是完全乾淨的,意即不能有任何像圍巾、眼鏡、口罩等等的遮蔽物擋住人臉。如此嚴格的假設較不適用於現實的狀況。因為,在現實生活的應用中,建立人臉 辨識資料庫時所蒐集到的人臉影像,並不能要求影像皆處在如此理想的環境下拍攝。例如:通常會因為照明條件、遮蔽,或取像角度等因素影響,無法取得完整的人臉影像。In the past, many research scholars have proposed different techniques to solve the problem of face recognition. But before in the research of face recognition, there is a very strict assumption that the picture when training the face model needs to be completely clean, meaning that there can be no cover like scarves, glasses, masks, etc. to block the face. . Such strict assumptions are less applicable to real-life situations. Because, in real life applications, building faces The face images collected when identifying the database do not require that the images be shot in such an ideal environment. For example, it is usually impossible to obtain a complete face image due to factors such as lighting conditions, obscuration, or image angle.

因此,目前實有必要提供一種新穎的人臉影像自動辨識方法,可以根據不理想的影像,建立有用的人臉模型,以供辨認之用。Therefore, it is necessary to provide a novel automatic face recognition method for creating a useful face model for identification purposes based on undesired images.

同時也需要有一種新穎的人臉影像自動辨認方法,可以在影像資料特徵不足的條件下,仍能做出正確的辨識。At the same time, there is a need for a novel method for automatic recognition of face images, which can still make correct recognition under the condition that the characteristics of image data are insufficient.

同時也需要有一種影像自動辨認方法,可以根據不理想的影像,建立有用的影像模型,以供辨認之用。At the same time, there is a need for an automatic image recognition method that can create useful image models for identification based on undesirable images.

同時也需要有一種新穎的影像自動辨認方法,可以在影像資料特徵不足的條件下,仍能做出正確的辨識。At the same time, there is a need for a novel automatic image recognition method, which can still make a correct identification under the condition that the image data is insufficient.

同時也需要有一種人臉影像自動辨認系統或影像自動辨認系統,可以執行上述方法。At the same time, there is also a need for an automatic face recognition system or an automatic image recognition system that can perform the above method.

本發明的目的即是在提供一種新穎的人臉影像自動辨認方法,可以在影像資料特徵不足的條件下,仍能做出正確的辨識。The object of the present invention is to provide a novel method for automatically recognizing facial images, which can still make correct recognition under the condition that the characteristics of the image data are insufficient.

本發明的目的也在提供一種影像自動辨認方法,可以根據不理想的影像,建立有用的影像模型,以供辨認之用。The object of the present invention is also to provide an automatic image recognition method for creating a useful image model for identification based on an undesired image.

本發明的目的也在提供一種新穎的影像自動辨認方法,可以在影像資料特徵不足的條件下,仍能做出正確的辨識。The object of the present invention is also to provide a novel image automatic recognition method, which can still make a correct identification under the condition that the image data characteristics are insufficient.

本發明的目的也在提供一種人臉影像自動辨認系統,用以執行上述方法。It is also an object of the present invention to provide a face image automatic recognition system for performing the above method.

本發明的目的也在提供一種影像自動辨認系統,用以執行上述方法。It is also an object of the present invention to provide an image automatic identification system for performing the above method.

根據本發明所揭示的影像自動辨識方法及系統,乃是針對在建立辨識用影像模型時,就已存在照明條件不佳、遭到遮蔽、片段影像等因素的訓練用影像,提出解決方案。根據本發明的方法,主要包括:藉由低秩分解,精準的比對像素的誤差;藉由分離出不屬於特定標的的像素,以獲取正確的辨識用影像模型,例如人臉影像模型。由於本發明的方法並沒有影像必須乾淨的假設,獲得的影像不需要經過前處理即可用來建立辨識用影像模型。因此本發明不需捨棄一些照明條件不佳、遭到遮蔽或片段的影像。換句話說,本發明能保留所有影像,成為有用的資訊。與先前技術必須經過挑選而捨棄一些遮蔽的圖片,完全不同。The image automatic identification method and system disclosed by the present invention proposes a solution for training images that have poor lighting conditions, masking, and segment images when establishing an image model for identification. The method according to the present invention mainly comprises: accurately correcting the error of the pixel by low rank decomposition; and obtaining a correct identification image model, such as a face image model, by separating pixels that do not belong to a specific target. Since the method of the present invention does not have the assumption that the image must be clean, the obtained image can be used to establish the identification image model without pre-processing. Therefore, the present invention does not need to discard some images of poor lighting conditions, obscuration or fragments. In other words, the present invention retains all images and becomes useful information. It's completely different from the previous technology that has to be chosen to discard some of the masked images.

此外,本發明引入結構相異性的概念,以更正確的找出屬於特定標的的特徵,例如臉型、眼睛、鼻子的大小及形狀等。比起只用低秩分解方法,本發明可以進一步提高辨識率。In addition, the present invention introduces the concept of structural dissimilarity to more accurately identify features that belong to a particular subject, such as face, eyes, size and shape of the nose, and the like. The present invention can further improve the recognition rate as compared with the low rank decomposition method alone.

本發明提出一種新穎的影像辨識用模型建立方法,特別是人臉影像辨識用模型的建立方法,包括如下步驟:輸入多數個分屬於n類別的數位化影像檔,各影像檔包括代表一影像之像素位址及其灰階資料,及該影像所屬類別資訊;對各類別之多數影像檔Di(0<in),抽取其共同特徵Ai;計算一類別共同特徵Ai與特定數量(m)其他類別共同特徵Aj(0<jmn-1,Dj≠Di)之差異特徵Σ(AjT Ai);及將該差異特徵Σ(AjT Ai)加入該類別之共同特徵Ai,得到該類別之模型影像AiThe invention provides a novel image recognition model establishing method, in particular, a method for establishing a face image recognition model, which comprises the following steps: inputting a plurality of digital image files belonging to the n category, each image file comprising a representative image. Pixel address and its grayscale data, and the category information of the image; most image files for each category Di(0<i n) extracting its common feature Ai; calculating a common feature Ai of a class and a common feature of a certain number (m) of other categories Aj (0<j m The difference characteristic n(Aj T Ai) of n-1, Dj≠Di); and the difference feature Σ(Aj T Ai) is added to the common feature Ai of the category to obtain a model image Ai * of the category.

其中,該Ai 的計算,可以最佳化解法(optimum solution)獲得。在本發明 的較佳實例中,是以下式經由單一步驟計算,得到該Ai Among them, the calculation of the Ai * can be obtained by an optimization solution. In a preferred embodiment of the invention, the following formula is calculated via a single step to obtain the Ai * :

s.t.D i =A i +E i .St D i = A i + E i .

在本發明的一種較佳實例,中該特定數量m即為n-1,即將該類別共同特徵Ai與其他所有已存在類別的共同特徵,均計算其差異特徵。In a preferred embodiment of the present invention, the specific number m is n-1, that is, the common feature of the common feature Ai of the class and all other existing categories are calculated for the difference feature.

在本發明的實施例中,另包括以相同方法得到各類別之模型影像A1 ~An 的步驟。In the embodiment of the present invention, the method of obtaining the model images A1 * to An * of each category in the same manner is further included.

本發明也提出一種利用同類別訓練影像的錯誤影像,提高模型影像獨立性的方法。該錯誤影像可加權後與模型影像結合。The invention also proposes a method for improving the independence of model images by using the same type of training image error images. The erroneous image can be weighted and combined with the model image.

本發明也揭示利用所得到的各類別模型影像,判斷一輸入影像與各類別模型影像之近似度的方法。該方法包括以下步驟:輸入一待處理數位化影像檔,該影像檔包括代表一影像之像素位址及其灰階資料;比較該影像與以上述方式獲得的,分屬多數類別之模型影像之差異,計算該待處理影像與各類別模型影像之近似度;及判斷該待處理影像與各類別模型影像之關係。The present invention also discloses a method for determining the degree of approximation of an input image and each type of model image using the obtained model image of each category. The method comprises the steps of: inputting a digital image file to be processed, the image file comprising a pixel address representing an image and gray scale data; comparing the image with the model image obtained by the above method and belonging to most categories Difference, calculating the approximate degree of the image to be processed and the image of each category; and determining the relationship between the image to be processed and the image of each category.

其中,該判斷該待處理影像與各類別模型影像關係之步驟包括判斷該待處理影像屬於近似度最高的類別。在本發明其他實例中,該判斷該待處理影像與各類別模型影像關係之步驟包括判斷該待處理影像屬於近似度高於一臨界值的類別。在本發明的多數實例中,該輸入之多數數位化影像檔為人臉影像檔,而以該人臉影像來源之人區分其類別。在此種實例中,所建立之模型影像為屬 於多數人之臉部模型影像。但在本發明其他實例中,所建立之模型影像為代表一物品、一場所或一事件之影像。The step of determining the relationship between the image to be processed and each model image includes determining that the image to be processed belongs to the category with the highest degree of similarity. In another example of the present invention, the step of determining the relationship between the image to be processed and each model image includes determining that the image to be processed belongs to a category whose approximation is higher than a threshold. In most instances of the invention, the majority of the digitized image files of the input are facial image files, and the categories from which the face image source is derived. In such an example, the model image created is a genus Image of the face model of most people. However, in other examples of the invention, the model image created is an image representing an item, a venue, or an event.

以下將以實例說明本發明之影像自動辨認方法,包括模型影像之建立方法以及利用所建立的模型影像辨認待處理影像的方法。唯須說明,本發明所舉之實例,目的在例示本發明的方法步驟及其適用之軟硬體元件及架構,使此行業之專家能夠據以實施。發明詳細說明的目的並非用來展示本發明的全貌或關鍵性元件,更不得用來限制本發明的範圍。Hereinafter, an image automatic recognition method of the present invention will be described by way of example, including a method for establishing a model image and a method for identifying a to-be-processed image by using the created model image. It should be noted that the examples of the present invention are directed to exemplifying the method steps of the present invention and the applicable hardware and software components and architectures thereof, enabling experts in the industry to implement them. The detailed description of the present invention is not intended to be exhaustive or to limit the scope of the invention.

本發明提出一種新穎的影像自動辨認技術,特別是人臉影像的自動辨認技術。本發明提出一種利用模型影像來辨認待處理影像的方法,可以獲得高度的正確率。The invention provides a novel image automatic recognition technology, in particular, an automatic recognition technology for face images. The invention proposes a method for identifying a to-be-processed image by using a model image, and a high accuracy rate can be obtained.

第1圖表示本發明模型影像建立方法流程圖。以下將以建立人臉模型影像為例,說明該模型影像的建立方法。Fig. 1 is a flow chart showing the method of creating a model image of the present invention. The following is an example of establishing a face model image to illustrate how to create the model image.

如第1圖所示,在建立模型影像時,首先在步驟101先取得多數個分屬於n(n為大於1的自然數)類別的數位化影像檔。在本實例中,因為目的是在作人臉影像比對或人臉模型影像建立,因此,在此步驟所取得的影像檔,是代表已知多數人的人臉影像;所稱之n類別是指n個已知人物。在此步驟所得到的影像檔,通常是包括代表一影像之像素位址及其灰階資料,及該影像所屬類別資訊。換言之,在人臉影像的情形,所得到的影像檔包括代表特定人之臉部影像之像素位址及其灰階資料,以及該特定人之名稱或代碼。在人臉辨認或其他應用場合,可能也許要包括代表像素色彩及其灰階之資料。其他適合使用在影像辨認或比對的資料,也可包括在該影像檔中。As shown in FIG. 1, when the model image is created, first, in step 101, a plurality of digitized image files belonging to the category n (n is a natural number greater than 1) are obtained. In this example, since the purpose is to create a face image comparison or a face model image, the image file obtained at this step is a face image representing a known majority; the so-called n category is Refers to n known characters. The image file obtained in this step usually includes a pixel address representing the image and its grayscale data, and the category information of the image. In other words, in the case of a face image, the obtained image file includes a pixel address representing a face image of a specific person and its grayscale material, and the name or code of the specific person. In face recognition or other applications, it may be necessary to include data representing the color of the pixel and its grayscale. Other materials suitable for image recognition or comparison may also be included in the image file.

在本發明的應用中,該影像檔通常是一以數位影像掃描器或數位攝影機取得的數位化影像本身。但是在本發明其他實例中,該數位化影像可已經過壓縮及/或抽取特徵。以減少影像檔的資料量或降低影像特徵的維度。In the application of the present invention, the image file is usually a digital image itself obtained by a digital image scanner or a digital camera. However, in other examples of the invention, the digitized image may have been compressed and/or extracted. To reduce the amount of data in the image file or to reduce the dimensions of the image features.

習知技術在作人臉影像比對時,有一種稱為「低秩矩陣分解法」(Low-rank matrix recovery)的技術用來找出同一類中的相同資訊。其方法是從原始影像資料D中移除稀疏雜訊,以產生低秩矩陣A,用來代表該原始影像。所產生的低秩矩陣A可保有原始影像的重要特徵。該方法可以簡化處理並提高處理速度,適合用來作各種影像辨認應用。該方法可以用在本發明,作為簡化處理的工具。因此,在步驟101所取得的影像檔,可能是已經壓縮及/或抽取特徵的影像。如果是影像的原始檔,在該步驟後也可增加影像的壓縮及/或抽取特徵的步驟。在應用上,各影像檔較好調整成具有相同會呈倍數的解析度及尺寸,以及相同方位的影像。在已知的影像處理技術中,已經有許多可用的影像處理軟體,可供以手動或自動方式達成該目的。相關技術在此不需贅述。Conventional techniques, when used for face image comparison, have a technique called "Low-rank matrix recovery" to find the same information in the same class. The method is to remove the sparse noise from the original image data D to generate a low rank matrix A for representing the original image. The resulting low rank matrix A preserves the important features of the original image. This method can simplify processing and increase processing speed, and is suitable for various image recognition applications. This method can be used in the present invention as a tool for simplifying processing. Therefore, the image file obtained in step 101 may be an image that has been compressed and/or extracted. If it is the original file of the image, the step of compressing and/or extracting the image may also be added after this step. In application, each image file is preferably adjusted to have the same resolution and size that is multiples, and images of the same orientation. Among the known image processing techniques, there are many image processing software available that can be used to achieve this purpose either manually or automatically. The related art need not be described here.

不過,在作人臉影像辨識或比對時,對不同人物取像所得的人臉影像通常會含有共同的(具相關性的)特徵,例如眼睛、鼻子的所在位置及區域。對「人臉」而言,這些影像位置或區域通常近似。因此,所產生的矩陣A可能無法涵有足夠的區別資訊。這種問題在其他類似影像辨認或比對應用上,也常發生。為解決這項問題,本發明提供一種提高各影像間結構上差異性的方法,以強調各個影像檔的獨立性,作為影像辨認的根據。However, when making facial image recognition or comparison, facial images obtained by capturing different characters usually contain common (related) features such as the location and area of the eyes and nose. For "faces", these image locations or areas are usually similar. Therefore, the generated matrix A may not contain enough distinguishing information. This problem also occurs in other similar image recognition or comparison applications. In order to solve this problem, the present invention provides a method for improving the structural difference between images to emphasize the independence of each image file as a basis for image recognition.

在步驟102,系統對所取得的影像檔,按其所屬類別Di(0<in)抽取各類別所屬影像的共同特徵Ai。在抽取共同特徵時,可對影像檔中座標相同(解析度及尺寸調整後)的像素,灰階值接近的像素,認為是共同特徵點。其他已知 的影像特徵抽取技術,也可應用在本步驟。如果該影像檔是前述低秩矩陣分析所得到的低秩矩陣,則可直接抽取其共同特徵,得到各類別的共同特徵檔Ai。此時所得到的共同特徵檔也是一低秩矩陣。In step 102, the system selects the obtained image file according to its category Di(0<i n) Extract the common feature Ai of the image to which each category belongs. When the common feature is extracted, the pixels with the same coordinates (resolution and size adjustment) in the image file, and the pixels whose gray scale values are close to each other are considered to be common feature points. Other known image feature extraction techniques can also be applied in this step. If the image file is the low rank matrix obtained by the low rank matrix analysis described above, the common features can be directly extracted to obtain the common feature file Ai of each category. The common signature obtained at this time is also a low rank matrix.

第2圖即表示將一組屬於同一人的臉部影像抽取共同特徵的結果。圖中,上方為該組原始影像,下方為所得到的共同特徵影像。當然,第2圖僅是用來表示本發明的一種實例,並非用來限制本發明。Figure 2 shows the result of extracting a common feature from a group of facial images belonging to the same person. In the figure, the upper part is the original image of the group, and the lower part is the obtained common feature image. Of course, Figure 2 is only intended to illustrate an example of the invention and is not intended to limit the invention.

於步驟103計算一類別共同特徵Ai與特定數量(m,m為自然數)其他類別共同特徵Aj(0<jmn-1,Dj≠Di)之差異特徵Σ(AjT Ai)。雖然在本發明中,利用少數其他類別的共同特徵即可建立有用的模型影像檔,但在本發明的較佳實例中,可將該類共同特徵影像Ai,與其他n-1類別共同特徵影像逐一比較,找出其與所有其他類別影像之差異影像。如此可以提高辨識的正確率。於步驟104同時考慮該差異特徵Σ(AjT Ai)以及共同特徵Ai的情況下,以最佳化解法得到該類別之特徵影像Ai 。如此完成一類別影像的共同特徵檔的建立。所建立的共同特徵檔即可作為該類別的模型影像使用。所完成的共同特徵檔具有較高的獨立性,不但可以代表該類別的共同特徵,且提供較多的特徵,可以提高辨識的正確率。在人臉的辨識應用上,該共同特徵檔即代表特定人的臉部特徵,足以用來辨認一張待處理人臉影像是否屬於該特定人。In step 103, a common feature Ai and a specific number (m, m is a natural number) are used to calculate the common feature Aj (0<j). m The difference characteristic of n-1, Dj≠Di) (Aj T Ai). Although in the present invention, a useful model image file can be created using a common feature of a few other categories, in the preferred embodiment of the present invention, the common feature image Ai can be combined with other n-1 categories. Compare them one by one to find out the difference between them and all other categories of images. This can improve the accuracy of recognition. In the case where the difference feature Σ (Aj T Ai) and the common feature Ai are simultaneously considered in step 104, the feature image Ai * of the category is obtained by an optimization solution. This completes the creation of a common profile for a category of images. The created common profile can be used as a model image for this category. The completed common feature file has higher independence, can not only represent the common features of the category, but also provide more features, which can improve the correct rate of recognition. In the face recognition application, the common feature file represents a facial feature of a specific person, which is sufficient for identifying whether a pending face image belongs to the specific person.

在本發明的較佳實例中,可以利用下式,僅經由單一步驟計算得到該模型影像: In a preferred embodiment of the present invention, the model image can be calculated by a single step using only the following formula:

s.t.D i =A i +E i .St D i = A i + E i .

於步驟105,判斷是否所有n類別的影像檔都已經建立共同特徵檔。如否,則會到步驟103。如是,則在步驟106結束操作,完成所有類別的影像共同特徵檔的建立。利用上述方式所建立的影像模型檔,具有獨立性提高的類別共同特徵,在進行影像比對或辨認時,可提高正確率。At step 105, it is determined whether all of the n types of image files have established a common profile. If no, it will go to step 103. If so, the operation ends in step 106, and the creation of the image common feature files of all categories is completed. The image model file established by the above method has the common feature of the category with improved independence, and the correctness rate can be improved when the image is compared or recognized.

在利用本發明所建立的模型影像進行比對時,是將一張待比對影像與已經存檔的模型影像作比對,根據其間差異值的高低,判斷該帶比對影像應屬的類別。以人臉影像比對而言,就是該待比對影像屬於何人的臉部。When the model image created by the present invention is used for comparison, a pair of to-be-compared images is compared with the model image that has been archived, and according to the difference value between the two, the category of the pair of images to be compared is determined. In terms of face image comparison, it is the face of the person to whom the image belongs.

第3圖顯示一種利用本發明所建立的模型影像進行比對的比對方法流程圖。如第3圖所示,在步驟301輸入一待處理數位化影像檔。該影像檔通常包括一影像之像素位址及其灰階資料。於步驟302將該影像檔作必要的處理,使其具有特定之解析度,尺寸及方位。如有必要,可在該影像中定出定位點,以提高比對的速度與正確率。於步驟303取得代表多數類別影像的多數模型影像檔,各模型影像檔也包括一影像之像素位址及其灰階資料,及該影像所屬類別資訊。如果模型影像的內容並非影像之像素位址及其灰階資料,而是其他類型、格式資料,則在步驟304將該待比對影像處理成相同或相容格式。於步驟305將該待處理影像檔與所有類別之特徵影像比對,分別得到其差異資料。於步驟306選出小於一臨界值之差異值。於307判斷該待處理數位化影像屬於選出之差異值中,最小差異值所對應之類別。如此即完成比對。Figure 3 shows a flow chart of a comparison method for comparing images using the model images created by the present invention. As shown in FIG. 3, a digital image file to be processed is input in step 301. The image file usually includes a pixel address of an image and its grayscale data. In step 302, the image file is subjected to necessary processing to have a specific resolution, size, and orientation. If necessary, locate the anchor point in the image to increase the speed and accuracy of the alignment. In step 303, a plurality of model image files representing the majority of the images are obtained. Each model image file also includes a pixel address of the image and its grayscale data, and the category information of the image. If the content of the model image is not the pixel address of the image and its grayscale data, but other types, format data, then in step 304 the image to be compared is processed into the same or compatible format. In step 305, the image file to be processed is compared with the feature images of all categories to obtain the difference data. At step 306, a difference value less than a threshold is selected. It is determined at 307 that the digital image to be processed belongs to the selected difference value and the category corresponding to the minimum difference value. This completes the comparison.

步驟306與307可以避免所選出的類別差異值過大,而造成誤判。但事實上也可以在算出差異值後,選定差異值小於特定臨界值(及近似度大於臨界值)的類別,判斷該待處理影像屬於該等類別。此外,當然也可僅挑選出差異值最小(近似度最高)的類別,判斷該待處理影像屬於該類別。以人臉辨認應用為 例,判斷結果即為該待處理影像屬於何人之臉部影像。Steps 306 and 307 can prevent the selected category difference value from being too large, resulting in a false positive. In fact, after calculating the difference value, the category whose difference value is smaller than the specific threshold value (and the approximation degree is greater than the threshold value) may be selected to determine that the image to be processed belongs to the categories. In addition, it is of course possible to select only the category having the smallest difference value (the highest degree of approximation) and determine that the image to be processed belongs to the category. Face recognition application For example, the judgment result is the face image of the person to whom the image to be processed belongs.

為證明本發明的效果,以現行最具公信力之人臉資料庫AR Database(http:///www2.ece.ohio-state.edu/~aleix/ARdatabase.html ,Ohio State University,USA)中選取50名男人及50名女人的臉部影像,以表一所示的組合進行模形影像檔的建立及辨識。每組訓練用影像檔均為800張,以1,200張選自其中的影像進行辨認測試。In order to prove the effect of the present invention, the current most credible face database AR Database ( http:///www2.ece.ohio-state.edu/~aleix/ARdatabase.html , Ohio State University, USA) is selected. The facial images of 50 men and 50 women were created and identified by the combination shown in Table 1. Each group of training images has 800 images, and 1,200 images selected from them are used for identification test.

建立模型影像時,使用本發明方法(Ours)、低秩矩陣解析法(LR)、SRC、LLC+SRC、NN、Fisherface方法等,作為比對依據。實驗結果如第4、5、6圖所示。證明本發明的正確率優於習知方法,且本發明在維度為75時,即可達到相當高的正確率。足以簡化模型影像的建立與待比對影像的比對,並提高影像訓練及比對的速度。When constructing a model image, the method of the present invention (Ours), low rank matrix analysis (LR), SRC, LLC+SRC, NN, Fisherface method, etc. are used as the basis for comparison. The experimental results are shown in Figures 4, 5, and 6. It is proved that the correct rate of the present invention is superior to the conventional method, and the present invention achieves a relatively high correct rate when the dimension is 75. It is enough to simplify the comparison between the establishment of the model image and the comparison of the images, and improve the speed of image training and comparison.

以上說明目的是以實例揭露本發明之原理、操作、功能及效果。說明書的 內容只是用來例示本發明之若干較佳實例。習於斯藝之人士不能經由閱讀本件專利說明書,了解本發明之概念與精神,並在不脫出本發明範圍之下,完成各種不同之變化與引申。但諸如此類之應用,都屬於本發明專利範圍。The above description is intended to disclose the principles, operations, functions and effects of the invention. Instructions The content is only intended to illustrate several preferred examples of the invention. The person skilled in the art can not understand the concept and spirit of the present invention by reading this patent specification, and various changes and extensions are made without departing from the scope of the invention. However, applications such as these are within the scope of the invention.

第1圖表示本發明模型影像建立方法流程圖。Fig. 1 is a flow chart showing the method of creating a model image of the present invention.

第2圖表示將一組屬於同一人的臉部影像抽取共同特徵的結果。Figure 2 shows the result of extracting a common feature from a group of facial images belonging to the same person.

第3圖顯示一種利用本發明所建立的模型影像進行比對的比對方法流程圖。Figure 3 shows a flow chart of a comparison method for comparing images using the model images created by the present invention.

第4圖表示本發明與已知方法第一種比較實驗結果。Figure 4 shows the results of the first comparative experiment of the present invention and known methods.

第5圖表示本發明與已知方法第二種比較實驗結果。Figure 5 shows the results of a second comparative experiment of the present invention and a known method.

第6圖表示本發明與已知方法第三種比較實驗結果。Figure 6 shows the results of a third comparative experiment of the present invention and known methods.

Claims (9)

一種影像辨識用模型建立方法,包括如下步驟:輸入多數個分屬於n類別的數位化影像檔,各影像檔包括代表一影像之像素位址及其灰階資料,及該影像所屬類別資訊;對各類別之多數影像檔Di(0<in),抽取其共同特徵Ai;計算一類別共同特徵Ai與特定數量(m)其他類別共同特徵Aj(0<jmn-1,Dj≠Di)之差異特徵Σ(AjT Ai);及將該差異特徵Σ(AjT Ai)加入該類別之共同特徵Ai,得到該類別之模型影像Ai*;其中,該Ai*是以下式經由單一步驟計算得到: s.t.D i =A i +E i .A method for establishing a model for image recognition includes the following steps: inputting a plurality of digitized image files belonging to the n category, each image file including a pixel address representing the image and its grayscale data, and category information of the image; Most image files of each category Di(0<i n) extracting its common feature Ai; calculating a common feature Ai of a class and a common feature of a certain number (m) of other categories Aj (0<j m a difference characteristic n(Aj T Ai) of n-1, Dj≠Di); and adding the difference characteristic Σ(Aj T Ai) to the common feature Ai of the category to obtain a model image Ai* of the category; wherein the Ai * is calculated by a single step: stD i =A i +E i . 如申請專利範圍第1項之方法,其中之m為n-1。 The method of claim 1, wherein m is n-1. 如申請專利範圍第1或2項之方法,另包括以相同方法得到各類別之模型影像A1*~An*的步驟。 For example, the method of claim 1 or 2 of the patent scope includes the steps of obtaining the model images A1* to An* of each category in the same manner. 如申請專利範圍第1或2項之方法,其中,所輸入之多數個分屬於n類別的數位化影像檔為分屬n個人的臉部數位化影像檔。 The method of claim 1 or 2, wherein the digitized image file in which the majority of the input points belong to the n category is a face digital image file belonging to n individuals. 如申請專利範圍第3項中之方法,其中,所輸入之多數個分屬於n類別的數位化影像檔為分屬n個人的臉部數位化影像檔。 The method of claim 3, wherein the digitized image file in which the majority of the input points belong to the n category is a face digital image file belonging to n individuals. 一種判斷一輸入影像與多數類別模型影像之關係的方法,包括以下步驟:輸入一待處理數位化影像檔,該影像檔包括代表一影像之像素位址及其灰階資料; 取得以申請專利範圍第1或2項之方法所得之多數類別模型影像;比較該待處理影像與該之多數類別模型影像之差異,以計算該待處理影像與各類別模型影像之近似度;及判斷該待處理影像為屬於所得近似度最高之模型影像之類別。 A method for determining a relationship between an input image and a majority of model images includes the following steps: inputting a digital image file to be processed, the image file including a pixel address representing an image and grayscale data thereof; Obtaining a majority of the model image obtained by the method of claim 1 or 2; comparing the difference between the image to be processed and the image of the majority of the model to calculate the similarity between the image to be processed and the model image of each category; The image to be processed is determined to belong to the category of the model image with the highest degree of similarity. 一種判斷一輸入影像與多數類別模型影像之關係的方法,包括以下步驟:輸入一待處理數位化影像檔,該影像檔包括代表一影像之像素位址及其灰階資料;取得以申請專利範圍第3項方法所得之多數類別模型影像;比較該待處理影像與該之多數類別模型影像之差異,以計算該待處理影像與各類別模型影像之近似度;及判斷該待處理影像為屬於所得近似度最高之模型影像之類別。 A method for determining a relationship between an input image and a majority of model images includes the steps of: inputting a digital image file to be processed, the image file including a pixel address representing an image and grayscale data; obtaining a patent application scope The majority of the model image obtained by the third method; comparing the difference between the image to be processed and the image of the majority of the model to calculate the similarity between the image to be processed and the model image of each category; and determining that the image to be processed belongs to the The category of the most approximate model image. 一種判斷一輸入影像與多數類別模型影像之關係的方法,包括以下步驟:輸入一待處理數位化影像檔,該影像檔包括代表一影像之像素位址及其灰階資料;取得以申請專利範圍第1或2項之方法所得之多數類別模型影像;比較該待處理影像與該之多數類別模型影像之差異,以計算該待處理影像與各類別模型影像之近似度;及判斷該待處理影像為屬於所得近似度超出臨界值之模型影像之類別。 A method for determining a relationship between an input image and a majority of model images includes the steps of: inputting a digital image file to be processed, the image file including a pixel address representing an image and grayscale data; obtaining a patent application scope The majority model image obtained by the method of the first or second method; comparing the difference between the image to be processed and the image of the majority of the model to calculate the similarity between the image to be processed and the model image of each category; and determining the image to be processed It is a category of model images whose approximate approximation exceeds the critical value. 一種判斷一輸入影像與多數類別模型影像之關係的方法,包括以下步驟:輸入一待處理數位化影像檔,該影像檔包括代表一影像之像素位址及其灰階資料;取得以申請專利範圍第3項方法所得之多數類別模型影像; 比較該待處理影像與該之多數類別模型影像之差異,以計算該待處理影像與各類別模型影像之近似度;及判斷該待處理影像為屬於所得近似度超出臨界值之模型影像之類別。A method for determining a relationship between an input image and a majority of model images includes the steps of: inputting a digital image file to be processed, the image file including a pixel address representing an image and grayscale data; obtaining a patent application scope Image of most categories of models obtained by the third method; Comparing the difference between the image to be processed and the image of the majority of the model to calculate the similarity between the image to be processed and the model image of each category; and determining that the image to be processed is a category of the model image whose approximate degree of approximation exceeds a critical value.
TW101124625A 2012-07-09 2012-07-09 Automatic image identification method TWI450206B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW101124625A TWI450206B (en) 2012-07-09 2012-07-09 Automatic image identification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW101124625A TWI450206B (en) 2012-07-09 2012-07-09 Automatic image identification method

Publications (2)

Publication Number Publication Date
TW201403496A TW201403496A (en) 2014-01-16
TWI450206B true TWI450206B (en) 2014-08-21

Family

ID=50345580

Family Applications (1)

Application Number Title Priority Date Filing Date
TW101124625A TWI450206B (en) 2012-07-09 2012-07-09 Automatic image identification method

Country Status (1)

Country Link
TW (1) TWI450206B (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120121142A1 (en) * 2009-06-09 2012-05-17 Pradeep Nagesh Ultra-low dimensional representation for face recognition under varying expressions

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120121142A1 (en) * 2009-06-09 2012-05-17 Pradeep Nagesh Ultra-low dimensional representation for face recognition under varying expressions

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Ramirez, Ignacio, Pablo Sprechmann, and Guillermo Sapiro. "Classification and clustering via dictionary learning with structured incoherence and shared features." Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on. IEEE, 2010. *

Also Published As

Publication number Publication date
TW201403496A (en) 2014-01-16

Similar Documents

Publication Publication Date Title
WO2018166116A1 (en) Car damage recognition method, electronic apparatus and computer-readable storage medium
CN109635686B (en) Two-stage pedestrian searching method combining human face and appearance
CN109145742B (en) Pedestrian identification method and system
WO2015149534A1 (en) Gabor binary pattern-based face recognition method and device
WO2021051545A1 (en) Behavior identification model-based fall-down action determining method and apparatus, computer device, and storage medium
CN110008909B (en) Real-name system business real-time auditing system based on AI
WO2019061658A1 (en) Method and device for positioning eyeglass, and storage medium
CN105426815A (en) Living body detection method and device
KR20170006355A (en) Method of motion vector and feature vector based fake face detection and apparatus for the same
CN103679147A (en) Method and device for identifying model of mobile phone
CN105678213B (en) Dual-mode mask person event automatic detection method based on video feature statistics
EP3905104B1 (en) Living body detection method and device
WO2019006835A1 (en) Target recognition method based on compressed sensing
WO2013075295A1 (en) Clothing identification method and system for low-resolution video
CN110457996A (en) Moving Objects in Video Sequences based on VGG-11 convolutional neural networks distorts evidence collecting method
CN112200056A (en) Face living body detection method and device, electronic equipment and storage medium
CN112633221A (en) Face direction detection method and related device
CN114550268A (en) Depth-forged video detection method utilizing space-time characteristics
CN111882525A (en) Image reproduction detection method based on LBP watermark characteristics and fine-grained identification
US20160110909A1 (en) Method and apparatus for creating texture map and method of creating database
Singh et al. Feature based method for human facial emotion detection using optical flow based analysis
US8879805B2 (en) Automated image identification method
CN111985400A (en) Face living body identification method, device, equipment and storage medium
TWI450206B (en) Automatic image identification method
Işikdoğan et al. Automatic recognition of Turkish fingerspelling

Legal Events

Date Code Title Description
MM4A Annulment or lapse of patent due to non-payment of fees