WO2011092848A1 - Object detection device and face detection device - Google Patents

Object detection device and face detection device Download PDF

Info

Publication number
WO2011092848A1
WO2011092848A1 PCT/JP2010/051257 JP2010051257W WO2011092848A1 WO 2011092848 A1 WO2011092848 A1 WO 2011092848A1 JP 2010051257 W JP2010051257 W JP 2010051257W WO 2011092848 A1 WO2011092848 A1 WO 2011092848A1
Authority
WO
WIPO (PCT)
Prior art keywords
face
image
unit
feature
feature amount
Prior art date
Application number
PCT/JP2010/051257
Other languages
French (fr)
Japanese (ja)
Inventor
智明 吉永
茂喜 長屋
武洋 藤田
Original Assignee
株式会社日立製作所
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社日立製作所 filed Critical 株式会社日立製作所
Priority to PCT/JP2010/051257 priority Critical patent/WO2011092848A1/en
Priority to JP2011551643A priority patent/JPWO2011092848A1/en
Publication of WO2011092848A1 publication Critical patent/WO2011092848A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2148Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the process organisation or structure, e.g. boosting cascade
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Definitions

  • the present invention relates to an object detection device that detects a specific object such as a face from an image.
  • Non-Patent Document 1 and Patent Document 1 as a method for detecting a specific object such as a face from an input image.
  • Non-Patent Document 1 detects a face in an image by determining whether a face image of 20 ⁇ 20 pixels is a face or a non-face using a plurality of image features called haar. is there.
  • Patent Document 1 detects a face in a plurality of directions by preparing a different feature pattern for each face direction and estimating the face direction, and then performing face detection processing specialized for the estimated face direction.
  • Non-Patent Document 1 when a change occurs such as turning to the side or rotating with respect to the face to be detected, the reaction deteriorates in many image features. Arise. That is, the detection rate for a face in which a specific change has occurred is low.
  • Patent Document 1 since the direction estimation is performed using the feature pattern for each of a plurality of directions, there is a problem that the processing time increases, and further, if the face direction estimation is wrong, the face cannot be detected. For this reason, it has been necessary to improve the detection rate for a face in which a specific change such as rotation has occurred without adding an enormous feature amount evaluation calculation.
  • the present invention has been made in view of the above problems, and an object of the present invention is to improve the detection rate for a face in which a specific change has occurred.
  • the object detection rate can be increased.
  • FIG. 1 is a block diagram illustrating a configuration of an object detection apparatus according to a first embodiment. It is explanatory drawing which shows the example of a process of the image input part 101 of FIG. It is a figure which shows the example of the feature pattern processed by the feature evaluation part 102 of FIG. It is a table which shows an example of the data of the feature pattern used by the object discrimination
  • FIG. 10 is a diagram illustrating an example of feature amounts used in the third embodiment.
  • FIG. 10 is a configuration diagram illustrating a configuration of an object detection apparatus according to a fourth embodiment.
  • FIG. 10 is a configuration diagram illustrating a configuration of an object detection apparatus according to a fifth embodiment.
  • FIG. 10 is a diagram illustrating an example of a setting screen for performing parameter setting in the fifth embodiment.
  • the object detection unit 100 includes an image input unit 101, a feature pattern DB 110, a feature evaluation unit 102, a feature amount storage unit 103, an object determination unit 104, a changed object determination unit 105, and a determination result output unit 106.
  • Each unit described above may be configured by hardware. Further, it may be a module combining hardware and software.
  • the operation of the object detection unit shown in FIG. 1 will be described using a case where a face is detected as an example.
  • a face As an object to be detected by the object detection unit, in addition to the face, other objects such as a person, a car, and a sign may be targeted.
  • an operation for detecting a front face of a person from an input image will be described as an example of an object detection operation.
  • the image input unit 101 receives an imaging module such as a camera, a reproduction image of a pre-recorded image, and the like, and outputs an image region 203 for performing a discrimination process between a face and a non-face to the feature evaluation unit 102.
  • the feature pattern DB 110 stores a feature pattern for determining whether it is a face or a non-face.
  • the feature evaluation unit 102 calculates feature amounts for a plurality of feature patterns defined in the feature pattern DB 110 for the input image region 203 and stores them in the feature amount storage unit 103.
  • the feature amount storage unit 103 stores the feature amount obtained from the feature evaluation unit 102.
  • the object discriminating unit 104 performs discrimination processing of a face or a non-face based on a plurality of feature value values obtained by the feature evaluation unit 102. As a result, if it is determined as a face, the result is output to the determination result output unit 106, and if it is determined as a non-face, the result is output to the change object determination unit 105.
  • the changed object discriminating unit 105 discriminates whether the image area 203 discriminated as a non-face is a face or a non-face that has undergone a specific change, such as a sideways face, using the feature amount stored in the feature amount storage unit 103.
  • the change object determination unit 105 in order to find an object different from the object determination unit 104, a determination process different from that of the object determination unit 104 is performed.
  • the discrimination result in the change object discrimination unit 105 is output to the discrimination result output unit 106.
  • the image region determined as the face by the object discrimination unit 104 or the image region discriminated as the face having undergone the specific change by the changed object discrimination unit 105 is set as a face, and the other is determined as a non-face. Is output.
  • the faces 201 and 202 to be detected exist at arbitrary positions and in arbitrary sizes.
  • the image input unit 101 cuts out image regions 203 having a plurality of positions and sizes on the input image 200, for example, in a raster scan form, and outputs them to the feature evaluation unit 102.
  • the feature evaluation unit 102 and later, a process for determining whether a face is a non-face is performed on a plurality of image areas in a single input image 200. As a result, a face of an arbitrary size present at an arbitrary location in the input image 200 is detected.
  • the feature pattern DB 110 in FIG. 1 will be described.
  • the feature pattern DB 110 defines a plurality of image feature parameters used for face discrimination.
  • FIG. 3 is an example of a defined image feature pattern.
  • the feature pattern in FIG. 3 is composed of a black rectangle 301 and a white rectangle 302, and the feature amount is obtained by the difference in the sum of the pixel values in the rectangle.
  • FIG. 4 is an example of a face discrimination parameter table 400 in which the feature parameters are defined.
  • N image features h i (i ⁇ N) are defined to discriminate between a face and a non-face, and a weight ⁇ when performing face determination for each feature and a weight ⁇ when performing right oblique side face determination , A weight ⁇ for the left oblique side face determination is defined.
  • the feature evaluation unit 102 calculates a plurality of feature amounts for the input image region 203. If the image area 203 input to the image feature pattern h i defined in the feature pattern DB 110 is a vector I, the obtained feature quantity h i (I) can be obtained by Equation 1.
  • the feature evaluation unit 102 stores the N feature amounts obtained by the above calculation in the feature amount storage unit 103.
  • EOH Edge of Orientation Histograms
  • HOG Histogram of Orientation Gradients
  • the object discriminating unit 104 calculates face likelihood based on each feature quantity h i (I) obtained by the feature evaluation unit 102 and discriminates whether the image region 203 is a face or a non-face.
  • the AdaBoost classifier calculates the face likelihood F (I) by a linear sum function using the face discrimination weight ⁇ i for each feature h i described in the face discrimination parameter table 400 of FIG. To do.
  • a larger face likelihood F (I) indicates that the face is more likely to be a face. If the face likelihood F is greater than or equal to the threshold Th F , it is determined to be a face, and if it is less, it is determined to be a non-face. If the discrimination result is a face, the object discrimination unit 104 outputs the result to the discrimination result output unit 106. If the determination result is a non-face, the result is output to the change object determination unit 105.
  • FIG. 5 is a flowchart showing steps of determining the right oblique face and the left oblique face in the change object discriminating unit 105.
  • step 501 the feature quantity h i (I) for the image area 203 stored in the feature quantity storage unit 103 is read.
  • the likelihood R for the right oblique face is calculated from the feature amount.
  • the likelihood R is calculated in step 502 using the same feature amount as that of the object determination unit 104, but a different determination process is performed because a determination target is different.
  • the right oblique face likelihood R is calculated by the calculation of Equation 3 which is a function of a linear sum using the right oblique face weight ⁇ i for each feature h i described in the face discrimination parameter table 400 of FIG.
  • step 503 the right oblique face likelihood R is compared with the right oblique face discrimination threshold Thre R, and if it is equal to or greater than the threshold, the process proceeds to step 507, and if not, the process proceeds to step 504.
  • step 504 the likelihood L for the left oblique face is calculated as in the case of the right oblique face.
  • the left oblique face likelihood L is a mathematical expression that is a function of a linear sum using the left oblique face weight ⁇ i for each feature h i described in the face discrimination parameter table 400 of FIG. 4 is calculated.
  • step 505 the left oblique face likelihood L is compared with the left oblique face discrimination threshold Thre L, and if it is equal to or greater than the threshold, the process proceeds to step 507, and if not, the process proceeds to step 506.
  • FIG. 6 is a diagram showing the concept of face likelihood calculation in steps 502 and 504 of FIG.
  • FIG. 6A is a diagram showing calculation of face likelihood for a front face
  • FIGS. 6B and 6C show calculation of right diagonal face likelihood and left diagonal face likelihood, respectively.
  • FIG. 6 the weights for the feature amounts 601 to 605 are expressed by shading.
  • FIG. 6A almost equal weights are assigned to the feature amounts 601 to 605 in order to discriminate a general front face.
  • FIG. 6B shows that weights are small for feature amounts such as h1, h3, and h4, and features such as h2, h5, and the like are large. Yes.
  • Step 506 is a step visited when it is determined that neither the right diagonal face nor the left diagonal face is detected. For this reason, in step 506, the final discrimination result in the change object discrimination unit is determined as a non-face.
  • step 507 is visited when it is determined that the face is a right diagonal face or a left diagonal face. For this reason, step 507 determines the discrimination result as a face.
  • the face detection apparatus detects a face from an input image, and includes an image input unit that inputs an image, and a feature that calculates a feature amount for the image obtained by the image input unit.
  • a linear sum is calculated by the evaluation unit, the feature amount storage unit that stores the calculated feature amount, the feature amount obtained by the feature evaluation unit and the front face function, and based on the calculated linear sum,
  • An object discriminating unit that discriminates whether the image is a front face or a non-front face; and an image that has been discriminated as a non-front face by the object discriminating unit; Calculates a linear sum with an oblique face function with a different weighting factor, and determines whether the image is a non-face or an oblique face, and a discrimination result that outputs the result of the object discrimination Characterized by having an output part
  • the feature points to which the weighting coefficient of the function for partial occlusion face is included in the change object discriminating unit include part of the feature points of the front face, the feature amount is not recalculated and the calculation amount is reduced. Therefore, high-speed face detection can be realized.
  • Example 2 will be described with reference to FIG.
  • the present embodiment is an example in which face detection is performed at high speed by using a cascade type discriminator as shown in FIG.
  • the face / non-face discriminator in the object discriminating unit 104 has a configuration in which a plurality of discriminators 710 to 730 are connected in cascade.
  • the discriminator 1 (710) uses only the feature value set (701) from the feature value 1 to the feature value A (A ⁇ N) out of the N feature values stored in the feature value storage unit 103. Processing is performed to determine whether the image area is a face or a non-face. Here, when it is determined that the face is non-face (when the determination result of the discriminator 1 (710) is “False”), the processing ends. At this time, the feature amount calculation in the feature evaluation unit 102 is completed only after the feature amount A is processed.
  • the determination processing is left to the identifier 2 (720).
  • the discriminator 2 (720) performs discrimination processing using the feature amount set (702) from the feature amounts A + 1 to B, and similarly determines whether the image area is a face or a non-face.
  • the discrimination result is set as a face.
  • the discriminator S (730) determines False
  • the change object discriminating unit 105 determines whether the face is an oblique face or a non-face.
  • the changing object discriminating unit 105 uses the N feature amounts stored in the feature amount storage unit (103) used in the discriminating processes of the discriminators 1 to S to determine whether the left and right diagonal faces are not in accordance with the flowchart of FIG. Determine if it is a face.
  • FIG. 8A shows an example of images of feature amounts (801) to (803) in which a high weight value is set when the face of a person wearing sunglasses is detected in the change object discriminating unit (105).
  • FIG. FIG. 8B is a diagram showing an example of images of feature amounts (804) to (807) in which a high weight value is set when detecting the face of a person with a mask similarly.
  • a classifier for discriminating each change object is prepared in the feature pattern DB 110 in advance. It is possible to determine whether the face is equipped with these wearing items.
  • the third embodiment is a face detection device that detects a face from an input image, and includes an image input unit that inputs an image, and a feature that calculates a feature amount for the image obtained by the image input unit.
  • a linear sum is calculated by the evaluation unit, the feature amount storage unit that stores the calculated feature amount, the feature amount obtained by the feature evaluation unit and the front face function, and based on the calculated linear sum,
  • An object discriminating unit that discriminates whether the image is a front face or a non-front face; and an image that has been discriminated as a non-front face by the object discriminating unit; Calculates a linear sum with a partially occluded face function with a different weighting coefficient, and a change object discriminating unit that detects whether the image is a non-face or a face that is partially occluded by something.
  • a face detection device characterized by having That. With this face detection device, it is possible to determine a face with an obstruction, so that the face detection rate is improved.
  • the feature points to which the weighting coefficient of the function for partial occlusion face is included in the change object discriminating unit include part of the feature points of the front face, the feature amount is not recalculated and the calculation amount is reduced. Therefore, high-speed face detection can be realized.
  • the configuration of the apparatus according to the fourth embodiment will be described with reference to FIG.
  • the present embodiment is an example in which the object detection apparatus according to the first to third embodiments is mounted on an imaging apparatus such as a monitoring camera or a digital camera, a display, or a video recording apparatus.
  • FIG. 9 is a configuration diagram illustrating the configuration of the object detection device according to the fourth embodiment.
  • an object detection apparatus (900) includes an image input unit (909), an image memory (902), a CPU (903), a RAM (904), a ROM (905), a detection result recording unit (906), an interface ( 907) and an output device (908).
  • a target object is detected from an image obtained by a camera that is an imaging unit (901).
  • the CPU (903) in the object detection apparatus (900) of the present embodiment corresponds to the object detection unit (100) shown in FIG. 1 of the first embodiment, and executes each arithmetic processing of the object detection unit (100) as a program. This is realized by performing arithmetic processing by the CPU (903).
  • the CPU (903) performs arithmetic processing according to the detection method in the object detection unit (100) to detect the object.
  • the object detection result for each sequence is recorded in the detection result recording unit (906).
  • the detection result is converted into an appropriate form through the interface (907) and output to the output device (908).
  • the output device may be a display, a printer, a PC, or the like.
  • the object detection apparatus (900) according to the fourth embodiment is provided with the input device (1010) and the setting control unit (1020) shown in FIG. It is an example which comprises the object detection apparatus (1000) which can do.
  • a command for parameter adjustment in the change object determination process in the change object determination unit (105) of FIG. 1 is received from the input device (1010).
  • the setting control unit (1020) that has received this command performs parameter control such as ON / OFF control of discrimination processing for the deformed object in the changed object discrimination unit (105) and sensitivity adjustment.
  • parameter control such as ON / OFF control of discrimination processing for the deformed object in the changed object discrimination unit (105) and sensitivity adjustment.
  • FIG. 11 is a diagram showing an example of a parameter setting screen (1100) for performing parameter setting in deformed object discrimination with the input device (1010).
  • the parameter setting screen (1100) includes an object discrimination parameter (1101) and deformed object discrimination parameters (1102 to 1107).
  • the object discrimination parameter (1101) controls the sensitivity of face detection in the object discrimination unit (104). For example, when high sensitivity is set, it is easy to determine a face by loosening the threshold TF of the face discrimination parameter table (400) defined in the feature pattern DB (110).
  • the deformed object discriminating parameters (1102 to 1107) whether or not the changing object discriminating unit (105) performs discriminating processing on each deformed object such as a left-right slanted face and a left-right rotated face, and how sensitivity adjustment is performed.
  • the parameter information of each item set in this way is sent to the setting control unit (1020), and the setting control unit controls the detection process in the CPU.
  • Imaging unit 902 ... Image memory 903 ... CPU 904 ... RAM 905 ... ROM 906 ... Detection result recording unit 907 ... Interface 908 ... Output device 909 ... Image input unit 1000 ... Object detection device 1010 ... Input device 1020 ... Setting control unit 1100 ... Parameter setting screen 1101 ... Object discrimination parameters 1102 to 1107 ... Deformed object discrimination parameters

Abstract

The disclosed object detection device resolves the issue of deterioration of detection accuracy upon occurrence of deformation to an object that had occurred with conventional object detection devices by way of the following means. In an object detection device for determining whether an object or a non-object is represented by performing determination processing for feature values obtained from a plurality of feature patterns, for an image that has been determined not to represent an object, determination is performed by using a new classifier into which is input a feature value for which calculation has already been performed regarding whether the image represents a certain specific deformed object or a non-object. As a result, it is possible to detect with high precision even an object for which a specific deformation has been applied. In addition, simply by altering the classifier, high-speed detection becomes possible.

Description

オブジェクト検出装置および顔検出装置Object detection device and face detection device
 本発明は、画像中から顔などの特定オブジェクトを検出するオブジェクト検出装置に関する。 The present invention relates to an object detection device that detects a specific object such as a face from an image.
 入力画像中から顔などの特定オブジェクトを検出する方法として、例えば、非特許文献1や特許文献1がある。 There are, for example, Non-Patent Document 1 and Patent Document 1 as a method for detecting a specific object such as a face from an input image.
 非特許文献1に記載した技術は、例えば20×20画素の顔の画像をハールと呼ばれる複数の画像特徴を用いて顔か非顔か判別することで、画像中にある顔を検出するものである。 The technique described in Non-Patent Document 1 detects a face in an image by determining whether a face image of 20 × 20 pixels is a face or a non-face using a plurality of image features called haar. is there.
 特許文献1は、顔方向毎に異なる特徴パターンを用意して顔方向を推定したのちに、推定した顔方向に特化した顔検出処理を行うことで複数方向の顔を検出するものである。 Patent Document 1 detects a face in a plurality of directions by preparing a different feature pattern for each face direction and estimating the face direction, and then performing face detection processing specialized for the estimated face direction.
特開2008-173628号公報JP 2008-173628 A
 非特許文献1では、検出対象である顔に対して横を向く、回転するなどの変化が生じると、多くの画像特徴において反応が劣化することとなり、その結果、顔と判別されなくなり検出漏れが生じる。つまり、特定変化が生じた顔に対する検出率が低かった。 In Non-Patent Document 1, when a change occurs such as turning to the side or rotating with respect to the face to be detected, the reaction deteriorates in many image features. Arise. That is, the detection rate for a face in which a specific change has occurred is low.
 また、特許文献1では、複数方向毎の特徴パターンを用いて方向推定を行うため処理時間が増大し、更には顔方向推定を間違えると顔が検出できないという問題がある。このため、膨大な特徴量評価演算を追加することなく、回転などの特定の変化が生じた顔に対する検出率を向上させる必要があった。 Further, in Patent Document 1, since the direction estimation is performed using the feature pattern for each of a plurality of directions, there is a problem that the processing time increases, and further, if the face direction estimation is wrong, the face cannot be detected. For this reason, it has been necessary to improve the detection rate for a face in which a specific change such as rotation has occurred without adding an enormous feature amount evaluation calculation.
 本発明は上記問題を鑑みてなされたものであり、本発明の目的は、特定の変化が生じた顔に対する検出率を向上させることにある。 The present invention has been made in view of the above problems, and an object of the present invention is to improve the detection rate for a face in which a specific change has occurred.
 本願は、上記課題を解決する手段を複数開示している。その一つが、例えば請求の範囲に記載された構成である。 This application discloses a plurality of means for solving the above problems. One of them is, for example, the configuration described in the claims.
 本発明によれば、オブジェクト検出率を高めることができる。 According to the present invention, the object detection rate can be increased.
実施例1のオブジェクト検出装置の構成ブロック図である。1 is a block diagram illustrating a configuration of an object detection apparatus according to a first embodiment. 図1の画像入力部101の処理の例を示す説明図である。It is explanatory drawing which shows the example of a process of the image input part 101 of FIG. 図1の特徴評価部102で処理を行う特徴パターンの例を示す図である。It is a figure which shows the example of the feature pattern processed by the feature evaluation part 102 of FIG. 図1の特徴パターンDB110に格納されているオブジェクト判別で用いる特徴パターンのデータの一例を示すテーブルである。It is a table which shows an example of the data of the feature pattern used by the object discrimination | determination stored in feature pattern DB110 of FIG. 図1の変化オブジェクト判別部105で行われる処理の手順を示すフローチャートである。2 is a flowchart illustrating a procedure of processing performed by a change object determination unit 105 in FIG. 1. 図1の変化オブジェクト判別部105で用いられる各特徴量に対する重みの例を示す図である。It is a figure which shows the example of the weight with respect to each feature-value used by the change object discrimination | determination part 105 of FIG. 実施例2の高速オブジェクト判定を示す図である。It is a figure which shows the high-speed object determination of Example 2. FIG. 実施例3で用いられる特徴量の一例を示す図である。FIG. 10 is a diagram illustrating an example of feature amounts used in the third embodiment. 実施例4に関わるオブジェクト検出装置の構成を示す構成図である。FIG. 10 is a configuration diagram illustrating a configuration of an object detection apparatus according to a fourth embodiment. 実施例5に関わるオブジェクト検出装置の構成を示す構成図である。FIG. 10 is a configuration diagram illustrating a configuration of an object detection apparatus according to a fifth embodiment. 実施例5におけるパラメータ設定を行う設定画面の例を示す図である。FIG. 10 is a diagram illustrating an example of a setting screen for performing parameter setting in the fifth embodiment.
 以下、実施例を説明する。 Hereinafter, examples will be described.
 実施例1について図1を用いて説明する。オブジェクト検出部100は、画像入力部101、特徴パターンDB110、特徴評価部102、特徴量格納部103、オブジェクト判別部104、変化オブジェクト判別部105、判別結果出力部106から構成される。上記の各部はハードウェアによって構成されていてもよい。また、ハードウェアとソフトウェアを組合せたモジュールであってもよい。 Example 1 will be described with reference to FIG. The object detection unit 100 includes an image input unit 101, a feature pattern DB 110, a feature evaluation unit 102, a feature amount storage unit 103, an object determination unit 104, a changed object determination unit 105, and a determination result output unit 106. Each unit described above may be configured by hardware. Further, it may be a module combining hardware and software.
 図1に示したオブジェクト検出部の動作について、顔を検出する場合を例にして説明する。オブジェクト検出部にて検出するオブジェクトとしては、顔以外にも、人、車、標識等、その他のオブジェクトを対象としてもよい。説明の簡略化のため、以下ではオブジェクト検出動作の一例として、入力画像から人物の正面顔を検出する動作を例に説明する。 The operation of the object detection unit shown in FIG. 1 will be described using a case where a face is detected as an example. As an object to be detected by the object detection unit, in addition to the face, other objects such as a person, a car, and a sign may be targeted. For simplification of description, an operation for detecting a front face of a person from an input image will be described as an example of an object detection operation.
 画像入力部101は、カメラなどの撮像モジュールや、あらかじめ記録された画像の再生画像等を受信し、顔か非顔かの判別処理を行う画像領域203を特徴評価部102に出力する。特徴パターンDB110には顔か非顔かを判別するための特徴パターンが格納されている。特徴評価部102は、入力された画像領域203に対して特徴パターンDB110に定義された複数の特徴パターンに対する特徴量を算出して特徴量格納部103に格納する。特徴量格納部103では、特徴評価部102から得られた特徴量を格納する。オブジェクト判別部104では、特徴評価部102で得られた複数の特徴量の値を元に、顔か非顔かの判別処理を行う。その結果、顔と判別したら判別結果出力部106に結果を出力し、非顔と判別したら変化オブジェクト判別部105に結果を出力する。変化オブジェクト判別部105では、非顔と判別された画像領域203が、横向き顔などの特定変化が生じた顔か非顔かを特徴量格納部103に格納された特徴量を用いて判別する。変化オブジェクト判別部105では、オブジェクト判別部104とは異なるオブジェクトを見つけるため、オブジェクト判別部104とは異なる判別処理が行われる。変化オブジェクト判別部105における判別結果を判別結果出力部106に出力する。判別結果出力部106では、オブジェクト判別部104で顔と判別された画像領域または変化オブジェクト判別部105で特定変化が生じた顔と判別された画像領域を顔として、それ以外を非顔として判別結果を出力する。 The image input unit 101 receives an imaging module such as a camera, a reproduction image of a pre-recorded image, and the like, and outputs an image region 203 for performing a discrimination process between a face and a non-face to the feature evaluation unit 102. The feature pattern DB 110 stores a feature pattern for determining whether it is a face or a non-face. The feature evaluation unit 102 calculates feature amounts for a plurality of feature patterns defined in the feature pattern DB 110 for the input image region 203 and stores them in the feature amount storage unit 103. The feature amount storage unit 103 stores the feature amount obtained from the feature evaluation unit 102. The object discriminating unit 104 performs discrimination processing of a face or a non-face based on a plurality of feature value values obtained by the feature evaluation unit 102. As a result, if it is determined as a face, the result is output to the determination result output unit 106, and if it is determined as a non-face, the result is output to the change object determination unit 105. The changed object discriminating unit 105 discriminates whether the image area 203 discriminated as a non-face is a face or a non-face that has undergone a specific change, such as a sideways face, using the feature amount stored in the feature amount storage unit 103. In the change object determination unit 105, in order to find an object different from the object determination unit 104, a determination process different from that of the object determination unit 104 is performed. The discrimination result in the change object discrimination unit 105 is output to the discrimination result output unit 106. In the discrimination result output unit 106, the image region determined as the face by the object discrimination unit 104 or the image region discriminated as the face having undergone the specific change by the changed object discrimination unit 105 is set as a face, and the other is determined as a non-face. Is output.
 上記の構成をとることで、斜め横方向を向くなどの変化が生じたことで検出できなくなった顔に対して、その変化顔に対して固有の出力を示す特徴量に重点を置いた判別処理を行うことができ、多少の変化に対して頑健な顔検出を行うことが可能となる。更には、変化した顔に対する判別処理を正面顔検出において得られた特徴量を用いてそのまま行うことができるため、高速に上記の判別処理が可能となる。 With the above configuration, for a face that can no longer be detected due to a change such as being directed obliquely laterally, a discrimination process that focuses on the feature value indicating the unique output for the changed face It is possible to perform face detection that is robust against slight changes. Furthermore, since the discrimination process for the changed face can be performed as it is using the feature amount obtained in the front face detection, the above discrimination process can be performed at high speed.
 なお、顔に対する特定変化の一例として、顔が斜め横方向を向く、顔が回転する、顔の一部が隠れるなどの変化が考えられる。以下ではオブジェクト検出の一例として、左右斜め横方向を向いた顔を高精度に検出する例について説明する。本オブジェクト検出部において判別する顔の特定変化のパターンは、1つまたは複数でも良い。 Note that, as an example of the specific change with respect to the face, changes such as the face turning obliquely in the horizontal direction, the face rotating, or a part of the face being hidden can be considered. Hereinafter, as an example of object detection, an example will be described in which a face facing diagonally to the left and right is detected with high accuracy. One or a plurality of specific change patterns of the face determined by the object detection unit may be used.
 図1の画像入力部101の動作の詳細について図2を用いて説明する。画像入力部101で得られた入力画像200に対して、検出すべき顔201、202は任意の位置に任意の大きさで存在する。これに対応するため、画像入力部101では入力画像200上の複数位置・大きさの画像領域203を例えばラスタスキャン状に切出していき、特徴評価部102に出力する。特徴評価部102以降では、ある1枚の入力画像200中の複数の画像領域に対して顔か非顔かの判別処理が行われることになる。これによって、入力画像200中の任意の場所に存在する任意の大きさの顔を検出する。 Details of the operation of the image input unit 101 in FIG. 1 will be described with reference to FIG. With respect to the input image 200 obtained by the image input unit 101, the faces 201 and 202 to be detected exist at arbitrary positions and in arbitrary sizes. In order to cope with this, the image input unit 101 cuts out image regions 203 having a plurality of positions and sizes on the input image 200, for example, in a raster scan form, and outputs them to the feature evaluation unit 102. In the feature evaluation unit 102 and later, a process for determining whether a face is a non-face is performed on a plurality of image areas in a single input image 200. As a result, a face of an arbitrary size present at an arbitrary location in the input image 200 is detected.
 図1の特徴パターンDB110について説明する。特徴パターンDB110には顔判別に用いる複数の画像特徴パラメータが定義される。図3は、定義される画像特徴パターンの一例である。図3の特徴パターンは、黒矩形301と白矩形302から構成されており、この矩形内の画素値の総和の差分によって特徴量をもとめる。図4は、この特徴パラメータを定義した顔判別パラメータテーブル400の一例である。顔と非顔を判別するためN個の画像特徴h(i∈N)が定義されており、各特徴に対して顔判定を行う際の重みα、右斜め横顔判定を行う際の重みβ、左斜め横顔判定を行う際の重みγがそれぞれ定義されている。 The feature pattern DB 110 in FIG. 1 will be described. The feature pattern DB 110 defines a plurality of image feature parameters used for face discrimination. FIG. 3 is an example of a defined image feature pattern. The feature pattern in FIG. 3 is composed of a black rectangle 301 and a white rectangle 302, and the feature amount is obtained by the difference in the sum of the pixel values in the rectangle. FIG. 4 is an example of a face discrimination parameter table 400 in which the feature parameters are defined. N image features h i (i∈N) are defined to discriminate between a face and a non-face, and a weight α when performing face determination for each feature and a weight β when performing right oblique side face determination , A weight γ for the left oblique side face determination is defined.
 図1の特徴評価部102の動作の詳細について説明する。特徴評価部102では、入力された画像領域203に対する複数の特徴量を算出する。特徴パターンDB110内に定義された画像特徴パターンhに対して入力された画像領域203をベクトルIとすると、得られる特徴量h(I)は、数式1で求めることができる。 Details of the operation of the feature evaluation unit 102 in FIG. 1 will be described. The feature evaluation unit 102 calculates a plurality of feature amounts for the input image region 203. If the image area 203 input to the image feature pattern h i defined in the feature pattern DB 110 is a vector I, the obtained feature quantity h i (I) can be obtained by Equation 1.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 特徴評価部102では上記の演算によって得られたN個の特徴量を、特徴量格納部103に評価結果を格納する。 The feature evaluation unit 102 stores the N feature amounts obtained by the above calculation in the feature amount storage unit 103.
 なお、特徴評価部102で評価する特徴量として矩形特徴を用いた例を示したが、検出対象とするオブジェクトに応じて、EOH(Edge of Orientation Histograms)やHOG(Histogram of Orientation Gradients)などの異なる特徴量を用いても良い。 In addition, although the example which used the rectangular feature as the feature-value evaluated by the feature evaluation part 102 was shown, according to the object made into a detection target, EOH (Edge of Orientation Histograms), HOG (Histogram of Orientation Gradients), etc. differ. A feature amount may be used.
 図1のオブジェクト判別部104の動作の詳細について説明する。オブジェクト判別部104では、特徴評価部102で得られた各特徴量h(I)を元に顔尤度を算出して画像領域203が顔か非顔かの判別を行う。顔判別の例としてAdaBoost識別器では、図4の顔判別パラメータテーブル400に記述された各特徴hに対する顔判別の重みαを用いた線形和の関数で顔尤度F(I)を算出する。 Details of the operation of the object determination unit 104 in FIG. 1 will be described. The object discriminating unit 104 calculates face likelihood based on each feature quantity h i (I) obtained by the feature evaluation unit 102 and discriminates whether the image region 203 is a face or a non-face. As an example of face discrimination, the AdaBoost classifier calculates the face likelihood F (I) by a linear sum function using the face discrimination weight α i for each feature h i described in the face discrimination parameter table 400 of FIG. To do.
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 顔尤度F(I)が大きいほど顔らしいことを示しており、顔尤度Fが閾値Th以上であれば顔、未満であれば非顔と判別する。オブジェクト判別部104では、判別結果が顔であれば、その結果を判別結果出力部106に出力する。判別結果が非顔であれば、変化オブジェクト判別部105に結果を出力する。 A larger face likelihood F (I) indicates that the face is more likely to be a face. If the face likelihood F is greater than or equal to the threshold Th F , it is determined to be a face, and if it is less, it is determined to be a non-face. If the discrimination result is a face, the object discrimination unit 104 outputs the result to the discrimination result output unit 106. If the determination result is a non-face, the result is output to the change object determination unit 105.
 図1の変化オブジェクト判別部105の動作の詳細について図5を用いて説明する。図5は、変化オブジェクト判別部105において右斜め顔と左斜め顔を判別するステップを示すフローチャートである。 Details of the operation of the change object determination unit 105 in FIG. 1 will be described with reference to FIG. FIG. 5 is a flowchart showing steps of determining the right oblique face and the left oblique face in the change object discriminating unit 105.
 ステップ501では、特徴量格納部103に格納された画像領域203に対する特徴量hi(I)を読み込む。 In step 501, the feature quantity h i (I) for the image area 203 stored in the feature quantity storage unit 103 is read.
 ステップ502では、特徴量から右斜め顔に対する尤度Rを算出する。ステップ502における尤度Rの算出は、オブジェクト判別部104と同じ特徴量を用いるが、判別する対象が異なるため、異なる判別処理が行われる。右斜め顔尤度Rは、図4の顔判別パラメータテーブル400に記述された各特徴hに対する右斜め顔の重みβを使った線形和の関数である数式3の計算で算出される。 In step 502, the likelihood R for the right oblique face is calculated from the feature amount. The likelihood R is calculated in step 502 using the same feature amount as that of the object determination unit 104, but a different determination process is performed because a determination target is different. The right oblique face likelihood R is calculated by the calculation of Equation 3 which is a function of a linear sum using the right oblique face weight β i for each feature h i described in the face discrimination parameter table 400 of FIG.
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
 ステップ503では、右斜め顔尤度Rを右斜め顔判別閾値Threと比較して、閾値以上であればステップ507へ、それ以外ならステップ504へと進む。 In step 503, the right oblique face likelihood R is compared with the right oblique face discrimination threshold Thre R, and if it is equal to or greater than the threshold, the process proceeds to step 507, and if not, the process proceeds to step 504.
 ステップ504では、右斜め顔の場合と同様に左斜め顔に対する尤度Lを算出する。 In step 504, the likelihood L for the left oblique face is calculated as in the case of the right oblique face.
 左斜め顔尤度Lは、右斜め顔尤度Rは、図4の顔判別パラメータテーブル400に記述された各特徴hに対する左斜め顔の重みγを使った線形和の関数である数式4の計算で算出される。 The left oblique face likelihood L is a mathematical expression that is a function of a linear sum using the left oblique face weight γ i for each feature h i described in the face discrimination parameter table 400 of FIG. 4 is calculated.
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
 ステップ505では、左斜め顔尤度Lを左斜め顔判別閾値Threと比較して、閾値以上であればステップ507へ、それ以外ならステップ506へと進む。 In step 505, the left oblique face likelihood L is compared with the left oblique face discrimination threshold Thre L, and if it is equal to or greater than the threshold, the process proceeds to step 507, and if not, the process proceeds to step 506.
 図6は、図5のステップ502、504における顔尤度算出の概念を示す図である。図6(a)は正面顔に対する顔尤度の算出を示す図であり、これに対して図6(b)(c)は、それぞれ右斜め顔尤度と左斜め顔尤度の算出を示す図である。図6では、特徴量601~605に対する重みを濃淡で表現している。図6(a)では、一般的な正面顔を判別するため特徴量601~605に対してほぼ均等な重みが振られることになる。一方、図6(b)は、右斜め顔に対しては特徴が表れにくいh1、h3、h4などの特徴量に対しては重みが小さくなり、h2、h5などの重みが大きいことを示している。これにより、正面顔と非顔の判定とは異なる基準で、右斜め顔と非顔の判別を行うことが可能となる。 FIG. 6 is a diagram showing the concept of face likelihood calculation in steps 502 and 504 of FIG. FIG. 6A is a diagram showing calculation of face likelihood for a front face, while FIGS. 6B and 6C show calculation of right diagonal face likelihood and left diagonal face likelihood, respectively. FIG. In FIG. 6, the weights for the feature amounts 601 to 605 are expressed by shading. In FIG. 6A, almost equal weights are assigned to the feature amounts 601 to 605 in order to discriminate a general front face. On the other hand, FIG. 6B shows that weights are small for feature amounts such as h1, h3, and h4, and features such as h2, h5, and the like are large. Yes. As a result, it is possible to discriminate between the right diagonal face and the non-face based on different criteria from the front face and non-face determination.
 ステップ506は、右斜め顔でも左斜め顔でもないと判別された場合に訪れるステップである。このため、ステップ506では変化オブジェクト判別部における最終的な判別結果を非顔と決定する。一方、ステップ507は、右斜め顔または左斜め顔と判別した場合に訪れる。このためステップ507は、判別結果を顔と決定する。 Step 506 is a step visited when it is determined that neither the right diagonal face nor the left diagonal face is detected. For this reason, in step 506, the final discrimination result in the change object discrimination unit is determined as a non-face. On the other hand, step 507 is visited when it is determined that the face is a right diagonal face or a left diagonal face. For this reason, step 507 determines the discrimination result as a face.
 上記の処理フローをとることで、オブジェクト判別部104で非顔と判別された画像領域に対して、右斜め顔または左斜め顔かどうかを再判別することができる。この再判別処理は、あらかじめ算出して特徴量格納部103に格納された複数の特徴量を用いて行うため、新たに画像領域203に対する特徴量算出を要せず、高速に判別できる。 By taking the above processing flow, it is possible to re-determine whether the image area determined to be a non-face by the object determination unit 104 is a right diagonal face or a left diagonal face. This re-discrimination process is performed using a plurality of feature amounts calculated in advance and stored in the feature amount storage unit 103, so that it is not necessary to newly calculate a feature amount for the image region 203 and can be determined at high speed.
 なお、変形顔の判別に複数の特徴量に対する線形和を用いたが、PCAや非線形サポートベクターマシン(SVM)などの判別器を用いて、オブジェクト判別部104と変化オブジェクト判別部105で異なる判別手法を用いて判別してもよい。 In addition, although the linear sum with respect to several feature-values was used for the discrimination | determination of a deformed face, the discrimination methods different in the object discrimination | determination part 104 and the change object discrimination | determination part 105 using discriminators, such as PCA and a nonlinear support vector machine (SVM). You may discriminate | determine using.
 このように、実施例1では、入力画像から顔を検出する顔検出装置であって、画像を入力する画像入力部と、前記画像入力部で得られた画像に対して特徴量を算出する特徴評価部と、算出した前記特徴量を格納する特徴量格納部と、前記特徴評価部で得られた特徴量と正面顔用関数とで線形和を算出して、算出した線形和に基づいて前記画像が正面顔か非正面顔かを判別するオブジェクト判別部と、前記オブジェクト判別部で非正面顔と判別された画像に対して、前記特徴量格納部に格納された特徴量とオブジェクト判別部とは重み係数を変えた斜め顔用関数とで線形和を算出して、当該画像が非顔なのか、斜め顔なのかを判別する変化オブジェクト判別部と、前記オブジェクト判別の結果を出力する判別結果出力部を有することを特徴とする顔検出装置を開示している。この顔検出装置であれば、遮蔽物がある顔についての判定も可能になるので、顔検出率が向上する。 As described above, in the first embodiment, the face detection apparatus detects a face from an input image, and includes an image input unit that inputs an image, and a feature that calculates a feature amount for the image obtained by the image input unit. A linear sum is calculated by the evaluation unit, the feature amount storage unit that stores the calculated feature amount, the feature amount obtained by the feature evaluation unit and the front face function, and based on the calculated linear sum, An object discriminating unit that discriminates whether the image is a front face or a non-front face; and an image that has been discriminated as a non-front face by the object discriminating unit; Calculates a linear sum with an oblique face function with a different weighting factor, and determines whether the image is a non-face or an oblique face, and a discrimination result that outputs the result of the object discrimination Characterized by having an output part It discloses a face detection apparatus that. With this face detection device, it is possible to determine a face with an obstruction, so that the face detection rate is improved.
 さらに、変化オブジェクト判別部で、部分遮蔽顔用関数の重み付け係数を付与する特徴点は、正面顔の特徴点の一部を含むようにすれば、特徴量の再算出が省け、計算量が減るので、高速顔検出が実現できるようになる。 Furthermore, if the feature points to which the weighting coefficient of the function for partial occlusion face is included in the change object discriminating unit include part of the feature points of the front face, the feature amount is not recalculated and the calculation amount is reduced. Therefore, high-speed face detection can be realized.
 実施例2について図7を用いて説明する。本実施例は、オブジェクト判別部104で顔/非顔の判別を行うために図7に示すようにCascade型識別器を用いることで高速に顔検出を行う例である。 Example 2 will be described with reference to FIG. The present embodiment is an example in which face detection is performed at high speed by using a cascade type discriminator as shown in FIG.
 図7では、オブジェクト判別部104における顔と非顔の判別器が、複数個の識別器710~730を縦続接続した構成となっている。識別器1(710)では、特徴量格納部103に格納されるN個の特徴量のうち、特徴量1から特徴量A(A<N)までの特徴量セット(701)だけを用いて判別処理を行い、画像領域が顔か非顔かを判別する。ここで、非顔と判別された場合(識別器1(710)の判別結果が”False”の時)は処理を終了する。この際、特徴評価部102における特徴量算出も特徴量Aまでしか処理を行わずに終了する。一方、画像領域が顔と判別された場合(識別器1(710)の判別結果が”True”の時)は、判別処理を識別器2(720)に委ねる。識別器2(720)では、特徴量A+1からBまでの特徴量セット(702)を用いて判別処理を行って、同様に画像領域が顔か非顔かを判別する。上記の処理を最後の識別器S(730)まで実施して識別器S(730)でも顔と判別したら、判別結果を顔とする。ただし、識別器S(730)でFalseと判定した場合には、変化オブジェクト判別部105にて斜め顔か非顔かの判別処理を行う。変化オブジェクト判別部105では、識別器1~Sの判別処理に用いられて特徴量格納部(103)に格納されたN個全ての特徴量を用いて、図5のフローチャートに従って左右斜め顔か非顔かの判別を行う。 In FIG. 7, the face / non-face discriminator in the object discriminating unit 104 has a configuration in which a plurality of discriminators 710 to 730 are connected in cascade. The discriminator 1 (710) uses only the feature value set (701) from the feature value 1 to the feature value A (A <N) out of the N feature values stored in the feature value storage unit 103. Processing is performed to determine whether the image area is a face or a non-face. Here, when it is determined that the face is non-face (when the determination result of the discriminator 1 (710) is “False”), the processing ends. At this time, the feature amount calculation in the feature evaluation unit 102 is completed only after the feature amount A is processed. On the other hand, when the image area is determined to be a face (when the determination result of the identifier 1 (710) is “True”), the determination processing is left to the identifier 2 (720). The discriminator 2 (720) performs discrimination processing using the feature amount set (702) from the feature amounts A + 1 to B, and similarly determines whether the image area is a face or a non-face. When the above processing is performed up to the last discriminator S (730) and the discriminator S (730) discriminates the face, the discrimination result is set as a face. However, if the discriminator S (730) determines False, the change object discriminating unit 105 determines whether the face is an oblique face or a non-face. The changing object discriminating unit 105 uses the N feature amounts stored in the feature amount storage unit (103) used in the discriminating processes of the discriminators 1 to S to determine whether the left and right diagonal faces are not in accordance with the flowchart of FIG. Determine if it is a face.
 上記の構成によって、顔でも斜め顔でもなく明らかに非顔と判定できる領域に対しては、全ての特徴量演算を行わずに判別処理を途中終了できるため、高速に顔検出が可能となる。 With the above-described configuration, it is possible to detect the face at high speed because it is possible to end the discrimination process halfway without performing all the feature amount calculations for an area that is clearly determined to be a non-face that is neither a face nor an oblique face.
 実施例3について図8を用いて説明する。図8(a)は変化オブジェクト判別部(105)において、サングラスを掛けた人物の顔を検出する場合に高い重み値が設定される特徴量(801)~(803)のイメージの一例を示した図である。図8(b)は同様にマスクを付けた人物の顔を検出する場合に、高い重み値が設定される特徴量(804)~(807)のイメージの一例を示した図である。図1の構成を持つオブジェクト検出部100を用いることで、例示したサングラスやマスクに対して反応する特徴量に重みを付けた判別処理を行うことができ、顔に対する装着品の有無を検出することが可能となる。 Example 3 will be described with reference to FIG. FIG. 8A shows an example of images of feature amounts (801) to (803) in which a high weight value is set when the face of a person wearing sunglasses is detected in the change object discriminating unit (105). FIG. FIG. 8B is a diagram showing an example of images of feature amounts (804) to (807) in which a high weight value is set when detecting the face of a person with a mask similarly. By using the object detection unit 100 having the configuration of FIG. 1, it is possible to perform a discrimination process in which feature quantities that react to the exemplified sunglasses and mask are weighted, and to detect the presence / absence of a wearing item on the face. Is possible.
 変化オブジェクト判別部105においてこうした装着品の有無を判別し、判別結果出力部106において、顔に対する装着品有無の追加情報を出力することで、顔に対する特定のメタデータ情報を収集することが可能となる。 It is possible to collect specific metadata information for a face by determining whether or not such a wearing item is present in the change object discriminating unit 105 and outputting additional information on the presence or absence of the wearing item for the face in the discrimination result output unit 106. Become.
 なお、顔に対する装着品の一例としては、サングラスやマスクのほかに、メガネ、ヒゲ、帽子、眼帯などが考えられ、それぞれの変化オブジェクトを判別するための判別器をあらかじめ特徴パターンDB110に用意しておくことで、これらの装着品を装着した顔かどうかを判別することが可能となる。 In addition to sunglasses and masks, glasses, beards, hats, eye patches, and the like can be considered as examples of wearing items for the face. A classifier for discriminating each change object is prepared in the feature pattern DB 110 in advance. It is possible to determine whether the face is equipped with these wearing items.
 このように、実施例3では、入力画像から顔を検出する顔検出装置であって、画像を入力する画像入力部と、前記画像入力部で得られた画像に対して特徴量を算出する特徴評価部と、算出した前記特徴量を格納する特徴量格納部と、前記特徴評価部で得られた特徴量と正面顔用関数とで線形和を算出して、算出した線形和に基づいて前記画像が正面顔か非正面顔かを判別するオブジェクト判別部と、前記オブジェクト判別部で非正面顔と判別された画像に対して、前記特徴量格納部に格納された特徴量とオブジェクト判別部とは重み係数を変えた部分遮蔽顔用関数とで線形和を算出して、当該画像が非顔なのか、部分的に何かに遮蔽されただけの顔なのかを検出する変化オブジェクト判別部を有することを特徴とする顔検出装置を開示している。この顔検出装置であれば、遮蔽物がある顔についての判定も可能になるので、顔検出率が向上する。 As described above, the third embodiment is a face detection device that detects a face from an input image, and includes an image input unit that inputs an image, and a feature that calculates a feature amount for the image obtained by the image input unit. A linear sum is calculated by the evaluation unit, the feature amount storage unit that stores the calculated feature amount, the feature amount obtained by the feature evaluation unit and the front face function, and based on the calculated linear sum, An object discriminating unit that discriminates whether the image is a front face or a non-front face; and an image that has been discriminated as a non-front face by the object discriminating unit; Calculates a linear sum with a partially occluded face function with a different weighting coefficient, and a change object discriminating unit that detects whether the image is a non-face or a face that is partially occluded by something. Disclosed is a face detection device characterized by having That. With this face detection device, it is possible to determine a face with an obstruction, so that the face detection rate is improved.
 さらに、変化オブジェクト判別部で、部分遮蔽顔用関数の重み付け係数を付与する特徴点は、正面顔の特徴点の一部を含むようにすれば、特徴量の再算出が省け、計算量が減るので、高速顔検出が実現できるようになる。 Furthermore, if the feature points to which the weighting coefficient of the function for partial occlusion face is included in the change object discriminating unit include part of the feature points of the front face, the feature amount is not recalculated and the calculation amount is reduced. Therefore, high-speed face detection can be realized.
 実施例4に係る装置の構成について図9を用いて説明する。本実施例は、実施例1乃至3に係るオブジェクト検出装置を監視カメラやデジタルカメラなどの撮像装置や、ディスプレイ、映像記録装置に実装した例である。 The configuration of the apparatus according to the fourth embodiment will be described with reference to FIG. The present embodiment is an example in which the object detection apparatus according to the first to third embodiments is mounted on an imaging apparatus such as a monitoring camera or a digital camera, a display, or a video recording apparatus.
 図9は実施例4に係るオブジェクト検出装置の構成を示す構成図である。図9において、オブジェクト検出装置(900)は、画像入力部(909)、画像メモリ(902)、CPU(903)、RAM(904)、ROM(905)、検出結果記録部(906)、インタフェース(907)、出力装置(908)から構成されている。 FIG. 9 is a configuration diagram illustrating the configuration of the object detection device according to the fourth embodiment. In FIG. 9, an object detection apparatus (900) includes an image input unit (909), an image memory (902), a CPU (903), a RAM (904), a ROM (905), a detection result recording unit (906), an interface ( 907) and an output device (908).
 本実施例のオブジェクト検出装置(900)では、撮像部(901)であるカメラで得られた画像から対象となるオブジェクトを検出する。本実施例のオブジェクト検出装置(900)におけるCPU(903)が、実施例1の図1に示すオブジェクト検出部(100)に相当し、オブジェクト検出部(100)の各演算処理を、プログラムとして実行しCPU(903)により演算処理することにより実現している。 In the object detection apparatus (900) of the present embodiment, a target object is detected from an image obtained by a camera that is an imaging unit (901). The CPU (903) in the object detection apparatus (900) of the present embodiment corresponds to the object detection unit (100) shown in FIG. 1 of the first embodiment, and executes each arithmetic processing of the object detection unit (100) as a program. This is realized by performing arithmetic processing by the CPU (903).
 本実施例では、オブジェクト検出部(100)における検出方法に従ってCPU(903)において演算処理が行われ、オブジェクトが検出される。 In this embodiment, the CPU (903) performs arithmetic processing according to the detection method in the object detection unit (100) to detect the object.
 シーケンス毎のオブジェクト検出結果は、検出結果記録部(906)に記録されることとなる。検出結果は、インタフェース(907)を通じて適切な形にデータ変換され、出力装置(908)に出力される。ここで、出力装置としてはディスプレイや、プリンタ、PCなどが考えられる。 The object detection result for each sequence is recorded in the detection result recording unit (906). The detection result is converted into an appropriate form through the interface (907) and output to the output device (908). Here, the output device may be a display, a printer, a PC, or the like.
 本実施例では、コンピュータなどの情報処理装置により、オブジェクト検出装置としての演算処理を行うことが可能である。 In this embodiment, it is possible to perform arithmetic processing as an object detection device by an information processing device such as a computer.
 以上説明した実施例4によれば、画像中に存在するオブジェクトを高精度に検出するオブジェクト検出機能を有する撮像装置やディスプレイ、映像記録装置を実現できる。 According to the fourth embodiment described above, it is possible to realize an imaging device, a display, and a video recording device having an object detection function for detecting an object existing in an image with high accuracy.
 実施例5に係る装置の構成について、図10により説明する。本実施例は、実施例4に係るオブジェクト検出装置(900)に対して、図10に示す入力装置(1010)と設定制御部(1020)を設けることで、利用用途や環境に応じたパラメータ設定が可能なオブジェクト検出装置(1000)を構成する例である。 The configuration of the apparatus according to the fifth embodiment will be described with reference to FIG. In this embodiment, the object detection apparatus (900) according to the fourth embodiment is provided with the input device (1010) and the setting control unit (1020) shown in FIG. It is an example which comprises the object detection apparatus (1000) which can do.
 本実施例のオブジェクト検出装置(1000)では、入力装置(1010)から図1の変化オブジェクト判別部(105)における変化オブジェクト判別処理におけるパラメータ調整の命令を受ける。本命令を受け取った設定制御部(1020)において、変化オブジェクト判別部(105)における変形オブジェクトに対する判別処理のON/OFF制御や感度の調整などのパラメータ制御を行う。これによりオブジェクト検出装置において判別する予め用意した複数の変形オブジェクトに対して、判別する対象を制限または拡大することが可能となる。 In the object detection apparatus (1000) of the present embodiment, a command for parameter adjustment in the change object determination process in the change object determination unit (105) of FIG. 1 is received from the input device (1010). The setting control unit (1020) that has received this command performs parameter control such as ON / OFF control of discrimination processing for the deformed object in the changed object discrimination unit (105) and sensitivity adjustment. As a result, it is possible to limit or expand the discrimination target for a plurality of deformed objects prepared in advance to be discriminated in the object detection device.
 図11は、入力装置(1010)で変形オブジェクト判別におけるパラメータ設定を行うためのパラメータ設定画面(1100)の例を示した図である。パラメータ設定画面(1100)は、オブジェクト判別パラメータ(1101)と、変形オブジェクト判別パラメータ(1102~1107)によって構成される。オブジェト判別パラメータ(1101)でオブジェクト判別部(104)における顔検出の感度を制御する。例えば高感度に設定すると、特徴パターンDB(110)に定義された顔判別パラメータテーブル(400)の閾値Tをゆるめることで、顔と判定しやすくなる。 FIG. 11 is a diagram showing an example of a parameter setting screen (1100) for performing parameter setting in deformed object discrimination with the input device (1010). The parameter setting screen (1100) includes an object discrimination parameter (1101) and deformed object discrimination parameters (1102 to 1107). The object discrimination parameter (1101) controls the sensitivity of face detection in the object discrimination unit (104). For example, when high sensitivity is set, it is easy to determine a face by loosening the threshold TF of the face discrimination parameter table (400) defined in the feature pattern DB (110).
 変形オブジェト判別パラメータ(1102~1107)では左右斜め顔、左右回転顔などのそれぞれの変形オブジェクトに対して変化オブジェクト判別部(105)において判別処理を行うかどうか、行う際の感度調整はどうするかといったパラメータを設定する。こうして設定した各項目のパラメータ情報は、設定制御部(1020)に送られ、設定制御部においてCPUでの検出処理の制御を行う。 In the deformed object discriminating parameters (1102 to 1107), whether or not the changing object discriminating unit (105) performs discriminating processing on each deformed object such as a left-right slanted face and a left-right rotated face, and how sensitivity adjustment is performed. Set the parameters. The parameter information of each item set in this way is sent to the setting control unit (1020), and the setting control unit controls the detection process in the CPU.
 以上の構成をとることで検出したい変形オブジェクトに特化したオブジェクト検出を行うことが可能となる。これにより、製品に対する検出感度の調整や、使用する環境に応じた特性変更が行える。例えば、監視カメラ映像における顔検出を例に挙げると、カメラの設置状況に応じて、映像中に現れる顔にばらつきが生じる。カメラ設置位置や設置環境における人物動線などの関係から、下向き顔が多く現れる状況や、左向きの顔よりも右向きの顔が多く現れるといった状況が起こりうる。カメラ毎に判別する変形オブジェクトのパターンを設定することで、高精度に所望の顔検出が可能となる。 By adopting the above configuration, it is possible to perform object detection specialized for the deformed object to be detected. This makes it possible to adjust the detection sensitivity for the product and change the characteristics according to the environment in which it is used. For example, when face detection in a surveillance camera image is taken as an example, the face appearing in the image varies depending on the installation status of the camera. Depending on the relationship between the camera installation position and the person flow line in the installation environment, a situation in which many downward faces appear or a situation in which more right faces appear than left faces may occur. By setting a deformed object pattern to be determined for each camera, a desired face can be detected with high accuracy.
100…オブジェクト検出部
101…画像入力部
102…特徴評価部
103…特徴量格納部
104…オブジェクト判別部
105…変化オブジェクト判別部
106…判別結果出力部
110…特徴パターンDB
200…入力画像
201、202…顔
203…画像領域
301、302…矩形特徴
400…顔判別パラメータテーブル
501~507…ステップ
601~605…矩形特徴
701~703…特徴量セット
710…識別器1
720…識別器2
730…識別器S
801~807…矩形特徴
900…オブジェクト検出装置
901…撮像部
902…画像メモリ
903…CPU
904…RAM
905…ROM
906…検出結果記録部
907…インタフェース
908…出力装置
909…画像入力部
1000…オブジェクト検出装置
1010…入力装置
1020…設定制御部
1100…パラメータ設定画面
1101…オブジェト判別パラメータ
1102~1107…変形オブジェト判別パラメータ
 
DESCRIPTION OF SYMBOLS 100 ... Object detection part 101 ... Image input part 102 ... Feature evaluation part 103 ... Feature-value storage part 104 ... Object discrimination | determination part 105 ... Change object discrimination | determination part 106 ... Discrimination result output part 110 ... Feature pattern DB
200: input image 201, 202 ... face 203 ... image area 301, 302 ... rectangular feature 400 ... face discrimination parameter table 501 to 507 ... step 601 to 605 ... rectangular feature 701 to 703 ... feature quantity set 710 ... classifier 1
720 ... Classifier 2
730 ... Classifier S
801 to 807 ... Rectangular feature 900 ... Object detection device 901 ... Imaging unit 902 ... Image memory 903 ... CPU
904 ... RAM
905 ... ROM
906 ... Detection result recording unit 907 ... Interface 908 ... Output device 909 ... Image input unit 1000 ... Object detection device 1010 ... Input device 1020 ... Setting control unit 1100 ... Parameter setting screen 1101 ... Object discrimination parameters 1102 to 1107 ... Deformed object discrimination parameters

Claims (10)

  1.  入力画像から特定のオブジェクトを検出するオブジェクト検出装置であって、
     画像を入力する画像入力部と、
     前記画像入力部で得られた画像に対して特徴量を算出する特徴評価部と、
     算出した前記特徴量を格納する特徴量格納部と、
     前記特徴評価部で得られた特徴量を用いて、前記画像がオブジェクトか非オブジェクトかを判別するオブジェクト判別部と、
     前記オブジェクト判別部でオブジェクトではないと判別された画像に対して、前記特徴量格納部に格納された特徴量を用いて前記オブジェクト判別部とは異なる判別方法で前記画像がある特定の変化が生じたオブジェクトか非オブジェクトかを判別する変化オブジェクト判別部と、
     前記オブジェクト判別の結果を出力する判別結果出力部を有することを特徴とするオブジェクト検出装置。
    An object detection device for detecting a specific object from an input image,
    An image input unit for inputting an image;
    A feature evaluation unit that calculates a feature amount for the image obtained by the image input unit;
    A feature amount storage unit for storing the calculated feature amount;
    An object discriminating unit that discriminates whether the image is an object or a non-object by using the feature amount obtained by the feature evaluation unit;
    For an image that is determined not to be an object by the object determination unit, a specific change occurs in the image by a determination method that is different from the object determination unit using the feature amount stored in the feature amount storage unit. A change object discriminator for discriminating whether the object is a non-object,
    An object detection apparatus comprising: a discrimination result output unit for outputting the object discrimination result.
  2.  請求項1において、
     前記変化オブジェクト判別部は、判別する変形オブジェクトの種類や判別の感度を調整するための入力部を有することを特徴とするオブジェクト検出装置。
    In claim 1,
    The object detection apparatus, wherein the change object determination unit includes an input unit for adjusting a type of a deformed object to be determined and a sensitivity of the determination.
  3.  請求項1において、
     前記変化オブジェクト判別部は、前記特徴評価部で算出した特徴量を用いた線形判別処理を行うことで変形オブジェクトか非オブジェクトかを判別することを特徴とするオブジェクト検出装置。
    In claim 1,
    The object detection device, wherein the change object determination unit determines whether the object is a deformed object or a non-object by performing a linear determination process using the feature amount calculated by the feature evaluation unit.
  4.  請求項1において、
     前記変化オブジェクト判別部は、前記特徴評価部で算出した特徴量を入力としたサポートベクターマシンによって判別することを特徴とするオブジェクト検出装置。
    In claim 1,
    The object detection apparatus according to claim 1, wherein the change object determination unit performs determination using a support vector machine having the feature amount calculated by the feature evaluation unit as an input.
  5.  請求項1において、
     前記変化オブジェクト判別部において、平面内/平面外回転が生じたオブジェクトを判別することを特徴とするオブジェクト検出装置。
    In claim 1,
    An object detection apparatus characterized in that the change object discrimination unit discriminates an object that has undergone in-plane / out-plane rotation.
  6.  請求項1において、
     前記変化オブジェクト判別部において、前記オブジェクトの一部に遮蔽が生じたオブジェクトを判別することを特徴とするオブジェクト検出装置。
    In claim 1,
    The object detection apparatus, wherein the change object determination unit determines an object in which a part of the object is blocked.
  7.  入力画像から顔を検出する顔検出装置であって、
     画像を入力する画像入力部と、
     前記画像入力部で得られた画像に対して特徴量を算出する特徴評価部と、
     算出した前記特徴量を格納する特徴量格納部と、
     前記特徴評価部で得られた特徴量と正面顔用関数とで線形和を算出して、算出した線形和に基づいて前記画像が正面顔か非正面顔かを判別するオブジェクト判別部と、
     前記オブジェクト判別部で非正面顔と判別された画像に対して、前記特徴量格納部に格納された特徴量と前記オブジェクト判別部とは重み係数を変えた斜め顔用関数とで線形和を算出して、当該画像が非顔なのか、斜め顔なのかを判別する変化オブジェクト判別部と、
     前記オブジェクト判別の結果を出力する判別結果出力部を有することを特徴とする顔検出装置。
    A face detection device for detecting a face from an input image,
    An image input unit for inputting an image;
    A feature evaluation unit that calculates a feature amount for the image obtained by the image input unit;
    A feature amount storage unit for storing the calculated feature amount;
    An object discriminating unit that calculates a linear sum of the feature amount obtained by the feature evaluation unit and the front face function, and discriminates whether the image is a front face or a non-front face based on the calculated linear sum;
    For an image determined as a non-front face by the object determination unit, a linear sum is calculated from the feature amount stored in the feature amount storage unit and the object determination unit using an oblique face function with a different weighting coefficient. Then, a change object determination unit that determines whether the image is a non-face or an oblique face,
    A face detection apparatus comprising a discrimination result output unit for outputting the result of object discrimination.
  8.  請求項7において、
     前記変化オブジェクト判別部で、斜め顔用関数の重み付け係数を付与する特徴点は、正面顔の特徴点の一部を含むことを特徴とする顔検出装置。
    In claim 7,
    The face detection device, wherein the feature point that assigns the weighting coefficient of the oblique face function in the change object determination unit includes a part of the feature point of the front face.
  9.  入力画像から顔を検出する顔検出装置であって、
     画像を入力する画像入力部と、
     前記画像入力部で得られた画像に対して特徴量を算出する特徴評価部と、
     算出した前記特徴量を格納する特徴量格納部と、
     前記特徴評価部で得られた特徴量と正面顔用関数とで線形和を算出して、算出した線形和に基づいて前記画像が正面顔か非正面顔かを判別するオブジェクト判別部と、
     前記オブジェクト判別部で非正面顔と判別された画像に対して、前記特徴量格納部に格納された特徴量と前記オブジェクト判別部とは重み係数を変えた部分遮蔽顔用関数とで線形和を算出して、当該画像が非顔なのか、部分的に何かに遮蔽されただけの顔なのかを検出する変化オブジェクト判別部を有することを特徴とする顔検出装置。
    A face detection device for detecting a face from an input image,
    An image input unit for inputting an image;
    A feature evaluation unit that calculates a feature amount for the image obtained by the image input unit;
    A feature amount storage unit for storing the calculated feature amount;
    An object discriminating unit that calculates a linear sum of the feature amount obtained by the feature evaluation unit and the front face function, and discriminates whether the image is a front face or a non-front face based on the calculated linear sum;
    For an image determined as a non-front face by the object determination unit, the feature amount stored in the feature amount storage unit and the object determination unit perform a linear sum with a partial occlusion face function with a weighting factor changed. A face detection apparatus comprising: a change object determination unit that calculates and detects whether the image is a non-face or a face that is partially covered by something.
  10.  請求項9において、
     前記変化オブジェクト判別部で、部分遮蔽顔用関数の重み付け係数を付与する特徴点は、正面顔の特徴点の一部を含むことを特徴とする顔検出装置。
     
     
    In claim 9,
    The face detection device, wherein the feature point to which the weighting coefficient of the partial occlusion face function is included in the change object determination unit includes a part of the feature point of the front face.

PCT/JP2010/051257 2010-01-29 2010-01-29 Object detection device and face detection device WO2011092848A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2010/051257 WO2011092848A1 (en) 2010-01-29 2010-01-29 Object detection device and face detection device
JP2011551643A JPWO2011092848A1 (en) 2010-01-29 2010-01-29 Object detection device and face detection device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2010/051257 WO2011092848A1 (en) 2010-01-29 2010-01-29 Object detection device and face detection device

Publications (1)

Publication Number Publication Date
WO2011092848A1 true WO2011092848A1 (en) 2011-08-04

Family

ID=44318856

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2010/051257 WO2011092848A1 (en) 2010-01-29 2010-01-29 Object detection device and face detection device

Country Status (2)

Country Link
JP (1) JPWO2011092848A1 (en)
WO (1) WO2011092848A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018147262A (en) * 2017-03-06 2018-09-20 国立大学法人豊橋技術科学大学 Image feature quantity and three-dimensional shape retrieval system using the same

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002051316A (en) * 2000-05-22 2002-02-15 Matsushita Electric Ind Co Ltd Image communication terminal
JP2004206665A (en) * 2002-10-30 2004-07-22 Japan Science & Technology Agency Human face detecting method and human face detecting device
JP2006018706A (en) * 2004-07-05 2006-01-19 Nippon Telegr & Teleph Corp <Ntt> Subject image discriminator setting device, setting method and setting program thereof, subject image identification apparatus, and identification method and identification program thereof
JP2006065613A (en) * 2004-08-27 2006-03-09 Fuji Photo Film Co Ltd Apparatus and method for partitioning specific image area and program for allowing computer to execute specific image area partitioning processing

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002051316A (en) * 2000-05-22 2002-02-15 Matsushita Electric Ind Co Ltd Image communication terminal
JP2004206665A (en) * 2002-10-30 2004-07-22 Japan Science & Technology Agency Human face detecting method and human face detecting device
JP2006018706A (en) * 2004-07-05 2006-01-19 Nippon Telegr & Teleph Corp <Ntt> Subject image discriminator setting device, setting method and setting program thereof, subject image identification apparatus, and identification method and identification program thereof
JP2006065613A (en) * 2004-08-27 2006-03-09 Fuji Photo Film Co Ltd Apparatus and method for partitioning specific image area and program for allowing computer to execute specific image area partitioning processing

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018147262A (en) * 2017-03-06 2018-09-20 国立大学法人豊橋技術科学大学 Image feature quantity and three-dimensional shape retrieval system using the same

Also Published As

Publication number Publication date
JPWO2011092848A1 (en) 2013-05-30

Similar Documents

Publication Publication Date Title
CN106709895B (en) Image generation method and apparatus
EP3021575B1 (en) Image processing device and image processing method
US9158985B2 (en) Method and apparatus for processing image of scene of interest
JP4619927B2 (en) Face detection method, apparatus and program
JP6525545B2 (en) INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND COMPUTER PROGRAM
JP4479478B2 (en) Pattern recognition method and apparatus
JP5701005B2 (en) Object detection apparatus, object detection method, surveillance camera system, and program
US8958609B2 (en) Method and device for computing degree of similarly between data sets
US8155396B2 (en) Method, apparatus, and program for detecting faces
US20120093402A1 (en) Image processing
US20160004935A1 (en) Image processing apparatus and image processing method which learn dictionary
WO2014074959A1 (en) Real-time face detection using pixel pairs
JP5515871B2 (en) Object recognition apparatus and program
JP2006270301A (en) Scene change detecting apparatus and scene change detection program
WO2011092848A1 (en) Object detection device and face detection device
WO2009096208A1 (en) Object recognition system, object recognition method, and object recognition program
JP6399122B2 (en) Face detection apparatus and control method thereof
JP4455980B2 (en) Moving image processing method, moving image processing apparatus, moving image processing program, and recording medium recording the program
JP2007140695A (en) Suspicious face detection system, suspicious face detection method and suspicious face detection program
JP5283267B2 (en) Content identification method and apparatus
JP5599228B2 (en) Busy detection system and busy detection program
JP2011159097A (en) Method and device for detecting object
GB2467643A (en) Improved detection of people in real world videos and images.
JP2021111228A (en) Learning device, learning method, and program
JP5786838B2 (en) Image region dividing apparatus, method, and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10844609

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2011551643

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10844609

Country of ref document: EP

Kind code of ref document: A1