WO2020261820A1 - Image processing device, monitoring device, control system, image processing method, and program - Google Patents

Image processing device, monitoring device, control system, image processing method, and program Download PDF

Info

Publication number
WO2020261820A1
WO2020261820A1 PCT/JP2020/019848 JP2020019848W WO2020261820A1 WO 2020261820 A1 WO2020261820 A1 WO 2020261820A1 JP 2020019848 W JP2020019848 W JP 2020019848W WO 2020261820 A1 WO2020261820 A1 WO 2020261820A1
Authority
WO
WIPO (PCT)
Prior art keywords
face
specific individual
image processing
feature amount
unit
Prior art date
Application number
PCT/JP2020/019848
Other languages
French (fr)
Japanese (ja)
Inventor
相澤 知禎
Original Assignee
オムロン株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by オムロン株式会社 filed Critical オムロン株式会社
Publication of WO2020261820A1 publication Critical patent/WO2020261820A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems

Definitions

  • the present invention relates to an image processing device, a monitoring device, a control system, an image processing method, and a program.
  • Patent Document 1 discloses a robot device used as a service providing device that can switch to an appropriate service according to the situation of a target (person) to which the service is provided.
  • the robot device is equipped with a first camera, a second camera, and an information processing device including a CPU, and the CPU includes a face detection unit, an attribute determination unit, a person detection unit, a person position calculation unit, and an information processing unit. It is equipped with a movement vector detector and the like.
  • the robot device when the service is provided to a group of people who have a relationship such as communicating with each other, the first service of providing information based on close communication is performed. To determine. On the other hand, when the service is provided to a group of people whose relationship such as communication with each other is unknown, the second service provides information unilaterally without exchanging information. Decide to do. As a result, it is possible to provide appropriate services according to the situation of the service provision target.
  • the face detection unit is configured to detect a person's face using the first camera, and a known technique can be used for the face detection.
  • a known technique can be used for the face detection.
  • a part of the facial organs such as eyes, nose, and mouth is missing or greatly deformed due to injury, a large mole, wart, or body decoration such as tattoo is applied to the face.
  • Such specific individuals in other words, age difference, gender, and person, such as when the facial organs are displaced from their average position due to treatment or a disease such as a hereditary disease.
  • the present invention has been made in view of the above problems, and provides an image processing device, a monitoring device, a control system, an image processing method, and a program capable of improving the accuracy of face sensing for a specific individual as described above.
  • the purpose is.
  • the image processing apparatus (1) according to the present disclosure in order to achieve the above object is an image processing apparatus that processes an image input from an imaging unit.
  • a facial feature storage unit that stores the facial features of a specific individual and the normal facial features
  • a face detection unit that detects a face region while extracting a feature amount for detecting a face from the image, and a face detection unit.
  • a specific individual determination unit for determining whether or not the face in the face region is the face of the specific individual.
  • the specific individual determination unit determines that the face is the face of the specific individual
  • the first face image processing unit that performs face image processing for the specific individual
  • the first face image processing unit When it is determined by the specific individual determination unit that it is not the face of the specific individual, it is characterized by including a second face image processing unit that performs normal face image processing.
  • the face feature amount of the specific individual and the normal face feature amount are used as the learned face feature amount in the face feature amount storage unit.
  • the facial feature amount used when the person is a person other than the above) is stored, and the feature amount of the face region detected by the face detection unit and the facial feature amount of the specific individual are stored by the specific individual determination unit. It is used to determine whether or not the face in the face region is the face of the specific individual. By using the facial feature amount of the specific individual, it is possible to accurately determine whether or not the face is the face of the specific individual.
  • the face image processing of the specific individual can be accurately performed by the first face image processing unit.
  • the second face image processing unit performs the normal face image processing. It can be carried out with high accuracy. Therefore, both the specific individual and the ordinary person other than the specific individual can accurately perform the sensing of each face.
  • the specific individual determination unit correlates the feature amount extracted from the face region with the face feature amount of the specific individual. It is characterized in that an index indicating the above is calculated, and based on the calculated index, it is determined whether or not the face in the face region is the face of the specific individual.
  • an index showing the correlation between the feature amount extracted from the face area and the face feature amount of the specific individual is calculated, and the face area is based on the calculated index. It is determined whether or not the face of the specific individual is the face of the specific individual. Thereby, it is possible to efficiently determine whether or not the face in the face region is the face of the specific individual based on the index.
  • the index may be an index value indicating that the larger the value is, the higher the relationship is, for example, a correlation coefficient, the reciprocal of the square error, or other extraction from the face region. It may be an index value or the like indicating the degree of similarity of the relationship between the said feature amount and the face feature amount of the specific individual.
  • the face in the face region is the face of the specific individual.
  • the index is equal to or less than the predetermined threshold value, it is determined that the face in the face region is not the face of the specific individual.
  • the index is larger than a predetermined threshold value, it is determined that the face in the face region is the face of the specific individual, and when the index is equal to or less than the predetermined threshold value, the face is described. It is determined that the face in the face area is not the face of the specific individual.
  • the processing efficiency of the determination can be further improved by the process of comparing the index with the predetermined threshold value.
  • the specific individual determination unit is based on the result of determination for one frame of the image. It is characterized in that it is determined whether or not the face in the face region is the face of the specific individual. According to the image processing device (4), it is determined whether or not the face in the face region is the face of the specific individual based on the result of the determination for one frame of the image, so that the determination is high speed. Can be achieved.
  • the specific individual determination unit is based on the result of determination for a plurality of frames of the image. It is characterized in that it is determined whether or not the face in the face region is the face of the specific individual. According to the image processing device (5), it is determined whether or not the face in the face region is the face of the specific individual based on the result of the determination for a plurality of frames of the image, so that the accuracy of the determination is determined. Can be enhanced.
  • the image processing device (6) is any of the above image processing devices (1) to (5), and the face image processing includes face detection processing, face orientation estimation processing, and eye direction estimation processing. And at least one of the eye opening / closing detection processing is included.
  • the face image processing includes at least one of face detection processing, face orientation estimation processing, line-of-sight direction estimation processing, and eye opening / closing detection processing. It is possible to accurately perform processing for estimating and detecting various facial behaviors of the specific individual or a (normal) person other than the specific individual.
  • the monitoring device (1) is used for any of the above image processing devices (1) to (6), an image pickup unit for capturing an image to be input to the image processing device, and image processing by the image processing device. It is characterized by having an output unit that outputs based information. According to the monitoring device (1), not only the face of the normal person but also the face of the specific individual can be accurately monitored, and information based on the image processing can be output from the output unit. Therefore, it is possible to easily construct a monitoring system or the like that uses the information.
  • control system (1) is communicably connected to the monitoring device (1) and executes a predetermined process based on the information output from the monitoring device. It is characterized by having the above-mentioned control device. According to the control system (1), it is possible to execute a predetermined process by one or more of the control devices based on the information output from the monitoring device. Therefore, it is possible to construct a system that can utilize not only the monitoring result of the normal person but also the monitoring result of the specific individual.
  • control system (2) is the control system (1) described above.
  • the monitoring device is a device for monitoring the driver of the vehicle.
  • the control device is characterized by including an electronic control unit mounted on the vehicle. According to the control system (2), even when the driver of the vehicle is the specific individual, the face of the specific individual can be accurately monitored, and the electronic device is based on the monitoring result. It is possible to make the control unit appropriately execute a predetermined control. This makes it possible to construct a highly safe in-vehicle system that allows even the specific individual to drive with peace of mind.
  • the image processing method is an image processing method for processing an image input from an imaging unit.
  • a specific individual determination step for determining whether or not is the face of the specific individual and When the face of the specific individual is determined by the specific individual determination step, the first face image processing step of performing the face image processing for the specific individual and the first face image processing step When it is determined by the specific individual determination step that the face is not the face of the specific individual, it is characterized by including a second face image processing step of performing normal face image processing.
  • the face in the face region is formed by using the feature amount of the face region detected in the face detection step and the face feature amount of the specific individual in the specific individual determination step. Whether or not it is the face of the specific individual is determined.
  • the facial feature amount of the specific individual it is possible to accurately determine whether or not the face is the face of the specific individual.
  • the face image processing of the specific individual can be accurately performed by the first face image processing step.
  • the normal face image processing is performed with high accuracy by the second face image processing step. Can be done. Therefore, both the specific individual and a normal person other than the specific individual can accurately sense each face.
  • the program according to the present disclosure is a program for causing at least one or more computers to process an image input from an imaging unit.
  • a face detection step of detecting a face region while extracting facial features from the image, and The face of the face region is used by using the feature amount of the face region detected by the face detection step and the learned face feature amount of the specific individual who has been trained to detect the face of the specific individual.
  • a specific individual determination step for determining whether or not is the face of the specific individual and When the face of the specific individual is determined by the specific individual determination step, the first face image processing step of performing the face image processing for the specific individual and the first face image processing step When it is determined by the specific individual determination step that the face is not the face of the specific individual, the second face image processing step of performing normal face image processing is executed.
  • the above program determines whether or not the face in the face region is the face of the specific individual by using the feature amount of the face region and the face feature amount of the specific individual on at least one computer. It is possible to determine whether or not the face is the specific individual's face with high accuracy. Further, when it is determined that the face of the specific individual is the face, the face image processing of the specific individual can be performed with high accuracy. On the other hand, when it is determined that the face is not the specific individual's face, in other words, it is a normal face that is not the specific individual, the normal face image processing can be performed with high accuracy. Therefore, it is possible to construct a device or system capable of accurately sensing each face regardless of whether the specific individual or an ordinary person other than the specific individual.
  • the above program may be a program stored in a storage medium, a program that can be transferred via a communication network, or a program that is executed via a communication network. ..
  • the image processing apparatus can be widely applied to, for example, an apparatus or system for monitoring an object such as a person using a camera.
  • the image processing device operates or monitors, for example, various facilities such as machines and devices in a factory, in addition to devices and systems for monitoring drivers (operators) of various moving objects such as vehicles. It can also be applied to devices and systems that monitor people who perform predetermined work.
  • FIG. 1 is a schematic view showing an example of an in-vehicle system including the driver monitoring device according to the embodiment.
  • the in-vehicle system 1 includes a driver monitoring device 10 that monitors the state of the driver 3 of the vehicle 2 (for example, facial behavior), and one or more ECUs (Electronic Control Units) that control the running, steering, or braking of the vehicle 2. ) 40, and one or more sensors 41 for detecting the state of each part of the vehicle, the state around the vehicle, and the like are included, and these are connected via the communication bus 43.
  • a driver monitoring device 10 that monitors the state of the driver 3 of the vehicle 2 (for example, facial behavior), and one or more ECUs (Electronic Control Units) that control the running, steering, or braking of the vehicle 2.
  • ECUs Electronic Control Units
  • the in-vehicle system 1 is configured as, for example, an in-vehicle network system that communicates according to a CAN (Controller Area Network) protocol.
  • CAN Controller Area Network
  • the driver monitoring device 10 is an example of the "monitoring device" of the present invention, and the in-vehicle system 1 is an example of the "control system" of the present invention.
  • the driver monitoring device 10 transmits information based on image processing by the camera 11 for capturing the face of the driver 3, the image processing unit 12 that processes the image input from the camera 11, and the image processing unit 12, and the communication bus 43. It is configured to include a communication unit 16 that performs processing such as output to a predetermined ECU 40 via the above.
  • the image processing unit 12 is an example of the "image processing device” of the present invention.
  • the camera 11 is an example of the "imaging unit” of the present invention.
  • the driver monitoring device 10 detects the face of the driver 3 from the image captured by the camera 11, and detects the behavior of the face such as the direction of the face of the detected driver 3, the direction of the line of sight, or the open / closed state of the eyes.
  • the driver monitoring device 10 may determine the state of the driver 3, such as forward gaze, inattentiveness, dozing, backward facing, and prone, based on the detection results of these facial behaviors. Further, the driver monitoring device 10 outputs a signal based on the state determination of the driver 3 to the ECU 40, and the ECU 40 performs attention and warning processing to the driver 3 or operation control of the vehicle 2 (for example, deceleration) based on the signal. Control, guidance control to the road shoulder, etc.) may be executed.
  • the driver monitoring device 10 One of the purposes of the driver monitoring device 10 is to improve the accuracy of face sensing for a specific individual.
  • the driver 3 of the vehicle 2 has a part of facial organs such as eyes, nose, and mouth missing or greatly deformed due to, for example, an injury, or a large mole or wart on the face, or Accuracy of detecting the face from the image captured by the camera when the facial organs are displaced from the average position due to body decoration such as tattoo or a disease such as a hereditary disease.
  • body decoration such as tattoo or a disease such as a hereditary disease.
  • the driver monitoring device 10 is a general person who is common regardless of a specific individual, in other words, a difference in age, gender, race, etc. (individual difference). In order to improve the accuracy of face detection for a specific individual who has features different from the face features of (ordinary person), the following configuration was adopted.
  • the image processing unit 12 as the learned facial features that have been learned to detect the face from the image, the facial features of a specific individual and the normal facial features (in other words, a person other than the specific individual).
  • the amount of facial features used in the case of) is stored.
  • the image processing unit 12 performs face detection processing for detecting a face region while extracting a feature amount for detecting a face from an input image of the camera 11. Then, the image processing unit 12 determines whether or not the face in the face region is the face of the specific individual by using the detected feature amount of the face region and the face feature amount of the specific individual. Performs specific individual judgment processing.
  • a correlation coefficient is calculated and calculated as an index showing the relationship between the feature amount extracted from the face region and the face feature amount of the specific individual, for example, an index showing the correlation. Based on the number of relationships, it may be determined whether or not the face in the face region is the face of the specific individual. For example, when the correlation coefficient is larger than a predetermined threshold value, it is determined that the face in the face region is the face of the specific individual, and when the correlation coefficient is equal to or less than the predetermined threshold value, the face in the face region is It may be determined that it is not the face of the specific individual. In the specific individual determination process, an index other than the correlation coefficient may be used.
  • the specific individual determination process it may be determined whether or not the face in the face region is the face of the specific individual based on the result of determination for one frame of the input image from the camera 11. Based on the result of determination for a plurality of frames of the input image from the camera 11, it may be determined whether or not the face in the face region is the face of the specific individual.
  • the learned facial feature amount of the specific individual is stored in advance in the image processing unit 12, and the face feature amount of the specific individual is used to obtain the face of the specific individual. It is possible to accurately determine whether or not it is.
  • the image processing unit 12 executes the face image process for the specific individual, so that the face image process of the specific individual is accurately performed. It becomes possible to do.
  • the image processing unit 12 executes the normal face image processing. Therefore, the normal face image processing can be performed with high accuracy. Therefore, whether the driver 3 is a specific individual or an ordinary person other than the specific individual, it is possible to accurately perform sensing of each face.
  • FIG. 2 is a block diagram showing an example of the hardware configuration of the in-vehicle system 1 including the driver monitoring device 10 according to the embodiment.
  • the in-vehicle system 1 includes a driver monitoring device 10, 1 or more ECUs 40 for monitoring the state of the driver 3 of the vehicle 2, and 1 or more sensors 41, which are connected via a communication bus 43. Further, one or more actuators 42 are connected to the ECU 40.
  • the driver monitoring device 10 includes a camera 11, an image processing unit 12 that processes an image input from the camera 11, and a communication unit 16 for exchanging data and signals with an external ECU 40 and the like. There is.
  • the camera 11 is a device that captures an image including the face of the driver 3 seated in the driver's seat.
  • the image sensor unit may include an image sensor such as a CCD (Charge Coupled Device) and a CMOS (Complementary Metal Oxide Semiconductor), a filter, a microlens, and the like.
  • the image pickup device unit may be an element capable of forming an image pickup image by receiving light in a visible region, or an element capable of forming an image pickup image by receiving light in a near infrared region.
  • the light irradiation unit is configured to include a light emitting element such as an LED (Light Emitting Diode), and may include a near infrared LED or the like so that the driver's face can be imaged day or night.
  • the camera 11 captures an image at a predetermined frame rate (for example, several tens of frames per second), and the data of the captured image is input to the image processing unit 12.
  • the camera 11 may be an external type as well as an integrated type.
  • the image processing unit 12 is configured as an image processing device including one or more CPU (Central Processing Unit) 13, ROM (Read Only Memory) 14, and RAM (Random Access Memory) 15.
  • the ROM 14 includes a program storage unit 141 and a facial feature amount storage unit 142
  • the RAM 15 includes an image memory 151 for storing an input image from the camera 11.
  • the driver monitoring device 10 may be provided with another storage unit, and the storage unit may be used as the program storage unit 141, the facial feature amount storage unit 142, and the image memory 151.
  • the other storage unit may be a semiconductor memory or a storage medium that can be read by a disk drive or the like.
  • the CPU 13 is an example of a hardware processor, and by reading, interpreting, and executing data such as a program stored in the program storage unit 141 of the ROM 14 and the face feature amount stored in the face feature amount storage unit 142. , Processing of the image input from the camera 11, for example, face image processing such as face detection processing is performed. Further, the CPU 13 performs a process of outputting the result (for example, processing data, determination signal, control signal, etc.) obtained by the face image processing to the ECU 40 or the like via the communication unit 16.
  • the face feature amount storage unit 142 as the learned face feature amount that has been learned (for example, machine learning) to detect the face from the image, the face feature amount 142a of the specific individual shown in FIG.
  • the facial feature amount 142b is stored.
  • various feature quantities effective for detecting a face from an image can be used. For example, a feature amount (Haar-like feature amount) focusing on the difference in brightness (difference in average brightness between two rectangular areas of various sizes) in a local area of the face may be used.
  • a feature amount (LBP (Local Binary Pattern) feature amount) focusing on a combination of brightness distributions in the local region of the face may be used, or the distribution of the brightness in the local region of the face in the gradient direction may be used.
  • Features HOG (Histogram of Oriented Gradients) features focusing on the combination may be used.
  • the face feature amount stored in the face feature amount storage unit 142 is extracted as an effective feature amount for face detection by using, for example, various machine learning methods.
  • Machine learning is a process of finding a pattern inherent in data (learning data) by a computer.
  • AdaBoost may be used as an example of a statistical learning method.
  • AdaBoost selects a large number of discriminators (weak discriminators) with low discriminating ability, selects a weak discriminator with a small error rate from these many weak discriminators, adjusts parameters such as weights, and has a hierarchical structure. It is a learning algorithm that can construct a strong discriminator by setting.
  • the discriminator may be referred to as a discriminator, a classifier, or a learner.
  • the strong discriminator is configured to discriminate one feature amount effective for face detection by one weak discriminator, and a large number of weak discriminators and their combinations are selected by AdaBoost, and these are used hierarchically.
  • the structure may be constructed.
  • one weak discriminator may output information such as 1 for a face and 0 for a non-face.
  • a learning method called Real AdaBoost which can output a real number from 0 to 1 instead of 0 or 1, may be used.
  • a neural network having an input layer, an intermediate layer, and an output layer may be used.
  • a large number of face images captured under various conditions and a large number of non-face images (non-face images) are given as training data to a learning device equipped with such a learning algorithm, learning is repeated, weighting, etc.
  • a strong discriminator having a hierarchical structure capable of detecting a face with high accuracy.
  • one or more feature amounts used in the weak discriminators of each layer constituting such a strong discriminator can be used as the learned facial feature amounts.
  • the face feature amount 142a of a specific individual individually captures a face image of the specific individual at a predetermined place under various conditions (conditions such as various face orientations, line-of-sight directions, or eye open / closed states). Then, these a large number of captured images are input to the learning device as teacher data, and are parameters that indicate the facial features of a specific individual adjusted by the learning process.
  • the facial feature amount 142a of the specific individual may be, for example, a combination pattern of the difference in brightness of the local region of the face obtained by the learning process.
  • the facial feature amount 142a of a specific individual stored in the facial feature amount storage unit 142 may be only the facial feature amount of one specific individual, or can be used when a plurality of specific individuals drive the vehicle 2. , The facial features of a plurality of specific individuals may be stored.
  • the normal facial feature amount 142b is the above-mentioned learning device using images of a normal human face image captured under various conditions (conditions such as various face orientations, line-of-sight directions, or eye open / closed states) as teacher data. It is a parameter indicating the characteristics of a normal human face, which is input to and adjusted by the learning process.
  • the normal facial feature amount 142b may be, for example, a combination pattern of light and dark differences in a local region of the face obtained by a learning process. Further, as the normal facial feature amount 142b, the information registered in the predetermined facial feature amount database may be used.
  • the learned facial feature amount stored in the facial feature amount storage unit 142 is fetched from a server on the cloud or the like via a communication network such as the Internet or a mobile phone network and stored in the facial feature amount storage unit 142. It may be configured as such.
  • the ECU 40 is composed of a computer device including one or more processors, a memory, a communication module, and the like. Then, the processor mounted on the ECU 40 reads, interprets, and executes the program stored in the memory, so that predetermined control for the actuator 42 and the like is executed.
  • the ECU 40 includes, for example, at least one of a traveling system ECU, a driving support system ECU, a body system ECU, and an information system ECU.
  • the traveling system ECU includes, for example, a drive system ECU, a chassis system ECU, and the like.
  • the drive system ECU includes a control unit related to a "running" function such as engine control, motor control, fuel cell control, EV (Electric Vehicle) control, or transmission control.
  • the chassis-based ECU includes a control unit related to a "stop, turn” function such as brake control or steering control.
  • the driving support system ECU has, for example, an automatic braking support function, a lane keeping support function (also referred to as LKA / Lane Keep Assist), a constant speed driving / inter-vehicle distance support function (also referred to as ACC / Adaptive Cruise Control), and a forward collision warning function.
  • Lane departure warning function a blind spot monitoring function, traffic sign recognition function, etc., functions that automatically improve safety or realize comfortable driving by linking with driving ECUs (driving support function or automatic driving function) It may be configured to include at least one control unit with respect to.
  • the driving support system ECU includes, for example, Level 1 (driver assistance), Level 2 (partially automatic driving), and Level 3 (conditional automatic driving) at the automatic driving level presented by the American Society of Automotive Engineers of Japan (SAE). ) May be equipped with at least one of the functions. Further, the functions of level 4 (highly automatic driving) or level 5 (fully automatic driving) of the automatic driving level may be equipped, and only the functions of level 1 and 2 or only level 2 and 3 are equipped. May be good. Further, the in-vehicle system 1 may be configured as an automatic driving system.
  • the body system ECU may be configured to include at least one control unit related to the function of the vehicle body such as a door lock, a smart key, a power window, an air conditioner, a light, an instrument panel, or a winker.
  • the information system ECU may be configured to include, for example, an infotainment device, a telematics device, or an ITS (Intelligent Transport Systems) related device.
  • the infotainment device may include, for example, an HMI (Human Machine Interface) device that functions as a user interface, a car navigation device, an audio device, and the like.
  • the telematics device may include a communication unit or the like for communicating with the outside.
  • the ITS-related device may include an ETC (Electronic Toll Collection System), a communication unit for performing road-to-vehicle communication with a roadside machine such as an ITS spot, or vehicle-to-vehicle communication.
  • ETC Electronic Toll Collection System
  • the sensor 41 may include various in-vehicle sensors that acquire sensing data necessary for controlling the operation of the actuator 42 by the ECU 40.
  • vehicle speed sensors shift position sensors, accelerator opening sensors, brake pedal sensors, steering sensors, etc.
  • peripheral monitoring of external imaging cameras millimeter-wave radar (Radar), riders (LIDER), ultrasonic sensors, etc.
  • a sensor or the like may be included.
  • the actuator 42 is a device that executes operations related to traveling, steering, braking, etc. of the vehicle 2 based on a control signal from the ECU 40, and includes, for example, an engine, a motor, a transmission, a hydraulic cylinder, an electric cylinder, and the like.
  • FIG. 3 is a block diagram showing a functional configuration example of the image processing unit 12 of the driver monitoring device 10 according to the embodiment.
  • the image processing unit 12 includes an image input unit 21, a face detection unit 22, a specific individual determination unit 25, a first face image processing unit 26, a second face image processing unit 30, an output unit 34, and a face feature amount storage unit 142. It is configured to include.
  • the image input unit 21 performs a process of capturing an image including the face of the driver 3 captured by the camera 11.
  • the face detection unit 22 is configured to include a face detection unit 23 of a specific individual and a normal face detection unit 24, and performs a process of detecting a face region while extracting a feature amount for detecting a face from an input image.
  • the face detection unit 23 of the specific individual uses the face feature amount 142a of the specific individual read from the face feature amount storage unit 142 to perform a process of detecting the face region from the input image.
  • the normal face detection unit 24 uses the normal face feature amount 142b read from the face feature amount storage unit 142 to perform a process of detecting a face region from an input image.
  • the method of detecting the face area from the image is not particularly limited, but a method of detecting the face area at high speed and with high accuracy is adopted.
  • the face detection unit 22 extracts, for example, a feature amount for detecting a face in each search area while scanning a predetermined search area (search window) for the input image.
  • the face detection unit 22 may extract, for example, the difference in brightness (luminance difference) of a local region of the face, the edge strength, or the relationship between these local regions as a feature amount.
  • the face detection unit 22 uses the feature amount extracted from the search area, the normal face feature amount 142b read from the face feature amount storage unit 142, or the face feature amount 142a of a specific individual, and has a hierarchical structure (A detector (hierarchical structure that captures the details of the face from the hierarchy that roughly captures the face) determines whether the face is face or non-face, and performs processing to detect the face area from the image.
  • a detector hierarchical structure that captures the details of the face from the hierarchy that roughly captures the face
  • the specific individual determination unit 25 uses the feature amount of the face area detected by the face detection unit 22 and the face feature amount 142a of the specific individual read from the face feature amount storage unit 142 to detect the face in the face area. Performs a process of determining whether or not is the face of a specific individual.
  • the specific individual determination unit 25 calculates a correlation coefficient as an index showing the relationship between the feature amount extracted from the face region and the face feature amount 142a of the specific individual, for example, an index showing the correlation, and the calculated correlation coefficient. It may be determined whether or not the face in the face region is the face of a specific individual based on. For example, the correlation of feature quantities such as Haar-like features (luminance difference) of one or more local regions in the face region may be obtained.
  • the correlation coefficient is larger than a predetermined threshold value, it is determined that the face in the detected face area is the face of a specific individual, and when the correlation coefficient is equal to or less than the predetermined threshold value, the face in the detected face area is a specific individual. It may be determined that it is not the face of.
  • the specific individual determination unit 25 includes a learned learning device that has been machine-learned using the facial feature amount 142a of the specific individual as a parameter in order to determine whether or not the face is the face of the specific individual.
  • a learned learning device that has been machine-learned using the facial feature amount 142a of the specific individual as a parameter in order to determine whether or not the face is the face of the specific individual.
  • the determination information may be acquired from the learning device.
  • the learned learner may be configured to include a non-linear discriminator, or may be configured to include a linear discriminator.
  • the trained learner may be configured to include a support vector machine or may be configured to include a neural network.
  • the specific individual determination unit 25 may determine whether or not the face in the detected face region is the face of the specific individual based on the result of determination for one frame of the input image from the camera 11. Based on the result of determination for a plurality of frames of the input image from the camera 11, it may be determined whether or not the face in the detected face region is the face of a specific individual.
  • the first face image processing unit 26 performs face image processing for the specific individual.
  • the first face image processing unit 26 includes a face orientation estimation unit 27 of a specific individual, an eye opening / closing detection unit 28 of the specific individual, and a line-of-sight direction estimation unit 29 of the specific individual, but is still different.
  • the first face image processing unit 26 may perform any processing of the face image processing for the specific individual by using the face feature amount 142a of the specific individual.
  • the face feature amount storage unit 142 stores the learned feature amount that has been machine-learned to perform the face image processing for the specific individual, and uses the learned feature amount for the specific individual. You may perform any processing of the face image processing of.
  • the face orientation estimation unit 27 of the specific individual performs a process of estimating the face orientation of the specific individual.
  • the face orientation estimation unit 27 of the specific individual detects, for example, the position and shape of facial organs such as eyes, nose, mouth, and eyebrows from the face region detected by the face detection unit 23 of the specific individual, and the detected facial organs. Performs processing to estimate the orientation of the face based on the position and shape.
  • the method for detecting the facial organs from the facial region in the image is not particularly limited, but it is preferable to adopt a method capable of detecting the facial organs at high speed and with high accuracy.
  • a method of creating a three-dimensional face shape model, fitting it to a face region on a two-dimensional image, and detecting the position and shape of each organ of the face can be adopted.
  • a technique for fitting a three-dimensional face shape model to a human face in an image for example, the technique described in Japanese Patent Application Laid-Open No. 2007-249280 can be applied, but the technique is not limited thereto.
  • the face orientation estimation unit 27 of the specific individual can use the estimation data of the face orientation of the specific individual, for example, the pitch angle of vertical rotation (around the X axis) included in the parameters of the three-dimensional face shape model.
  • the yaw angle of the left-right rotation (around the Y axis) and the roll angle of the entire rotation (around the Z axis) may be output.
  • the eye opening / closing detection unit 28 of the specific individual performs a process of detecting the opening / closing state of the eyes of the specific individual.
  • the eye opening / closing detection unit 28 of the specific individual is based on the position and shape of the facial organs obtained by the face orientation estimation unit 27 of the specific individual, particularly the position and shape of the feature points (eyelids, pupils) of the eyes. Detects the open / closed state, for example, whether the eyes are open or closed. For the open / closed state of the eye, for example, the feature amount of the image of the eye (the position of the eyelid, the shape of the pupil (black eye), the area size of the white eye part and the black eye part, etc.) in various open / closed states of the eye is previously learned. It may be detected by learning using the data and evaluating the degree of similarity with the learned feature data.
  • the line-of-sight direction estimation unit 29 of the specific individual performs a process of estimating the line-of-sight direction of the specific individual.
  • the line-of-sight direction estimation unit 29 of a specific individual is based on, for example, the orientation of the face of the driver 3 and the position and shape of the facial organs of the driver 3, particularly the position and shape of the feature points of the eyes (outer corners of eyes, inner corners of eyes, pupils).
  • Estimate the direction of the line of sight is the direction in which the driver 3 is looking, and is determined by, for example, a combination of the direction of the face and the direction of the eyes.
  • the direction of the line of sight is, for example, the feature amount of the image of the eye in various combinations of face orientation and eye orientation (relative position of outer corner, inner corner of eye, pupil, relative position of white eye portion and black eye portion, shading, etc. (Texture, etc.) may be detected by learning in advance using a learning device and evaluating the degree of similarity with the learned feature amount data.
  • the line-of-sight direction estimation unit 29 of the specific individual estimates the size and center position of the eyeball from the size and orientation of the face, the position of the eyes, etc., using the fitting result of the three-dimensional face shape model, and the pupil. The position of the eyeball may be detected, and the vector connecting the center of the eyeball and the center of the pupil may be detected as the line-of-sight direction.
  • the second face image processing unit 30 When the specific individual determination unit 25 determines that the face is not the face of a specific individual, the second face image processing unit 30 performs normal face image processing.
  • the second face image processing unit 30 includes a normal face orientation estimation unit 31, a normal eye opening / closing detection unit 32, and a normal line-of-sight direction estimation unit 33, but has yet another face behavior. It may include a configuration for estimating or detecting.
  • the second face image processing unit 30 may perform any processing of the normal face image processing using the normal face feature amount 142b.
  • the face feature amount storage unit 142 stores the learned feature amount obtained by machine learning for performing the normal face image processing, and the learned feature amount is used for the normal face image processing. Any of the above processes may be performed.
  • the processing performed by the normal face orientation estimation unit 31, the normal eye opening / closing detection unit 32, and the normal line-of-sight direction estimation unit 33 includes the face orientation estimation unit 27 of the specific individual and the eye opening / closing detection of the specific individual. Since it is basically the same as the unit 28 and the line-of-sight direction estimation unit 29 of a specific individual, the description thereof will be omitted here.
  • the output unit 34 performs a process of outputting information based on the image processing by the image processing unit 12 to the ECU 40 or the like.
  • the information based on the image processing may be, for example, information on the behavior of the face such as the direction of the face of the driver 3, the direction of the line of sight, or the open / closed state of the eyes, or the driver 3 determined based on the detection result of the behavior of the face. Information on the state of (for example, forward gaze, inattentiveness, dozing, backward facing, prone, etc.) may be used. Further, the information based on the image processing may be a predetermined control signal (control signal for performing caution or warning processing, control signal for performing operation control of the vehicle 2, etc.) based on the state determination of the driver 3.
  • FIG. 4 is a flowchart showing an example of a processing operation performed by the CPU 13 of the image processing unit 12 in the driver monitoring device 10 according to the embodiment.
  • the camera 11 captures an image of several tens of frames per second, and this processing is performed for each frame or every frame at regular intervals.
  • the CPU 13 operates as an image input unit 21, performs a process of reading an image (an image including the face of the driver 3) captured by the camera 11, and proceeds to step S2.
  • the CPU 13 operates as a normal face detection unit 24, performs normal face detection processing on the input image, and proceeds to step S3.
  • the CPU 13 scans a predetermined search area (search window) for an input image and extracts a feature amount for detecting a face in each search area. Then, the CPU 13 determines whether the face is a face or a non-face by using the feature amount extracted from the search area and the normal face feature amount 142b read from the face feature amount storage unit 142, and detects the face area from the image. Perform the process of
  • step S3 the CPU 13 operates as the face detection unit 23 of the specific individual, performs face detection processing of the specific individual on the input image, and proceeds to step S4.
  • the CPU 13 scans a predetermined search area (search window) for an input image and extracts a feature amount for detecting a face in each search area. Then, the CPU 13 determines whether the face is a face or a non-face by using the feature amount extracted from the search area and the face feature amount 142a of the specific individual read from the face feature amount storage unit 142, and determines the face area from the image. Perform the detection process.
  • the processes of steps S2 and S3 may be performed in parallel in one step or may be performed in combination.
  • step S4 the CPU 13 operates as the specific individual determination unit 25, and uses the feature amount of the face region detected in steps S2 and S3 and the face feature amount 142a of the specific individual read from the face feature amount storage unit 142. Then, a process of determining whether or not the face in the face area is a face of a specific individual is performed, and the process proceeds to step S5.
  • step S5 the CPU 13 determines whether or not the result of the determination process in step S4 is the face of a specific individual, and if it is determined that the result is the face of a specific individual, the process proceeds to step S6.
  • step S6 the CPU 13 operates as a face orientation estimation unit 27 of a specific individual, and for example, detects and detects the position and shape of facial organs such as eyes, nose, mouth, and eyebrows from the face area detected in step S3.
  • the orientation of the face is estimated based on the position and shape of the facial organ, and the process proceeds to step S7.
  • step S7 the CPU 13 operates as an eye opening / closing detection unit 28 for a specific individual, and is based on, for example, the position and shape of the facial organs obtained in step S6, particularly the position and shape of eye feature points (eyelids, pupils).
  • the open / closed state of the eyes for example, whether the eyes are open or closed is detected, and the process proceeds to step S8.
  • step S8 the CPU 13 operates as a line-of-sight direction estimation unit 29 of a specific individual, and for example, the orientation of the face, the position and shape of the facial organs obtained in step S6, particularly the feature points of the eyes (outer corners of eyes, inner corners of eyes, pupils).
  • the direction of the line of sight is estimated based on the position and shape, and then the process is finished.
  • step S5 if the CPU 13 determines that it is not the face of a specific individual, in other words, it is a normal face, the process proceeds to step S9.
  • step S9 the CPU 13 operates as a normal face orientation estimation unit 31, and for example, detects the position and shape of facial organs such as eyes, nose, mouth, and eyebrows from the face area detected in step S2, and detects the face.
  • the orientation of the face is estimated based on the position and shape of the organ, and the process proceeds to step S10.
  • step S10 the CPU 13 operates as a normal eye opening / closing detection unit 32, and for example, based on the position and shape of the facial organs obtained in step S9, particularly the position and shape of the feature points (eyelids, pupils) of the eyes.
  • the open / closed state of the eyes for example, whether the eyes are open or closed is detected, and the process proceeds to step S11.
  • step S11 the CPU 13 operates as a normal line-of-sight direction estimation unit 33, and for example, the orientation of the face and the position and shape of the facial organs obtained in step S9, particularly the positions of the feature points of the eyes (outer corners of eyes, inner corners of eyes, pupils).
  • the direction of the line of sight is estimated based on the shape and shape, and then the process is completed.
  • FIG. 5 is a flowchart showing an example of a specific individual determination processing operation performed by the CPU 13 of the image processing unit 12 in the driver monitoring device 10 according to the embodiment. This processing operation is an example of the specific individual determination processing operation in step S4 shown in FIG. 4, and is an example of the processing operation in the case of determining with one input image (1 frame).
  • step S21 the CPU 13 reads the feature amount extracted from the face area detected by the face detection processes of steps S2 and S3 shown in FIG. 4, and in the next step S22, learns from the face feature amount storage unit 142.
  • the completed facial feature amount 142a of the specific individual is read, and the process proceeds to step S23.
  • step S23 the CPU 13 performs a process of calculating the correlation coefficient between the feature amount extracted from the face area read in step S21 and the face feature amount 142a of the specific individual read in step S22, and processes in step S24. To proceed.
  • step S24 the CPU 13 determines whether or not the calculated correlation coefficient is larger than a predetermined threshold value for determining whether or not the individual is a specific individual, and the correlation coefficient is larger than the predetermined threshold value, in other words. If it is determined that the feature amount extracted from the face area and the face feature amount 142a of the specific individual have a high correlation (in other words, the similarity is high), the process proceeds to step S25. In step S25, the CPU 13 determines that the face detected in the face area is the face of a specific individual, and then ends the process.
  • step S24 the correlation coefficient is equal to or less than a predetermined threshold value, in other words, the correlation between the feature amount extracted from the face region and the face feature amount 142a of the specific individual is low (in other words, the degree of similarity). Is low), the process proceeds to step S26.
  • step S26 the CPU 13 determines that the face is not a specific individual's face, in other words, a normal face, and then ends the process.
  • FIG. 6 is a flowchart showing an example of a specific individual determination processing operation performed by the CPU 13 of the image processing unit 12 in the driver monitoring device 10 according to the embodiment.
  • This processing operation is another example of the specific individual determination processing operation in step S4 shown in FIG. 4, and is a processing operation example in the case of determining with a plurality of input images (multiple frames).
  • step S31 the CPU 13 sets the counter (numSp) of the image in which the face of a specific individual is detected to 0, and in step S32, the counter (i) of the input image to be determined is set to 0 and step S33. Proceed to processing.
  • step S33 the CPU 13 performs a specific individual determination process (for example, the processes of steps S21 to S26 shown in FIG. 5) for one input image (1 frame), and proceeds to step S34.
  • step S34 the CPU 13 determines whether or not the result of the determination process in step S33 is the face of a specific individual, and if it is determined that the result is the face of a specific individual, the process proceeds to step S35.
  • step S35 the CPU 13 adds 1 to the counter (numSp) of the face image of the specific individual and proceeds with the process in step S36, and in step S36, adds 1 to the counter (i) of the input image and steps.
  • the process proceeds to S37.
  • step S34 if the CPU 13 determines that the face is not the face of a specific individual, the process proceeds to step S36, and in step S36, 1 is added to the counter (i) of the input image and the process proceeds to step S37. .. That is, in this case, 1 is not added to the counter (numSp) of the face image of the specific individual.
  • step S37 the CPU 13 determines whether or not the input image counter (i) is less than the predetermined number of images N, and the counter (i) is less than N (in other words, the predetermined number of images (N). If it is determined that the determination of (sheets) has not been completed), the process returns to step S33, and the determination process of the specific individual in the next input image is repeated. On the other hand, in step S37, if the CPU 13 determines that the counter (i) of the input image is not less than N (in other words, the determination of the predetermined number of images (N) has been completed), the processing is performed in step S38. Proceed.
  • step S38 the CPU 13 determines whether or not the counter (numSp) of the face image of the specific individual is larger than a predetermined threshold value for determining the specific individual.
  • the predetermined threshold value may be set to N / 2 or a value larger than N / 2 when the number of input images used for determination is N (N frames), for example. ..
  • step S38 if the CPU 13 determines that the face image counter (numSp) of the specific individual is larger than the predetermined threshold value, the process proceeds to step S39, and in step S39, it is determined that the face of the specific individual is the face of the specific individual. After finishing the determination process of the specific individual, the face image processing for the specific individual is performed.
  • step S38 if the CPU 13 determines that the counter (numSp) of the specific individual face image is equal to or less than a predetermined threshold value, the process proceeds to step S40, and in step S40, the face is not the face of the specific individual (in other words, For example, it is a normal face), the judgment process of a specific individual is completed, and then the normal face image processing is performed.
  • the face feature amount 142a of a specific individual and the normal face feature amount 142b are stored as the learned face feature amount in the face feature amount storage unit 142.
  • the specific individual determination unit 25 determines whether or not the face in the face region is the face of a specific individual by using the feature amount of the face region detected by the face detection unit 22 and the face feature amount 142a of the specific individual. Will be done. Therefore, by using the face feature amount 142a of the specific individual, it is possible to accurately determine whether or not the face is the face of the specific individual, and it is possible to reduce the load on the determination process.
  • the first face image processing unit 26 can accurately perform the face image processing of the specific individual.
  • the specific individual determination unit 25 determines that the face is not a specific individual's face, in other words, a normal face (a face of a person other than the specific individual)
  • the second face image processing unit 30 determines that the face is a normal face. Image processing can be performed with high accuracy. Therefore, whether the driver 3 is a specific individual or an ordinary person other than the specific individual, it is possible to accurately perform sensing of each face.
  • the specific individual determination unit 25 calculates a correlation coefficient as an index showing the correlation between the feature amount extracted from the face region and the face feature amount 142a of the specific individual, and based on the calculated correlation coefficient, It is determined whether or not the face in the face region is the face of the specific individual. As a result, it is possible to efficiently determine whether or not the face in the face region is the face of the specific individual based on the correlation coefficient, and a process of comparing the correlation coefficient with a predetermined threshold value. Therefore, the processing efficiency of the determination can be further improved.
  • the in-vehicle system 1 includes a driver monitoring device 10 and one or more ECUs 40 that execute a predetermined process based on the monitoring result output from the driver monitoring device 10. Therefore, based on the result of the monitoring, the ECU 40 can appropriately execute a predetermined control. This makes it possible to construct a highly safe in-vehicle system that allows even a specific individual to drive with peace of mind.
  • the embodiments of the present invention have been described in detail above, the above description is merely an example of the present invention in all respects. Needless to say, various improvements and changes can be made without departing from the scope of the present invention.
  • the case where the image processing device according to the present invention is applied to the driver monitoring device 10 has been described, but the application example is not limited to this.
  • the image processing apparatus according to the present invention can be applied.
  • Embodiments of the present invention may also be described as, but are not limited to, the following appendices.
  • Appendix 1 An image processing device (12) that processes an image input from the image pickup unit (11).
  • a face feature storage unit that stores a specific individual's face feature (142a) and a normal face feature (142b) as learned face features that have been learned to detect a face from the image.
  • (142) and A face detection unit (22) that detects a face region while extracting a feature amount for detecting a face from the image, and a face detection unit (22).
  • a specific individual determination unit that determines whether or not the face in the face region is the face of the specific individual by using the detected feature amount of the face region and the face feature amount (142a) of the specific individual. (25) and When the specific individual determination unit (25) determines that the face is the specific individual's face, the first face image processing unit (26) that performs face image processing for the specific individual and Image processing characterized by including a second face image processing unit (30) that performs normal face image processing when the specific individual determination unit (25) determines that the face is not the face of the specific individual. apparatus.
  • the storage unit (14) A face feature storage unit that stores a specific individual's face feature (142a) and a normal face feature (142b) as learned face features that have been learned to detect a face from the image.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Collating Specific Patterns (AREA)
  • Traffic Control Systems (AREA)

Abstract

An image processing device for processing an image inputted from an imaging unit, the image processing device being provided with: a facial feature amount storage unit for storing, as learned facial feature amounts, a facial feature amount of a specific individual and a normal facial feature amount; a face detection unit for detecting a face region while extracting the feature amount of a face from the image; a specific individual assessment unit for assessing, using the feature amount of the detected face region and the facial feature amount of the specific individual, whether the face in the face region is the face of the specific individual; a first facial image processing unit for performing facial image processing for a specific individual when it is assessed that the face is the face of the specific individual; and a second facial image processing unit for performing normal facial image processing when it is assessed that the face is not the face of the specific individual.

Description

画像処理装置、モニタリング装置、制御システム、画像処理方法、及びプログラムImage processing equipment, monitoring equipment, control systems, image processing methods, and programs
 本発明は、画像処理装置、モニタリング装置、制御システム、画像処理方法、及びプログラムに関する。 The present invention relates to an image processing device, a monitoring device, a control system, an image processing method, and a program.
 下記の特許文献1には、サービスを提供する対象(人物)の状況に応じて、適切なサービスに切り替え可能なサービス提供装置として利用されるロボット装置が開示されている。
 前記ロボット装置には、第1カメラと、第2カメラと、CPUを含む情報処理装置とが装備され、前記CPUには、顔検出部、属性判定部、人物検出部、人物位置算出部、及び移動ベクトル検出部などが装備されている。
Patent Document 1 below discloses a robot device used as a service providing device that can switch to an appropriate service according to the situation of a target (person) to which the service is provided.
The robot device is equipped with a first camera, a second camera, and an information processing device including a CPU, and the CPU includes a face detection unit, an attribute determination unit, a person detection unit, a person position calculation unit, and an information processing unit. It is equipped with a movement vector detector and the like.
 前記ロボット装置によれば、サービスの提供対象が、互いに意思疎通を行うなどの関係が成立している人物の集合である場合は、密なやり取りに基づいた情報を提供する第1サービスを行うことを決定する。一方、サービスの提供対象が、互いに意思疎通を行うなどの関係が成立しているか否かが不明な人物の集合である場合は、やり取りを行わずに、一方的に情報を提供する第2サービスを行うことを決定する。これにより、サービスの提供対象の状況に応じて、適切なサービスを行うことができるとしている。 According to the robot device, when the service is provided to a group of people who have a relationship such as communicating with each other, the first service of providing information based on close communication is performed. To determine. On the other hand, when the service is provided to a group of people whose relationship such as communication with each other is unknown, the second service provides information unilaterally without exchanging information. Decide to do. As a result, it is possible to provide appropriate services according to the situation of the service provision target.
特開2014-14899号公報Japanese Unexamined Patent Publication No. 2014-14899
 前記ロボット装置では、前記顔検出部が、前記第1カメラを用いて人物の顔検出を行う構成になっており、該顔検出には、公知の技術を利用することができるとしている。
 しかしながら、従来の顔検出技術では、ケガなどにより、目、鼻、口などの顔器官の一部が欠損、若しくは大きく変形している場合、顔に大きなホクロやイボ、若しくはタトゥーなどの身体装飾が施されている場合、又は遺伝性の疾患などの病気により、前記顔器官の配置が平均的な位置からずれている場合など、このような特定個人(換言すれば、年齢差、性別、及び人種などの違いにかかわらずに共通する一般的な人の顔特徴とは異なっている特徴を有する特定の個人)に対する顔検出の精度が低下してしまうという課題があった。
In the robot device, the face detection unit is configured to detect a person's face using the first camera, and a known technique can be used for the face detection.
However, with conventional face detection technology, if a part of the facial organs such as eyes, nose, and mouth is missing or greatly deformed due to injury, a large mole, wart, or body decoration such as tattoo is applied to the face. Such specific individuals (in other words, age difference, gender, and person), such as when the facial organs are displaced from their average position due to treatment or a disease such as a hereditary disease. There is a problem that the accuracy of face detection for a specific individual (a specific individual) having characteristics different from those of a general person, which is common regardless of the difference in species, is lowered.
 本発明は上記課題に鑑みなされたものであって、上記のような特定個人に対する顔センシングの精度を向上させることができる画像処理装置、モニタリング装置、制御システム、画像処理方法、及びプログラムを提供することを目的としている。
 上記目的を達成するために本開示に係る画像処理装置(1)は、撮像部から入力される画像を処理する画像処理装置であって、
 前記画像から顔を検出するための学習を行った学習済みの顔特徴量として、特定個人の顔特徴量と、通常の顔特徴量とが記憶される顔特徴量記憶部と、
 前記画像から顔を検出するための特徴量を抽出しながら顔領域を検出する顔検出部と、
 検出された前記顔領域の前記特徴量と、前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定部と、
 該特定個人判定部により前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理部と、
 前記特定個人判定部により前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理部とを備えていることを特徴としている。
The present invention has been made in view of the above problems, and provides an image processing device, a monitoring device, a control system, an image processing method, and a program capable of improving the accuracy of face sensing for a specific individual as described above. The purpose is.
The image processing apparatus (1) according to the present disclosure in order to achieve the above object is an image processing apparatus that processes an image input from an imaging unit.
As the learned facial features that have been learned to detect the face from the image, a facial feature storage unit that stores the facial features of a specific individual and the normal facial features, and
A face detection unit that detects a face region while extracting a feature amount for detecting a face from the image, and a face detection unit.
Using the detected feature amount of the face region and the face feature amount of the specific individual, a specific individual determination unit for determining whether or not the face in the face region is the face of the specific individual.
When the specific individual determination unit determines that the face is the face of the specific individual, the first face image processing unit that performs face image processing for the specific individual and the first face image processing unit
When it is determined by the specific individual determination unit that it is not the face of the specific individual, it is characterized by including a second face image processing unit that performs normal face image processing.
 上記画像処理装置(1)によれば、前記顔特徴量記憶部に前記学習済みの顔特徴量として、前記特定個人の顔特徴量と、前記通常の顔特徴量(換言すれば、前記特定個人以外の人である場合に用いる顔特徴量)とが記憶され、前記特定個人判定部により、前記顔検出部で検出された前記顔領域の前記特徴量と、前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かが判定される。前記特定個人の顔特徴量を用いることにより、前記特定個人の顔であるか否かを精度良く判定することができる。 According to the image processing device (1), the face feature amount of the specific individual and the normal face feature amount (in other words, the specific individual) are used as the learned face feature amount in the face feature amount storage unit. The facial feature amount used when the person is a person other than the above) is stored, and the feature amount of the face region detected by the face detection unit and the facial feature amount of the specific individual are stored by the specific individual determination unit. It is used to determine whether or not the face in the face region is the face of the specific individual. By using the facial feature amount of the specific individual, it is possible to accurately determine whether or not the face is the face of the specific individual.
 また、前記特定個人の顔であると判定された場合、前記第1顔画像処理部により前記特定個人の顔画像処理を精度良く実施することができる。一方、前記特定個人の顔ではない、換言すれば、通常の顔(前記特定個人以外の人の顔)であると判定された場合、前記第2顔画像処理部により前記通常の顔画像処理を精度良く実施することができる。したがって、前記特定個人であっても、特定個人以外の通常の人であっても、それぞれの顔のセンシングを精度良く実施することができる。 Further, when it is determined that the face of the specific individual is the face, the face image processing of the specific individual can be accurately performed by the first face image processing unit. On the other hand, when it is determined that the face is not the face of the specific individual, in other words, the face of a person other than the specific individual, the second face image processing unit performs the normal face image processing. It can be carried out with high accuracy. Therefore, both the specific individual and the ordinary person other than the specific individual can accurately perform the sensing of each face.
 また本開示に係る画像処理装置(2)は、上記画像処理装置(1)において、前記特定個人判定部が、前記顔領域から抽出された前記特徴量と前記特定個人の顔特徴量との相関を示す指標を算出し、算出した前記指標に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定するものであることを特徴としている。
 上記画像処理装置(2)によれば、前記顔領域から抽出された前記特徴量と前記特定個人の顔特徴量との相関を示す指標を算出し、算出した前記指標に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かが判定される。これにより、前記指標に基づいて前記顔領域の顔が前記特定個人の顔であるか否かを効率良く判定することができる。前記指標は、その値が大きいほど関係性が高くなることを示す指標値、例えば、相関係数であってもよいし、二乗誤差の逆数であってもよいし、その他、前記顔領域から抽出された前記特徴量と前記特定個人の顔特徴量との関係の類似度を示す指標値などであってもよい。
Further, in the image processing device (2) according to the present disclosure, in the image processing device (1), the specific individual determination unit correlates the feature amount extracted from the face region with the face feature amount of the specific individual. It is characterized in that an index indicating the above is calculated, and based on the calculated index, it is determined whether or not the face in the face region is the face of the specific individual.
According to the image processing apparatus (2), an index showing the correlation between the feature amount extracted from the face area and the face feature amount of the specific individual is calculated, and the face area is based on the calculated index. It is determined whether or not the face of the specific individual is the face of the specific individual. Thereby, it is possible to efficiently determine whether or not the face in the face region is the face of the specific individual based on the index. The index may be an index value indicating that the larger the value is, the higher the relationship is, for example, a correlation coefficient, the reciprocal of the square error, or other extraction from the face region. It may be an index value or the like indicating the degree of similarity of the relationship between the said feature amount and the face feature amount of the specific individual.
 また本開示に係る画像処理装置(3)は、上記画像処理装置(2)において、前記特定個人判定部が、前記指標が所定の閾値より大きい場合、前記顔領域の顔が前記特定個人の顔であると判定し、前記指標が前記所定の閾値以下の場合、前記顔領域の顔が前記特定個人の顔ではないと判定するものであることを特徴としている。
 上記画像処理装置(3)によれば、前記指標が所定の閾値より大きい場合、前記顔領域の顔が前記特定個人の顔であると判定され、前記指標が前記所定の閾値以下の場合、前記顔領域の顔が前記特定個人の顔ではないと判定される。前記指標と前記所定の閾値とを比較する処理により、前記判定の処理効率を更に高めることができる。
Further, in the image processing device (3) according to the present disclosure, in the image processing device (2), when the specific individual determination unit indicates that the index is larger than a predetermined threshold value, the face in the face region is the face of the specific individual. When the index is equal to or less than the predetermined threshold value, it is determined that the face in the face region is not the face of the specific individual.
According to the image processing device (3), when the index is larger than a predetermined threshold value, it is determined that the face in the face region is the face of the specific individual, and when the index is equal to or less than the predetermined threshold value, the face is described. It is determined that the face in the face area is not the face of the specific individual. The processing efficiency of the determination can be further improved by the process of comparing the index with the predetermined threshold value.
 また本開示に係る画像処理装置(4)は、上記画像処理装置(1)~(3)のいずれかにおいて、前記特定個人判定部が、前記画像の1フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定するものであることを特徴としている。
 上記画像処理装置(4)によれば、前記画像の1フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かが判定されるので、該判定の高速化を図ることができる。
Further, in the image processing device (4) according to the present disclosure, in any of the image processing devices (1) to (3), the specific individual determination unit is based on the result of determination for one frame of the image. It is characterized in that it is determined whether or not the face in the face region is the face of the specific individual.
According to the image processing device (4), it is determined whether or not the face in the face region is the face of the specific individual based on the result of the determination for one frame of the image, so that the determination is high speed. Can be achieved.
 また本開示に係る画像処理装置(5)は、上記画像処理装置(1)~(3)のいずれかにおいて、前記特定個人判定部が、前記画像の複数フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定するものであることを特徴としている。
 上記画像処理装置(5)によれば、前記画像の複数フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かが判定されるので、該判定の精度を高めることができる。
Further, in the image processing device (5) according to the present disclosure, in any of the image processing devices (1) to (3), the specific individual determination unit is based on the result of determination for a plurality of frames of the image. It is characterized in that it is determined whether or not the face in the face region is the face of the specific individual.
According to the image processing device (5), it is determined whether or not the face in the face region is the face of the specific individual based on the result of the determination for a plurality of frames of the image, so that the accuracy of the determination is determined. Can be enhanced.
 また本開示に係る画像処理装置(6)は、上記画像処理装置(1)~(5)のいずれかにおいて、前記顔画像処理には、顔検出処理、顔向き推定処理、視線方向推定処理、及び目開閉検出処理のうちの少なくともいずれかが含まれていることを特徴としている。
 上記画像処理装置(6)によれば、前記顔画像処理には、顔検出処理、顔向き推定処理、視線方向推定処理、及び目開閉検出処理のうちの少なくともいずれかが含まれているので、前記特定個人、又は前記特定個人以外の(通常の)人のさまざまな顔の挙動を推定したり、検出したりする処理を精度良く行うことができる。
Further, the image processing device (6) according to the present disclosure is any of the above image processing devices (1) to (5), and the face image processing includes face detection processing, face orientation estimation processing, and eye direction estimation processing. And at least one of the eye opening / closing detection processing is included.
According to the image processing apparatus (6), the face image processing includes at least one of face detection processing, face orientation estimation processing, line-of-sight direction estimation processing, and eye opening / closing detection processing. It is possible to accurately perform processing for estimating and detecting various facial behaviors of the specific individual or a (normal) person other than the specific individual.
 また本開示に係るモニタリング装置(1)は、上記画像処理装置(1)~(6)のいずれかと、該画像処理装置に入力する画像を撮像する撮像部と、前記画像処理装置による画像処理に基づく情報を出力する出力部とを備えていることを特徴としている。
 上記モニタリング装置(1)によれば、前記通常の人の顔だけでなく、前記特定個人の顔を精度良くモニタリングすることができ、また、前記出力部から前記画像処理に基づく情報が出力可能なため、該情報を利用するモニタリングシステムなどを容易に構築することが可能となる。
Further, the monitoring device (1) according to the present disclosure is used for any of the above image processing devices (1) to (6), an image pickup unit for capturing an image to be input to the image processing device, and image processing by the image processing device. It is characterized by having an output unit that outputs based information.
According to the monitoring device (1), not only the face of the normal person but also the face of the specific individual can be accurately monitored, and information based on the image processing can be output from the output unit. Therefore, it is possible to easily construct a monitoring system or the like that uses the information.
 また本開示に係る制御システム(1)は、上記モニタリング装置(1)と、該モニタリング装置と通信可能に接続され、該モニタリング装置から出力される前記情報に基づいて、所定の処理を実行する1以上の制御装置とを備えていることを特徴としている。
 上記制御システム(1)によれば、前記モニタリング装置から出力される前記情報に基づいて、1以上の前記制御装置で所定の処理を実行させることが可能となる。したがって、前記通常の人のモニタリング結果だけでなく、前記特定個人のモニタリング結果を利用することができるシステムを構築することができる。
Further, the control system (1) according to the present disclosure is communicably connected to the monitoring device (1) and executes a predetermined process based on the information output from the monitoring device. It is characterized by having the above-mentioned control device.
According to the control system (1), it is possible to execute a predetermined process by one or more of the control devices based on the information output from the monitoring device. Therefore, it is possible to construct a system that can utilize not only the monitoring result of the normal person but also the monitoring result of the specific individual.
 また本開示に係る制御システム(2)は、上記制御システム(1)において、
 前記モニタリング装置が、車両のドライバをモニタリングするための装置であり、
 前記制御装置が、前記車両に搭載される電子制御ユニットを含むことを特徴としている。
 上記制御システム(2)によれば、前記車両のドライバが前記特定個人である場合であっても、前記特定個人の顔を精度良くモニタリングすることができ、そのモニタリングの結果に基づいて、前記電子制御ユニットに所定の制御を適切に実行させることが可能となる。これにより、前記特定個人であっても安心して運転することができる安全性の高い車載システムを構築することが可能となる。
Further, the control system (2) according to the present disclosure is the control system (1) described above.
The monitoring device is a device for monitoring the driver of the vehicle.
The control device is characterized by including an electronic control unit mounted on the vehicle.
According to the control system (2), even when the driver of the vehicle is the specific individual, the face of the specific individual can be accurately monitored, and the electronic device is based on the monitoring result. It is possible to make the control unit appropriately execute a predetermined control. This makes it possible to construct a highly safe in-vehicle system that allows even the specific individual to drive with peace of mind.
 また本開示に係る画像処理方法は、撮像部から入力される画像を処理する画像処理方法であって、
 前記画像から顔の特徴量を抽出しながら顔領域を検出する顔検出ステップと、
 該顔検出ステップにより検出された前記顔領域の前記特徴量と、特定個人の顔を検出するための学習を行った学習済みの前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定ステップと、
 該特定個人判定ステップにより前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理ステップと、
 前記特定個人判定ステップにより前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理ステップとを含むことを特徴としている。
Further, the image processing method according to the present disclosure is an image processing method for processing an image input from an imaging unit.
A face detection step of detecting a face region while extracting facial features from the image, and
The face of the face region is used by using the feature amount of the face region detected by the face detection step and the learned face feature amount of the specific individual who has been trained to detect the face of the specific individual. A specific individual determination step for determining whether or not is the face of the specific individual, and
When the face of the specific individual is determined by the specific individual determination step, the first face image processing step of performing the face image processing for the specific individual and the first face image processing step
When it is determined by the specific individual determination step that the face is not the face of the specific individual, it is characterized by including a second face image processing step of performing normal face image processing.
 上記画像処理方法によれば、前記特定個人判定ステップにより、前記顔検出ステップで検出された前記顔領域の前記特徴量と、前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かが判定される。前記特定個人の顔特徴量を用いることにより、前記特定個人の顔であるか否かを精度良く判定することができる。
 また、前記特定個人の顔であると判定された場合、前記第1顔画像処理ステップにより前記特定個人の顔画像処理を精度良く実施することができる。一方、前記特定個人の顔ではない、換言すれば、前記特定個人ではない通常の顔であると判定された場合、前記第2顔画像処理ステップにより前記通常の顔画像処理を精度良く実施することができる。したがって、前記特定個人であっても、特定個人以外の通常の人であっても、それぞれの顔のセンシングを精度良く実施することができる。
According to the image processing method, the face in the face region is formed by using the feature amount of the face region detected in the face detection step and the face feature amount of the specific individual in the specific individual determination step. Whether or not it is the face of the specific individual is determined. By using the facial feature amount of the specific individual, it is possible to accurately determine whether or not the face is the face of the specific individual.
Further, when it is determined that the face of the specific individual is the face, the face image processing of the specific individual can be accurately performed by the first face image processing step. On the other hand, when it is determined that the face is not the specific individual's face, in other words, it is a normal face that is not the specific individual, the normal face image processing is performed with high accuracy by the second face image processing step. Can be done. Therefore, both the specific individual and a normal person other than the specific individual can accurately sense each face.
 また本開示に係るプログラムは、撮像部から入力される画像の処理を少なくとも1以上のコンピュータに実行させるためのプログラムであって、
 前記少なくとも1以上のコンピュータに、
 前記画像から顔の特徴量を抽出しながら顔領域を検出する顔検出ステップと、
 該顔検出ステップにより検出された前記顔領域の前記特徴量と、特定個人の顔を検出するための学習を行った学習済みの前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定ステップと、
 該特定個人判定ステップにより前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理ステップと、
 前記特定個人判定ステップにより前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理ステップとを実行させることを特徴としている。
Further, the program according to the present disclosure is a program for causing at least one or more computers to process an image input from an imaging unit.
To at least one of the above computers
A face detection step of detecting a face region while extracting facial features from the image, and
The face of the face region is used by using the feature amount of the face region detected by the face detection step and the learned face feature amount of the specific individual who has been trained to detect the face of the specific individual. A specific individual determination step for determining whether or not is the face of the specific individual, and
When the face of the specific individual is determined by the specific individual determination step, the first face image processing step of performing the face image processing for the specific individual and the first face image processing step
When it is determined by the specific individual determination step that the face is not the face of the specific individual, the second face image processing step of performing normal face image processing is executed.
 上記プログラムによれば、前記少なくとも1以上のコンピュータに、前記顔領域の前記特徴量と、前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定させることができ、前記特定個人の顔であるか否かを精度良く判定させることができる。
 また、前記特定個人の顔であると判定された場合、前記特定個人の顔画像処理を精度良く実施させることができる。一方、前記特定個人の顔ではない、換言すれば、前記特定個人ではない通常の顔であると判定された場合、前記通常の顔画像処理を精度良く実施させることができる。したがって、前記特定個人であっても、特定個人以外の通常の人であっても、それぞれの顔のセンシングを精度良く実施することができる装置やシステムを構築することができる。なお、上記プログラムは、記憶媒体に保存されたプログラムであってもよいし、通信ネットワークを介して転送可能なプログラムであってもよいし、通信ネットワークを介して実行されるプログラムであってもよい。
According to the above program, whether or not the face in the face region is the face of the specific individual by using the feature amount of the face region and the face feature amount of the specific individual on at least one computer. It is possible to determine whether or not the face is the specific individual's face with high accuracy.
Further, when it is determined that the face of the specific individual is the face, the face image processing of the specific individual can be performed with high accuracy. On the other hand, when it is determined that the face is not the specific individual's face, in other words, it is a normal face that is not the specific individual, the normal face image processing can be performed with high accuracy. Therefore, it is possible to construct a device or system capable of accurately sensing each face regardless of whether the specific individual or an ordinary person other than the specific individual. The above program may be a program stored in a storage medium, a program that can be transferred via a communication network, or a program that is executed via a communication network. ..
本発明の実施の形態に係るドライバモニタリング装置を含む車載システムの一例を示す模式図である。It is a schematic diagram which shows an example of the in-vehicle system including the driver monitoring apparatus which concerns on embodiment of this invention. 実施の形態に係るドライバモニタリング装置を含む車載システムのハードウェア構成の一例を示すブロック図である。It is a block diagram which shows an example of the hardware composition of the in-vehicle system including the driver monitoring apparatus which concerns on embodiment. 実施の形態に係るドライバモニタリング装置の画像処理部の機能構成例を示すブロック図である。It is a block diagram which shows the functional structure example of the image processing part of the driver monitoring apparatus which concerns on embodiment. 実施の形態に係るドライバモニタリング装置の画像処理部が行う処理動作の一例を示すフローチャートである。It is a flowchart which shows an example of the processing operation performed by the image processing part of the driver monitoring apparatus which concerns on embodiment. 実施の形態に係るドライバモニタリング装置の画像処理部が行う特定個人判定処理動作の一例を示すフローチャートである。It is a flowchart which shows an example of the specific individual judgment processing operation performed by the image processing part of the driver monitoring apparatus which concerns on embodiment. 実施の形態に係るドライバモニタリング装置の画像処理部が行う特定個人判定処理動作の別の一例を示すフローチャートである。It is a flowchart which shows another example of the specific individual determination processing operation performed by the image processing part of the driver monitoring apparatus which concerns on embodiment.
 以下、本発明に係る画像処理装置、モニタリング装置、制御システム、画像処理方法、及びプログラムの実施の形態を図面に基づいて説明する。
 本発明に係る画像処理装置は、例えば、カメラを用いて人などの対象物をモニタリングする装置やシステムに広く適用可能である。本発明に係る画像処理装置は、例えば、車両などの各種移動体のドライバ(操縦者)をモニタリングする装置やシステムの他、工場内の機械や装置などの各種設備を操作したり、監視したり、所定の作業をしたりする人などをモニタリングする装置やシステムなどにも適用可能である。
Hereinafter, an image processing device, a monitoring device, a control system, an image processing method, and an embodiment of a program according to the present invention will be described with reference to the drawings.
The image processing apparatus according to the present invention can be widely applied to, for example, an apparatus or system for monitoring an object such as a person using a camera. The image processing device according to the present invention operates or monitors, for example, various facilities such as machines and devices in a factory, in addition to devices and systems for monitoring drivers (operators) of various moving objects such as vehicles. It can also be applied to devices and systems that monitor people who perform predetermined work.
[適用例]
 図1は、実施の形態に係るドライバモニタリング装置を含む車載システムの一例を示す模式図である。本適用例では、本発明に係る画像処理装置をドライバモニタリング装置10に適用した例について説明する。
 車載システム1は、車両2のドライバ3の状態(例えば、顔の挙動など)をモニタリングするドライバモニタリング装置10、車両2の走行、操舵、又は制動などの制御を行う1以上のECU(Electronic Control Unit)40、及び車両各部の状態、又は車両周囲の状態などを検出する1以上のセンサ41を含んで構成され、これらが通信バス43を介して接続されている。車載システム1は、例えば、CAN(Controller Area Network)プロトコルに従って通信する車載ネットワークシステムとして構成されている。なお、車載システム1の通信規格には、CAN以外の他の通信規格が採用されてもよい。ドライバモニタリング装置10が、本発明の「モニタリング装置」の一例であり、車載システム1が、本発明の「制御システム」の一例である。
[Application example]
FIG. 1 is a schematic view showing an example of an in-vehicle system including the driver monitoring device according to the embodiment. In this application example, an example in which the image processing apparatus according to the present invention is applied to the driver monitoring apparatus 10 will be described.
The in-vehicle system 1 includes a driver monitoring device 10 that monitors the state of the driver 3 of the vehicle 2 (for example, facial behavior), and one or more ECUs (Electronic Control Units) that control the running, steering, or braking of the vehicle 2. ) 40, and one or more sensors 41 for detecting the state of each part of the vehicle, the state around the vehicle, and the like are included, and these are connected via the communication bus 43. The in-vehicle system 1 is configured as, for example, an in-vehicle network system that communicates according to a CAN (Controller Area Network) protocol. As the communication standard of the in-vehicle system 1, a communication standard other than CAN may be adopted. The driver monitoring device 10 is an example of the "monitoring device" of the present invention, and the in-vehicle system 1 is an example of the "control system" of the present invention.
 ドライバモニタリング装置10は、ドライバ3の顔を撮像するためのカメラ11と、カメラ11から入力される画像を処理する画像処理部12と、画像処理部12による画像処理に基づく情報を、通信バス43を介して所定のECU40に出力する処理などを行う通信部16とを含んで構成されている。画像処理部12が、本発明の「画像処理装置」の一例である。カメラ11が、本発明の「撮像部」の一例である。 The driver monitoring device 10 transmits information based on image processing by the camera 11 for capturing the face of the driver 3, the image processing unit 12 that processes the image input from the camera 11, and the image processing unit 12, and the communication bus 43. It is configured to include a communication unit 16 that performs processing such as output to a predetermined ECU 40 via the above. The image processing unit 12 is an example of the "image processing device" of the present invention. The camera 11 is an example of the "imaging unit" of the present invention.
 ドライバモニタリング装置10は、カメラ11で撮像された画像からドライバ3の顔を検出し、検出されたドライバ3の顔の向き、視線の方向、又は目の開閉状態などの顔の挙動を検出する。ドライバモニタリング装置10は、これら顔の挙動の検出結果に基づいて、ドライバ3の状態、例えば、前方注視、脇見、居眠り、後ろ向き、突っ伏しなどの状態を判定してもよい。また、ドライバモニタリング装置10が、これらドライバ3の状態判定に基づく信号をECU40に出力し、ECU40が、前記信号に基づいてドライバ3への注意や警告処理、又は車両2の動作制御(例えば、減速制御、又は路肩への誘導制御など)などを実行するように構成してもよい。 The driver monitoring device 10 detects the face of the driver 3 from the image captured by the camera 11, and detects the behavior of the face such as the direction of the face of the detected driver 3, the direction of the line of sight, or the open / closed state of the eyes. The driver monitoring device 10 may determine the state of the driver 3, such as forward gaze, inattentiveness, dozing, backward facing, and prone, based on the detection results of these facial behaviors. Further, the driver monitoring device 10 outputs a signal based on the state determination of the driver 3 to the ECU 40, and the ECU 40 performs attention and warning processing to the driver 3 or operation control of the vehicle 2 (for example, deceleration) based on the signal. Control, guidance control to the road shoulder, etc.) may be executed.
 ドライバモニタリング装置10では、特定個人に対する顔センシングの精度を向上させることを目的の一つとしている。
 従来のドライバモニタリング装置では、車両2のドライバ3が、例えば、ケガなどにより、目、鼻、口などの顔器官の一部が欠損、若しくは大きく変形していたり、顔に大きなホクロやイボ、若しくはタトゥーなどの身体装飾が施されていたり、又は遺伝性の疾患などの病気により、前記顔器官の配置が平均的な位置からずれていたりした場合、カメラで撮像された画像から顔を検出する精度が低下してしまうという課題があった。
One of the purposes of the driver monitoring device 10 is to improve the accuracy of face sensing for a specific individual.
In the conventional driver monitoring device, the driver 3 of the vehicle 2 has a part of facial organs such as eyes, nose, and mouth missing or greatly deformed due to, for example, an injury, or a large mole or wart on the face, or Accuracy of detecting the face from the image captured by the camera when the facial organs are displaced from the average position due to body decoration such as tattoo or a disease such as a hereditary disease. There was a problem that the
 また、顔検出精度が低下すると、顔向き推定処理など、顔検出後の処理も適切に行われないこととなるため、ドライバ3の脇見や居眠りなどの状態判定も適切に行うことができなくなり、また、前記状態判定に基づいてECU40が実行すべき各種の制御も適切に行うことができなくなる虞があるという課題があった。
 係る課題を解決すべく、実施の形態に係るドライバモニタリング装置10では、特定個人、換言すれば、年齢差、性別、及び人種などの違い(個人差)にかかわらずに共通する一般的な人(普通の人)の顔特徴とは異なる特徴を有している特定の個人に対する顔検出の精度を向上させるために、以下の構成を採用した。
Further, if the face detection accuracy is lowered, the post-face detection processing such as the face orientation estimation processing is not properly performed, so that the driver 3 cannot properly perform the state determination such as inattentiveness or dozing. Further, there is a problem that various controls to be executed by the ECU 40 based on the state determination may not be appropriately performed.
In order to solve the problem, the driver monitoring device 10 according to the embodiment is a general person who is common regardless of a specific individual, in other words, a difference in age, gender, race, etc. (individual difference). In order to improve the accuracy of face detection for a specific individual who has features different from the face features of (ordinary person), the following configuration was adopted.
 画像処理部12には、画像から顔を検出するための学習を行った学習済みの顔特徴量として、特定個人の顔特徴量と、通常の顔特徴量(換言すれば、特定個人以外の人である場合に用いる顔特徴量)とが記憶されている。
 画像処理部12が、カメラ11の入力画像から顔を検出するための特徴量を抽出しながら顔領域を検出する顔検出処理を行う。そして、画像処理部12が、検出された前記顔領域の特徴量と、前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定処理を行う。
In the image processing unit 12, as the learned facial features that have been learned to detect the face from the image, the facial features of a specific individual and the normal facial features (in other words, a person other than the specific individual). The amount of facial features used in the case of) is stored.
The image processing unit 12 performs face detection processing for detecting a face region while extracting a feature amount for detecting a face from an input image of the camera 11. Then, the image processing unit 12 determines whether or not the face in the face region is the face of the specific individual by using the detected feature amount of the face region and the face feature amount of the specific individual. Performs specific individual judgment processing.
 前記特定個人判定処理では、前記顔領域から抽出された特徴量と前記特定個人の顔特徴量との関係を示す指標、例えば、相関を示す指標として、相関係数を算出し、算出した前記相関係数に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定してもよい。
 例えば、前記相関係数が所定の閾値より大きい場合、前記顔領域の顔が前記特定個人の顔であると判定し、前記相関係数が前記所定の閾値以下の場合、前記顔領域の顔が前記特定個人の顔ではないと判定してもよい。なお、前記特定個人判定処理では、前記相関係数以外の指標を用いてもよい。
In the specific individual determination process, a correlation coefficient is calculated and calculated as an index showing the relationship between the feature amount extracted from the face region and the face feature amount of the specific individual, for example, an index showing the correlation. Based on the number of relationships, it may be determined whether or not the face in the face region is the face of the specific individual.
For example, when the correlation coefficient is larger than a predetermined threshold value, it is determined that the face in the face region is the face of the specific individual, and when the correlation coefficient is equal to or less than the predetermined threshold value, the face in the face region is It may be determined that it is not the face of the specific individual. In the specific individual determination process, an index other than the correlation coefficient may be used.
 また、前記特定個人判定処理では、カメラ11からの入力画像の1フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定してもよいし、カメラ11からの入力画像の複数フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定してもよい。
 このように、ドライバモニタリング装置10では、画像処理部12に、学習済みの特定個人の顔特徴量を予め記憶しておき、特定個人の顔特徴量を用いることにより、前記特定個人の顔であるか否かを精度良く判定することが可能となる。
Further, in the specific individual determination process, it may be determined whether or not the face in the face region is the face of the specific individual based on the result of determination for one frame of the input image from the camera 11. Based on the result of determination for a plurality of frames of the input image from the camera 11, it may be determined whether or not the face in the face region is the face of the specific individual.
As described above, in the driver monitoring device 10, the learned facial feature amount of the specific individual is stored in advance in the image processing unit 12, and the face feature amount of the specific individual is used to obtain the face of the specific individual. It is possible to accurately determine whether or not it is.
 また、前記特定個人判定処理により前記特定個人の顔であると判定された場合、画像処理部12は、特定個人用の顔画像処理を実行するので、前記特定個人の顔画像処理を精度良く実施することが可能となる。
 一方、前記特定個人の顔ではない、換言すれば、通常の顔(換言すれば、特定個人以外の顔)であると判定された場合、画像処理部12は、通常の顔画像処理を実行するので、前記通常の顔画像処理を精度良く実施することができる。したがって、ドライバ3が、特定個人であっても、特定個人以外の通常の人であっても、それぞれの顔のセンシングを精度良く実施することができる。
Further, when the face of the specific individual is determined by the specific individual determination process, the image processing unit 12 executes the face image process for the specific individual, so that the face image process of the specific individual is accurately performed. It becomes possible to do.
On the other hand, when it is determined that the face is not the face of the specific individual, in other words, a normal face (in other words, a face other than the specific individual), the image processing unit 12 executes the normal face image processing. Therefore, the normal face image processing can be performed with high accuracy. Therefore, whether the driver 3 is a specific individual or an ordinary person other than the specific individual, it is possible to accurately perform sensing of each face.
[ハードウェア構成例]
 図2は、実施の形態に係るドライバモニタリング装置10を含む車載システム1のハードウェア構成の一例を示すブロック図である。
 車載システム1は、車両2のドライバ3の状態をモニタリングするドライバモニタリング装置10、1以上のECU40、及び1以上のセンサ41を含んで構成され、これらが通信バス43を介して接続されている。また、ECU40には、1以上のアクチュエータ42が接続されている。
 ドライバモニタリング装置10は、カメラ11と、カメラ11から入力される画像を処理する画像処理部12と、外部のECU40などとデータや信号のやり取りを行うための通信部16とを含んで構成されている。
[Hardware configuration example]
FIG. 2 is a block diagram showing an example of the hardware configuration of the in-vehicle system 1 including the driver monitoring device 10 according to the embodiment.
The in-vehicle system 1 includes a driver monitoring device 10, 1 or more ECUs 40 for monitoring the state of the driver 3 of the vehicle 2, and 1 or more sensors 41, which are connected via a communication bus 43. Further, one or more actuators 42 are connected to the ECU 40.
The driver monitoring device 10 includes a camera 11, an image processing unit 12 that processes an image input from the camera 11, and a communication unit 16 for exchanging data and signals with an external ECU 40 and the like. There is.
 カメラ11は、運転席に着座しているドライバ3の顔を含む画像を撮像する装置であり、例えば、レンズ部、撮像素子部、光照射部、インターフェース部、これら各部を制御するカメラ制御部などを含んで構成され得る。前記撮像素子部は、CCD(Charge Coupled Device)、CMOS(Complementary Metal Oxide Semiconductor)などの撮像素子、フィルタ、マイクロレンズなどを含んで構成され得る。前記撮像素子部は、可視領域の光を受けて撮像画像を形成できる素子でもよいし、近赤外領域の光を受けて撮像画像を形成できる素子でもよい。前記光照射部は、LED(Light Emitting Diode)などの発光素子を含んで構成され、昼夜を問わずドライバの顔を撮像できるように近赤外線LEDなどを含んでもよい。カメラ11は、所定のフレームレート(例えば、毎秒数十フレーム)で画像を撮像し、撮像された画像のデータが画像処理部12に入力される。カメラ11は、一体式の他、外付け式のものであってもよい。 The camera 11 is a device that captures an image including the face of the driver 3 seated in the driver's seat. For example, a lens unit, an image sensor unit, a light irradiation unit, an interface unit, a camera control unit that controls each of these units, and the like. Can be configured to include. The image sensor unit may include an image sensor such as a CCD (Charge Coupled Device) and a CMOS (Complementary Metal Oxide Semiconductor), a filter, a microlens, and the like. The image pickup device unit may be an element capable of forming an image pickup image by receiving light in a visible region, or an element capable of forming an image pickup image by receiving light in a near infrared region. The light irradiation unit is configured to include a light emitting element such as an LED (Light Emitting Diode), and may include a near infrared LED or the like so that the driver's face can be imaged day or night. The camera 11 captures an image at a predetermined frame rate (for example, several tens of frames per second), and the data of the captured image is input to the image processing unit 12. The camera 11 may be an external type as well as an integrated type.
 画像処理部12は、1以上のCPU(Central Processing Unit)13、ROM(Read Only Memory)14、及びRAM(Random Access Memory)15を含む画像処理装置として構成されている。ROM14は、プログラム記憶部141と顔特徴量記憶部142とを含み、RAM15は、カメラ11からの入力画像を記憶する画像メモリ151を含んで構成されている。なお、ドライバモニタリング装置10に、別の記憶部を設け、その記憶部をプログラム記憶部141、顔特徴量記憶部142、及び画像メモリ151として用いてもよい。前記別の記憶部は、半導体メモリでもよいし、ディスクドライブなどで読み込み可能な記憶媒体でもよい。 The image processing unit 12 is configured as an image processing device including one or more CPU (Central Processing Unit) 13, ROM (Read Only Memory) 14, and RAM (Random Access Memory) 15. The ROM 14 includes a program storage unit 141 and a facial feature amount storage unit 142, and the RAM 15 includes an image memory 151 for storing an input image from the camera 11. The driver monitoring device 10 may be provided with another storage unit, and the storage unit may be used as the program storage unit 141, the facial feature amount storage unit 142, and the image memory 151. The other storage unit may be a semiconductor memory or a storage medium that can be read by a disk drive or the like.
 CPU13は、ハードウェアプロセッサの一例であり、ROM14のプログラム記憶部141に記憶されているプログラム、顔特徴量記憶部142に記憶されている顔特徴量などのデータを読み込み、解釈し実行することで、カメラ11から入力された画像の処理、例えば、顔検出処理などの顔画像処理を行う。また、CPU13は、該顔画像処理により得られた結果(例えば、処理データ、判定信号、又は制御信号など)を、通信部16を介してECU40などに出力する処理などを行う。 The CPU 13 is an example of a hardware processor, and by reading, interpreting, and executing data such as a program stored in the program storage unit 141 of the ROM 14 and the face feature amount stored in the face feature amount storage unit 142. , Processing of the image input from the camera 11, for example, face image processing such as face detection processing is performed. Further, the CPU 13 performs a process of outputting the result (for example, processing data, determination signal, control signal, etc.) obtained by the face image processing to the ECU 40 or the like via the communication unit 16.
 顔特徴量記憶部142には、画像から顔を検出するための学習(例えば、機械学習)を行った学習済みの顔特徴量として、図3に示す特定個人の顔特徴量142aと、通常の顔特徴量142bとが記憶されている。
 学習済みの顔特徴量には、画像から顔を検出するのに有効な各種の特徴量を用いることができる。例えば、顔の局所的な領域の明暗差(さまざまな大きさの2つの矩形領域の平均輝度の差)に着目した特徴量(Haar-like特徴量)を用いてもよい。又は、顔の局所的な領域の輝度の分布の組み合わせに着目した特徴量(LBP (Local Binary Pattern) 特徴量)を用いてもよいし、顔の局所的な領域の輝度の勾配方向の分布の組み合わせに着目した特徴量(HOG (Histogram of Oriented Gradients) 特徴量)などを用いてもよい。
In the face feature amount storage unit 142, as the learned face feature amount that has been learned (for example, machine learning) to detect the face from the image, the face feature amount 142a of the specific individual shown in FIG. The facial feature amount 142b is stored.
As the learned facial features, various feature quantities effective for detecting a face from an image can be used. For example, a feature amount (Haar-like feature amount) focusing on the difference in brightness (difference in average brightness between two rectangular areas of various sizes) in a local area of the face may be used. Alternatively, a feature amount (LBP (Local Binary Pattern) feature amount) focusing on a combination of brightness distributions in the local region of the face may be used, or the distribution of the brightness in the local region of the face in the gradient direction may be used. Features (HOG (Histogram of Oriented Gradients) features) focusing on the combination may be used.
 顔特徴量記憶部142に記憶される顔特徴量は、例えば、各種の機械学習による手法を用いて、顔検出に有効な特徴量として抽出されたものである。機械学習とは、データ(学習データ)に内在するパターンをコンピュータにより見つけ出す処理である。例えば、統計的な学習手法の一例としてAdaBoostを用いてもよい。AdaBoostは、判別能力の低い判別器(弱判別器)を多数選び出し、これら多数の弱判別器の中からエラー率が小さい弱判別器を選択し、重みなどのパラメータを調整し、階層的な構造にすることで、強判別器を構築することのできる学習アルゴリズムである。なお、判別器は、識別器、分類器、又は学習器と称されてもよい。 The face feature amount stored in the face feature amount storage unit 142 is extracted as an effective feature amount for face detection by using, for example, various machine learning methods. Machine learning is a process of finding a pattern inherent in data (learning data) by a computer. For example, AdaBoost may be used as an example of a statistical learning method. AdaBoost selects a large number of discriminators (weak discriminators) with low discriminating ability, selects a weak discriminator with a small error rate from these many weak discriminators, adjusts parameters such as weights, and has a hierarchical structure. It is a learning algorithm that can construct a strong discriminator by setting. The discriminator may be referred to as a discriminator, a classifier, or a learner.
 強判別器は、例えば、顔の検出に有効な1つの特徴量を1つの弱判別器によって判別する構成とし、AdaBoostにより多数の弱判別器とその組み合わせを選び出し、これらを用いて、階層的な構造を構築したものとしてもよい。なお、1つの弱判別器は、例えば、顔の場合は1、非顔の場合は0という情報を出力してもよい。また、学習手法には、顔らしさを0または1ではなく、0から1の実数で出力可能なReal AdaBoostという学習手法を用いてもよい。また、これら学習手法には、入力層、中間層、及び出力層を有するニューラルネットワークを用いてもよい。 For example, the strong discriminator is configured to discriminate one feature amount effective for face detection by one weak discriminator, and a large number of weak discriminators and their combinations are selected by AdaBoost, and these are used hierarchically. The structure may be constructed. Note that one weak discriminator may output information such as 1 for a face and 0 for a non-face. Further, as the learning method, a learning method called Real AdaBoost, which can output a real number from 0 to 1 instead of 0 or 1, may be used. Further, as these learning methods, a neural network having an input layer, an intermediate layer, and an output layer may be used.
 このような学習アルゴリズムが搭載された学習装置に、さまざまな条件で撮像された多数の顔画像と多数の顔以外の画像(非顔画像)とを学習データとして与え、学習を繰り返し、重みなどのパラメータを調整して最適化を図ることにより、顔を高精度に検出可能な階層構造を有する強判別器を構築することが可能となる。そして、このような強判別器を構成する各階層の弱判別器で用いられる1以上の特徴量を、学習済みの顔特徴量として用いることができる。 A large number of face images captured under various conditions and a large number of non-face images (non-face images) are given as training data to a learning device equipped with such a learning algorithm, learning is repeated, weighting, etc. By adjusting the parameters and optimizing it, it is possible to construct a strong discriminator having a hierarchical structure capable of detecting a face with high accuracy. Then, one or more feature amounts used in the weak discriminators of each layer constituting such a strong discriminator can be used as the learned facial feature amounts.
 特定個人の顔特徴量142aは、例えば、予め所定の場所で、特定個人の顔画像をさまざまな条件(さまざまな顔の向き、視線の方向、又は目の開閉状態などの条件)で個別に撮像し、これら多数の撮像画像を教師データとして、上記学習装置に入力し、学習処理によって調整された、特定個人の顔の特徴を示すパラメータである。特定個人の顔特徴量142aは、例えば、学習処理によって得られた、顔の局所的な領域の明暗差の組み合わせパターンなどでもよい。顔特徴量記憶部142に記憶される特定個人の顔特徴量142aは、1人の特定個人の顔特徴量だけでもよいし、複数の特定個人が車両2を運転する場合などに対応できるように、複数人の特定個人の顔特徴量が記憶されてもよい。 For example, the face feature amount 142a of a specific individual individually captures a face image of the specific individual at a predetermined place under various conditions (conditions such as various face orientations, line-of-sight directions, or eye open / closed states). Then, these a large number of captured images are input to the learning device as teacher data, and are parameters that indicate the facial features of a specific individual adjusted by the learning process. The facial feature amount 142a of the specific individual may be, for example, a combination pattern of the difference in brightness of the local region of the face obtained by the learning process. The facial feature amount 142a of a specific individual stored in the facial feature amount storage unit 142 may be only the facial feature amount of one specific individual, or can be used when a plurality of specific individuals drive the vehicle 2. , The facial features of a plurality of specific individuals may be stored.
 通常の顔特徴量142bは、通常の人の顔画像をさまざまな条件(さまざまな顔の向き、視線の方向、又は目の開閉状態などの条件)で撮像した画像を教師データとして、上記学習装置に入力し、学習処理によって調整された、通常の人の顔の特徴を示すパラメータである。通常の顔特徴量142bは、例えば、学習処理によって得られた、顔の局所的な領域の明暗差の組み合わせパターンなどでもよい。また、通常の顔特徴量142bは、所定の顔特徴量データベースに登録されている情報を用いてもよい。 The normal facial feature amount 142b is the above-mentioned learning device using images of a normal human face image captured under various conditions (conditions such as various face orientations, line-of-sight directions, or eye open / closed states) as teacher data. It is a parameter indicating the characteristics of a normal human face, which is input to and adjusted by the learning process. The normal facial feature amount 142b may be, for example, a combination pattern of light and dark differences in a local region of the face obtained by a learning process. Further, as the normal facial feature amount 142b, the information registered in the predetermined facial feature amount database may be used.
 顔特徴量記憶部142に記憶される学習済みの顔特徴量は、例えば、クラウド上のサーバなどからインターネット、携帯電話網などの通信ネットワークを介して取り込んで、顔特徴量記憶部142に記憶される構成としてもよい。
 ECU40は、1以上のプロセッサ、メモリ、及び通信モジュールなどを含むコンピュータ装置で構成されている。そして、ECU40に搭載されたプロセッサが、メモリに記憶されたプログラムを読み込み、解釈し実行することで、アクチュエータ42などに対する所定の制御が実行されるようになっている。
The learned facial feature amount stored in the facial feature amount storage unit 142 is fetched from a server on the cloud or the like via a communication network such as the Internet or a mobile phone network and stored in the facial feature amount storage unit 142. It may be configured as such.
The ECU 40 is composed of a computer device including one or more processors, a memory, a communication module, and the like. Then, the processor mounted on the ECU 40 reads, interprets, and executes the program stored in the memory, so that predetermined control for the actuator 42 and the like is executed.
 ECU40は、例えば、走行系ECU、運転支援系ECU、ボディ系ECU、及び情報系ECUのうちの少なくともいずれかを含んで構成されている。
 前記走行系ECUには、例えば、駆動系ECU、シャーシ系ECUなどが含まれている。前記駆動系ECUには、例えば、エンジン制御、モータ制御、燃料電池制御、EV(Electric Vehicle)制御、又はトランスミッション制御等の「走る」機能に関する制御ユニットが含まれている。前記シャーシ系ECUには、例えば、ブレーキ制御、又はステアリング制御等の「止まる、曲がる」機能に関する制御ユニットが含まれている。
The ECU 40 includes, for example, at least one of a traveling system ECU, a driving support system ECU, a body system ECU, and an information system ECU.
The traveling system ECU includes, for example, a drive system ECU, a chassis system ECU, and the like. The drive system ECU includes a control unit related to a "running" function such as engine control, motor control, fuel cell control, EV (Electric Vehicle) control, or transmission control. The chassis-based ECU includes a control unit related to a "stop, turn" function such as brake control or steering control.
 前記運転支援系ECUは、例えば、自動ブレーキ支援機能、車線維持支援機能(LKA/Lane Keep Assistともいう)、定速走行・車間距離支援機能(ACC/Adaptive Cruise Controlともいう)、前方衝突警告機能、車線逸脱警報機能、死角モニタリング機能、交通標識認識機能等、走行系ECUなどとの連携により自動的に安全性の向上、又は快適な運転を実現する機能(運転支援機能、又は自動運転機能)に関する制御ユニットを少なくとも1つ以上含んで構成され得る。 The driving support system ECU has, for example, an automatic braking support function, a lane keeping support function (also referred to as LKA / Lane Keep Assist), a constant speed driving / inter-vehicle distance support function (also referred to as ACC / Adaptive Cruise Control), and a forward collision warning function. , Lane departure warning function, blind spot monitoring function, traffic sign recognition function, etc., functions that automatically improve safety or realize comfortable driving by linking with driving ECUs (driving support function or automatic driving function) It may be configured to include at least one control unit with respect to.
 前記運転支援系ECUには、例えば、米国自動車技術会(SAE)が提示している自動運転レベルにおけるレベル1(ドライバ支援)、レベル2(部分的自動運転)、及びレベル3(条件付自動運転)の少なくともいずれかの機能が装備されてもよい。さらに、自動運転レベルのレベル4(高度自動運転)、又はレベル5(完全自動運転)の機能が装備されてもよいし、レベル1、2のみ、又はレベル2、3のみの機能が装備されてもよい。また、車載システム1を自動運転システムとして構成してもよい。 The driving support system ECU includes, for example, Level 1 (driver assistance), Level 2 (partially automatic driving), and Level 3 (conditional automatic driving) at the automatic driving level presented by the American Society of Automotive Engineers of Japan (SAE). ) May be equipped with at least one of the functions. Further, the functions of level 4 (highly automatic driving) or level 5 (fully automatic driving) of the automatic driving level may be equipped, and only the functions of level 1 and 2 or only level 2 and 3 are equipped. May be good. Further, the in-vehicle system 1 may be configured as an automatic driving system.
 前記ボディ系ECUは、例えば、ドアロック、スマートキー、パワーウインドウ、エアコン、ライト、メーターパネル、又はウインカ等の車体の機能に関する制御ユニットを少なくとも1つ以上含んで構成され得る。
 前記情報系ECUは、例えば、インフォテイメント装置、テレマティクス装置、又はITS(Intelligent Transport Systems)関連装置を含んで構成され得る。前記インフォテイメント装置には、例えば、ユーザインターフェースとして機能するHMI(Human Machine Interface)装置の他、カーナビゲーション装置、オーディオ機器などが含まれてもよい。前記テレマティクス装置には、外部と通信するための通信ユニットなどが含まれてもよい。前記ITS関連装置には、ETC(Electronic Toll Collection System)、又はITSスポットなどの路側機との路車間通信、若しくは車々間通信などを行うための通信ユニットなどが含まれてもよい。
The body system ECU may be configured to include at least one control unit related to the function of the vehicle body such as a door lock, a smart key, a power window, an air conditioner, a light, an instrument panel, or a winker.
The information system ECU may be configured to include, for example, an infotainment device, a telematics device, or an ITS (Intelligent Transport Systems) related device. The infotainment device may include, for example, an HMI (Human Machine Interface) device that functions as a user interface, a car navigation device, an audio device, and the like. The telematics device may include a communication unit or the like for communicating with the outside. The ITS-related device may include an ETC (Electronic Toll Collection System), a communication unit for performing road-to-vehicle communication with a roadside machine such as an ITS spot, or vehicle-to-vehicle communication.
 センサ41には、ECU40でアクチュエータ42の動作制御を行うために必要となるセンシングデータを取得する各種の車載センサが含まれ得る。例えば、車速センサ、シフトポジションセンサ、アクセル開度センサ、ブレーキペダルセンサ、ステアリングセンサなどの他、車外撮像用カメラ、ミリ波等のレーダー(Radar)、ライダー(LIDER)、超音波センサなどの周辺監視センサなどが含まれてもよい。 The sensor 41 may include various in-vehicle sensors that acquire sensing data necessary for controlling the operation of the actuator 42 by the ECU 40. For example, in addition to vehicle speed sensors, shift position sensors, accelerator opening sensors, brake pedal sensors, steering sensors, etc., peripheral monitoring of external imaging cameras, millimeter-wave radar (Radar), riders (LIDER), ultrasonic sensors, etc. A sensor or the like may be included.
 アクチュエータ42は、ECU40からの制御信号に基づいて、車両2の走行、操舵、又は制動などに関わる動作を実行する装置であり、例えば、エンジン、モータ、トランスミッション、油圧又は電動シリンダー等が含まれる。 The actuator 42 is a device that executes operations related to traveling, steering, braking, etc. of the vehicle 2 based on a control signal from the ECU 40, and includes, for example, an engine, a motor, a transmission, a hydraulic cylinder, an electric cylinder, and the like.
[機能構成例]
 図3は、実施の形態に係るドライバモニタリング装置10の画像処理部12の機能構成例を示すブロック図である。
 画像処理部12は、画像入力部21、顔検出部22、特定個人判定部25、第1顔画像処理部26、第2顔画像処理部30、出力部34、及び顔特徴量記憶部142を含んで構成されている。
 画像入力部21は、カメラ11で撮像されたドライバ3の顔を含む画像を取り込む処理を行う。
[Functional configuration example]
FIG. 3 is a block diagram showing a functional configuration example of the image processing unit 12 of the driver monitoring device 10 according to the embodiment.
The image processing unit 12 includes an image input unit 21, a face detection unit 22, a specific individual determination unit 25, a first face image processing unit 26, a second face image processing unit 30, an output unit 34, and a face feature amount storage unit 142. It is configured to include.
The image input unit 21 performs a process of capturing an image including the face of the driver 3 captured by the camera 11.
 顔検出部22は、特定個人の顔検出部23と、通常の顔検出部24とを含んで構成され、入力画像から顔を検出するための特徴量を抽出しながら顔領域を検出する処理を行う。
 特定個人の顔検出部23は、顔特徴量記憶部142から読み込んだ特定個人の顔特徴量142aを用いて、入力画像から顔領域を検出する処理を行う。
 通常の顔検出部24は、顔特徴量記憶部142から読み込んだ通常の顔特徴量142bを用いて、入力画像から顔領域を検出する処理を行う。
The face detection unit 22 is configured to include a face detection unit 23 of a specific individual and a normal face detection unit 24, and performs a process of detecting a face region while extracting a feature amount for detecting a face from an input image. Do.
The face detection unit 23 of the specific individual uses the face feature amount 142a of the specific individual read from the face feature amount storage unit 142 to perform a process of detecting the face region from the input image.
The normal face detection unit 24 uses the normal face feature amount 142b read from the face feature amount storage unit 142 to perform a process of detecting a face region from an input image.
 画像から顔領域を検出する手法は特に限定されないが、高速で高精度に顔領域を検出する手法が採用される。顔検出部22は、例えば、入力画像に対して所定の探索領域(探索窓)を走査させながら、それぞれの探索領域で顔を検出するための特徴量を抽出する。顔検出部22は、例えば、顔の局所的な領域の明暗差(輝度差)、エッジ強度、又はこれら局所的領域間の関連性などを特徴量として抽出してよい。そして、顔検出部22は、探索領域から抽出した特徴量と、顔特徴量記憶部142から読み込んだ通常の顔特徴量142b、又は特定個人の顔特徴量142aを用いて、階層的な構造(顔をおおまかにとらえる階層から顔の細部をとらえる階層構造)の検出器で顔か非顔かを判断し、画像中から顔領域を検出する処理を行う。 The method of detecting the face area from the image is not particularly limited, but a method of detecting the face area at high speed and with high accuracy is adopted. The face detection unit 22 extracts, for example, a feature amount for detecting a face in each search area while scanning a predetermined search area (search window) for the input image. The face detection unit 22 may extract, for example, the difference in brightness (luminance difference) of a local region of the face, the edge strength, or the relationship between these local regions as a feature amount. Then, the face detection unit 22 uses the feature amount extracted from the search area, the normal face feature amount 142b read from the face feature amount storage unit 142, or the face feature amount 142a of a specific individual, and has a hierarchical structure ( A detector (hierarchical structure that captures the details of the face from the hierarchy that roughly captures the face) determines whether the face is face or non-face, and performs processing to detect the face area from the image.
 特定個人判定部25は、顔検出部22で検出された顔領域の特徴量と、顔特徴量記憶部142から読み込んだ特定個人の顔特徴量142aとを用いて、検出された顔領域の顔が特定個人の顔であるか否かを判定する処理を行う。
 特定個人判定部25は、顔領域から抽出された特徴量と特定個人の顔特徴量142aとの関係を示す指標、例えば、相関を示す指標として、相関係数を算出し、算出した相関係数に基づいて、顔領域の顔が特定個人の顔であるか否かを判定してもよい。例えば、顔領域内における1以上の局所的な領域のHaar-like特徴(輝度差)などの特徴量の相関を求めてもよい。そして、相関係数が所定の閾値より大きい場合、検出した顔領域の顔が特定個人の顔であると判定し、相関係数が所定の閾値以下の場合、検出した顔領域の顔が特定個人の顔ではないと判定してもよい。
The specific individual determination unit 25 uses the feature amount of the face area detected by the face detection unit 22 and the face feature amount 142a of the specific individual read from the face feature amount storage unit 142 to detect the face in the face area. Performs a process of determining whether or not is the face of a specific individual.
The specific individual determination unit 25 calculates a correlation coefficient as an index showing the relationship between the feature amount extracted from the face region and the face feature amount 142a of the specific individual, for example, an index showing the correlation, and the calculated correlation coefficient. It may be determined whether or not the face in the face region is the face of a specific individual based on. For example, the correlation of feature quantities such as Haar-like features (luminance difference) of one or more local regions in the face region may be obtained. Then, when the correlation coefficient is larger than a predetermined threshold value, it is determined that the face in the detected face area is the face of a specific individual, and when the correlation coefficient is equal to or less than the predetermined threshold value, the face in the detected face area is a specific individual. It may be determined that it is not the face of.
 また、特定個人判定部25は、特定個人の顔であるか否かを判定するために、特定個人の顔特徴量142aをパラメータに用いて機械学習を行った学習済みの学習器を備え、該学習済みの学習器に、顔領域から抽出された特徴量を入力し、前記学習済みの学習器の演算処理、例えば、特徴量の相関分析処理を行うことで、特定個人の顔であるか否かの判定情報を当該学習器から取得する構成としてもよい。前記学習済みの学習器は、非線形判別器を含んで構成してもよいし、線形判別器を含んで構成してもよい。例えば、前記学習済みの学習器は、サポートベクターマシンを含んで構成してもよいし、ニューラルネットワークを含んで構成してもよい。 Further, the specific individual determination unit 25 includes a learned learning device that has been machine-learned using the facial feature amount 142a of the specific individual as a parameter in order to determine whether or not the face is the face of the specific individual. By inputting the feature amount extracted from the face area into the learned learner and performing arithmetic processing of the learned learner, for example, correlation analysis processing of the feature amount, whether or not the face is a specific individual's face. The determination information may be acquired from the learning device. The learned learner may be configured to include a non-linear discriminator, or may be configured to include a linear discriminator. For example, the trained learner may be configured to include a support vector machine or may be configured to include a neural network.
 また、特定個人判定部25では、カメラ11からの入力画像の1フレームに対する判定の結果に基づいて、検出した顔領域の顔が特定個人の顔であるか否かを判定してもよいし、カメラ11からの入力画像の複数フレームに対する判定の結果に基づいて、検出した顔領域の顔が特定個人の顔であるか否かを判定してもよい。
 第1顔画像処理部26は、特定個人判定部25により特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う。第1顔画像処理部26は、特定個人の顔向き推定部27と、特定個人の目開閉検出部28と、特定個人の視線方向推定部29とを含んで構成されているが、さらに別の顔挙動を推定したり、検出したりする構成を含んでもよい。また、第1顔画像処理部26は、特定個人の顔特徴量142aを用いて、特定個人用の顔画像処理のいずれかの処理を行ってもよい。また、顔特徴量記憶部142に、特定個人用の顔画像処理を行うための機械学習を行った学習済みの特徴量を記憶しておき、該学習済みの特徴量を用いて、特定個人用の顔画像処理のいずれかの処理を行ってもよい。
Further, the specific individual determination unit 25 may determine whether or not the face in the detected face region is the face of the specific individual based on the result of determination for one frame of the input image from the camera 11. Based on the result of determination for a plurality of frames of the input image from the camera 11, it may be determined whether or not the face in the detected face region is the face of a specific individual.
When the specific individual determination unit 25 determines that the face is a specific individual's face, the first face image processing unit 26 performs face image processing for the specific individual. The first face image processing unit 26 includes a face orientation estimation unit 27 of a specific individual, an eye opening / closing detection unit 28 of the specific individual, and a line-of-sight direction estimation unit 29 of the specific individual, but is still different. It may include a configuration for estimating or detecting facial behavior. In addition, the first face image processing unit 26 may perform any processing of the face image processing for the specific individual by using the face feature amount 142a of the specific individual. Further, the face feature amount storage unit 142 stores the learned feature amount that has been machine-learned to perform the face image processing for the specific individual, and uses the learned feature amount for the specific individual. You may perform any processing of the face image processing of.
 特定個人の顔向き推定部27は、特定個人の顔の向きを推定する処理を行う。特定個人の顔向き推定部27は、例えば、特定個人の顔検出部23で検出された顔領域から目、鼻、口、眉などの顔器官の位置や形状を検出し、検出した顔器官の位置や形状に基づいて、顔の向きを推定する処理を行う。
 画像中の顔領域から顔器官を検出する手法は特に限定されないが、高速で高精度に顔器官を検出できる手法を採用することが好ましい。例えば、3次元顔形状モデルを作成し、これを2次元画像上の顔の領域にフィッティングさせ、顔の各器官の位置と形状を検出する手法が採用され得る。画像中の人の顔に3次元顔形状モデルをフィッティングさせる技術として、例えば、特開2007-249280号公報に記載された技術を適用することができるが、これに限定されるものではない。
The face orientation estimation unit 27 of the specific individual performs a process of estimating the face orientation of the specific individual. The face orientation estimation unit 27 of the specific individual detects, for example, the position and shape of facial organs such as eyes, nose, mouth, and eyebrows from the face region detected by the face detection unit 23 of the specific individual, and the detected facial organs. Performs processing to estimate the orientation of the face based on the position and shape.
The method for detecting the facial organs from the facial region in the image is not particularly limited, but it is preferable to adopt a method capable of detecting the facial organs at high speed and with high accuracy. For example, a method of creating a three-dimensional face shape model, fitting it to a face region on a two-dimensional image, and detecting the position and shape of each organ of the face can be adopted. As a technique for fitting a three-dimensional face shape model to a human face in an image, for example, the technique described in Japanese Patent Application Laid-Open No. 2007-249280 can be applied, but the technique is not limited thereto.
 また、特定個人の顔向き推定部27は、特定個人の顔の向きの推定データとして、例えば、上記3次元顔形状モデルのパラメータに含まれている、上下回転(X軸回り)のピッチ角、左右回転(Y軸回り)のヨー角、及び全体回転(Z軸回り)のロール角を出力してもよい。
 特定個人の目開閉検出部28は、特定個人の目の開閉状態を検出する処理を行う。特定個人の目開閉検出部28は、例えば、特定個人の顔向き推定部27で求めた顔器官の位置や形状、特に目の特徴点(瞼、瞳孔)の位置や形状に基づいて、目の開閉状態、例えば、目を開けているか、閉じているかを検出する。目の開閉状態は、例えば、さまざまな目の開閉状態における目の画像の特徴量(瞼の位置、瞳孔(黒目)の形状、又は、白目部分と黒目部分の領域サイズなど)を予め学習器を用いて学習し、これら学習済みの特徴量データとの類似度を評価することで検出してもよい。
Further, the face orientation estimation unit 27 of the specific individual can use the estimation data of the face orientation of the specific individual, for example, the pitch angle of vertical rotation (around the X axis) included in the parameters of the three-dimensional face shape model. The yaw angle of the left-right rotation (around the Y axis) and the roll angle of the entire rotation (around the Z axis) may be output.
The eye opening / closing detection unit 28 of the specific individual performs a process of detecting the opening / closing state of the eyes of the specific individual. The eye opening / closing detection unit 28 of the specific individual, for example, is based on the position and shape of the facial organs obtained by the face orientation estimation unit 27 of the specific individual, particularly the position and shape of the feature points (eyelids, pupils) of the eyes. Detects the open / closed state, for example, whether the eyes are open or closed. For the open / closed state of the eye, for example, the feature amount of the image of the eye (the position of the eyelid, the shape of the pupil (black eye), the area size of the white eye part and the black eye part, etc.) in various open / closed states of the eye is previously learned. It may be detected by learning using the data and evaluating the degree of similarity with the learned feature data.
 特定個人の視線方向推定部29は、特定個人の視線の方向を推定する処理を行う。特定個人の視線方向推定部29は、例えば、ドライバ3の顔の向き、及びドライバ3の顔器官の位置や形状、特に目の特徴点(目尻、目頭、瞳孔)の位置や形状に基づいて、視線の方向を推定する。視線の方向とは、ドライバ3が見ている方向のことであり、例えば、顔の向きと目の向きとの組み合わせによって求められる。 The line-of-sight direction estimation unit 29 of the specific individual performs a process of estimating the line-of-sight direction of the specific individual. The line-of-sight direction estimation unit 29 of a specific individual is based on, for example, the orientation of the face of the driver 3 and the position and shape of the facial organs of the driver 3, particularly the position and shape of the feature points of the eyes (outer corners of eyes, inner corners of eyes, pupils). Estimate the direction of the line of sight. The direction of the line of sight is the direction in which the driver 3 is looking, and is determined by, for example, a combination of the direction of the face and the direction of the eyes.
 また、視線の方向は、例えば、さまざまな顔の向きと目の向きとの組み合わせにおける目の画像の特徴量(目尻、目頭、瞳孔の相対位置、又は白目部分と黒目部分の相対位置、濃淡、テクスチャーなど)とを予め学習器を用いて学習し、これら学習した特徴量データとの類似度を評価することで検出してもよい。また、特定個人の視線方向推定部29は、前記3次元顔形状モデルのフィッティング結果などを用いて、顔の大きさや向きと目の位置などから眼球の大きさと中心位置とを推定するとともに、瞳孔の位置を検出し、眼球の中心と瞳孔の中心とを結ぶベクトルを視線方向として検出してもよい。 In addition, the direction of the line of sight is, for example, the feature amount of the image of the eye in various combinations of face orientation and eye orientation (relative position of outer corner, inner corner of eye, pupil, relative position of white eye portion and black eye portion, shading, etc. (Texture, etc.) may be detected by learning in advance using a learning device and evaluating the degree of similarity with the learned feature amount data. In addition, the line-of-sight direction estimation unit 29 of the specific individual estimates the size and center position of the eyeball from the size and orientation of the face, the position of the eyes, etc., using the fitting result of the three-dimensional face shape model, and the pupil. The position of the eyeball may be detected, and the vector connecting the center of the eyeball and the center of the pupil may be detected as the line-of-sight direction.
 第2顔画像処理部30は、特定個人判定部25により特定個人の顔ではないと判定された場合、通常の顔画像処理を行う。第2顔画像処理部30は、通常の顔向き推定部31と、通常の目開閉検出部32と、通常の視線方向推定部33とを含んで構成されているが、さらに別の顔挙動を推定したり、検出したりする構成を含んでもよい。また、第2顔画像処理部30は、通常の顔特徴量142bを用いて、通常の顔画像処理のいずれかの処理を行ってもよい。また、顔特徴量記憶部142に、通常の顔画像処理を行うための機械学習を行った学習済みの特徴量を記憶しておき、該学習済みの特徴量を用いて、通常の顔画像処理のいずれかの処理を行ってもよい。なお、通常の顔向き推定部31と、通常の目開閉検出部32と、通常の視線方向推定部33とで行われる処理は、特定個人の顔向き推定部27と、特定個人の目開閉検出部28と、特定個人の視線方向推定部29と基本的に同様であるので、ここではその説明を省略する。 When the specific individual determination unit 25 determines that the face is not the face of a specific individual, the second face image processing unit 30 performs normal face image processing. The second face image processing unit 30 includes a normal face orientation estimation unit 31, a normal eye opening / closing detection unit 32, and a normal line-of-sight direction estimation unit 33, but has yet another face behavior. It may include a configuration for estimating or detecting. In addition, the second face image processing unit 30 may perform any processing of the normal face image processing using the normal face feature amount 142b. Further, the face feature amount storage unit 142 stores the learned feature amount obtained by machine learning for performing the normal face image processing, and the learned feature amount is used for the normal face image processing. Any of the above processes may be performed. The processing performed by the normal face orientation estimation unit 31, the normal eye opening / closing detection unit 32, and the normal line-of-sight direction estimation unit 33 includes the face orientation estimation unit 27 of the specific individual and the eye opening / closing detection of the specific individual. Since it is basically the same as the unit 28 and the line-of-sight direction estimation unit 29 of a specific individual, the description thereof will be omitted here.
 出力部34は、画像処理部12による画像処理に基づく情報をECU40などに出力する処理を行う。画像処理に基づく情報は、例えば、ドライバ3の顔の向き、視線の方向、又は目の開閉状態などの顔の挙動に関する情報でもよいし、顔の挙動の検出結果に基づいて判定されたドライバ3の状態(例えば、前方注視、脇見、居眠り、後ろ向き、突っ伏しなどの状態)に関する情報でもよい。また、画像処理に基づく情報は、ドライバ3の状態判定に基づく、所定の制御信号(注意や警告処理を行うための制御信号、又は車両2の動作制御を行うための制御信号など)でもよい。 The output unit 34 performs a process of outputting information based on the image processing by the image processing unit 12 to the ECU 40 or the like. The information based on the image processing may be, for example, information on the behavior of the face such as the direction of the face of the driver 3, the direction of the line of sight, or the open / closed state of the eyes, or the driver 3 determined based on the detection result of the behavior of the face. Information on the state of (for example, forward gaze, inattentiveness, dozing, backward facing, prone, etc.) may be used. Further, the information based on the image processing may be a predetermined control signal (control signal for performing caution or warning processing, control signal for performing operation control of the vehicle 2, etc.) based on the state determination of the driver 3.
[処理動作例]
 図4は、実施の形態に係るドライバモニタリング装置10における画像処理部12のCPU13が行う処理動作の一例を示すフローチャートである。カメラ11では、例えば、毎秒数十フレームの画像が撮像され、各フレーム、又は一定間隔のフレーム毎に本処理が行われる。
 まず、ステップS1では、CPU13は、画像入力部21として動作し、カメラ11で撮像された画像(ドライバ3の顔を含む画像)を読み込む処理を行い、ステップS2に処理を進める。
 ステップS2では、CPU13は、通常の顔検出部24として動作し、入力画像に対して通常の顔検出処理を行い、ステップS3に処理を進める。CPU13は、例えば、入力画像に対して所定の探索領域(探索窓)を走査させながら、それぞれの探索領域で顔を検出するための特徴量を抽出する。そして、CPU13は、探索領域から抽出した特徴量と、顔特徴量記憶部142から読み込んだ通常の顔特徴量142bとを用いて、顔か非顔かを判断し、画像中から顔領域を検出する処理を行う。
[Processing operation example]
FIG. 4 is a flowchart showing an example of a processing operation performed by the CPU 13 of the image processing unit 12 in the driver monitoring device 10 according to the embodiment. For example, the camera 11 captures an image of several tens of frames per second, and this processing is performed for each frame or every frame at regular intervals.
First, in step S1, the CPU 13 operates as an image input unit 21, performs a process of reading an image (an image including the face of the driver 3) captured by the camera 11, and proceeds to step S2.
In step S2, the CPU 13 operates as a normal face detection unit 24, performs normal face detection processing on the input image, and proceeds to step S3. For example, the CPU 13 scans a predetermined search area (search window) for an input image and extracts a feature amount for detecting a face in each search area. Then, the CPU 13 determines whether the face is a face or a non-face by using the feature amount extracted from the search area and the normal face feature amount 142b read from the face feature amount storage unit 142, and detects the face area from the image. Perform the process of
 ステップS3では、CPU13は、特定個人の顔検出部23として動作し、入力画像に対して特定個人の顔検出処理を行い、ステップS4に処理を進める。CPU13は、例えば、入力画像に対して所定の探索領域(探索窓)を走査させながら、それぞれの探索領域で顔を検出するための特徴量を抽出する。そして、CPU13は、探索領域から抽出した特徴量と、顔特徴量記憶部142から読み込んだ特定個人の顔特徴量142aとを用いて、顔か非顔かを判断し、画像中から顔領域を検出する処理を行う。なお、ステップS2とS3の処理は、1つのステップ内で並列的に行ってもよいし、組み合わせて行ってもよい。 In step S3, the CPU 13 operates as the face detection unit 23 of the specific individual, performs face detection processing of the specific individual on the input image, and proceeds to step S4. For example, the CPU 13 scans a predetermined search area (search window) for an input image and extracts a feature amount for detecting a face in each search area. Then, the CPU 13 determines whether the face is a face or a non-face by using the feature amount extracted from the search area and the face feature amount 142a of the specific individual read from the face feature amount storage unit 142, and determines the face area from the image. Perform the detection process. The processes of steps S2 and S3 may be performed in parallel in one step or may be performed in combination.
 ステップS4では、CPU13は、特定個人判定部25として動作し、ステップS2、3で検出された顔領域の特徴量と、顔特徴量記憶部142から読み込んだ特定個人の顔特徴量142aとを用いて、顔領域の顔が特定個人の顔であるか否かを判定する処理を行い、ステップS5に処理を進める。
 ステップS5では、CPU13は、ステップS4での判定処理の結果が、特定個人の顔であるか否かを判断し、特定個人の顔であると判断すれば、ステップS6に処理を進める。
In step S4, the CPU 13 operates as the specific individual determination unit 25, and uses the feature amount of the face region detected in steps S2 and S3 and the face feature amount 142a of the specific individual read from the face feature amount storage unit 142. Then, a process of determining whether or not the face in the face area is a face of a specific individual is performed, and the process proceeds to step S5.
In step S5, the CPU 13 determines whether or not the result of the determination process in step S4 is the face of a specific individual, and if it is determined that the result is the face of a specific individual, the process proceeds to step S6.
 ステップS6では、CPU13は、特定個人の顔向き推定部27として動作し、例えば、ステップS3で検出した顔領域から目、鼻、口、眉などの顔器官の位置や形状を検出し、検出した顔器官の位置や形状に基づいて、顔の向きを推定し、ステップS7に処理を進める。
 ステップS7では、CPU13は、特定個人の目開閉検出部28として動作し、例えば、ステップS6で求めた顔器官の位置や形状、特に目の特徴点(瞼、瞳孔)の位置や形状に基づいて、目の開閉状態、例えば、目を開けているか、閉じているかを検出し、ステップS8に処理を進める。
In step S6, the CPU 13 operates as a face orientation estimation unit 27 of a specific individual, and for example, detects and detects the position and shape of facial organs such as eyes, nose, mouth, and eyebrows from the face area detected in step S3. The orientation of the face is estimated based on the position and shape of the facial organ, and the process proceeds to step S7.
In step S7, the CPU 13 operates as an eye opening / closing detection unit 28 for a specific individual, and is based on, for example, the position and shape of the facial organs obtained in step S6, particularly the position and shape of eye feature points (eyelids, pupils). , The open / closed state of the eyes, for example, whether the eyes are open or closed is detected, and the process proceeds to step S8.
 ステップS8では、CPU13は、特定個人の視線方向推定部29として動作し、例えば、ステップS6で求めた顔の向き、顔器官の位置や形状、特に目の特徴点(目尻、目頭、瞳孔)の位置や形状に基づいて、視線の方向を推定し、その後処理を終える。
 一方ステップS5において、CPU13は、特定個人の顔ではない、換言すれば、通常の顔であると判断すれば、ステップS9に処理を進める。
In step S8, the CPU 13 operates as a line-of-sight direction estimation unit 29 of a specific individual, and for example, the orientation of the face, the position and shape of the facial organs obtained in step S6, particularly the feature points of the eyes (outer corners of eyes, inner corners of eyes, pupils). The direction of the line of sight is estimated based on the position and shape, and then the process is finished.
On the other hand, in step S5, if the CPU 13 determines that it is not the face of a specific individual, in other words, it is a normal face, the process proceeds to step S9.
 ステップS9では、CPU13は、通常の顔向き推定部31として動作し、例えば、ステップS2で検出した顔領域から目、鼻、口、眉などの顔器官の位置や形状を検出し、検出した顔器官の位置や形状に基づいて、顔の向きを推定し、ステップS10に処理を進める。
 ステップS10では、CPU13は、通常の目開閉検出部32として動作し、例えば、ステップS9で求めた顔器官の位置や形状、特に目の特徴点(瞼、瞳孔)の位置や形状に基づいて、目の開閉状態、例えば、目を開けているか、閉じているかを検出し、ステップS11に処理を進める。
In step S9, the CPU 13 operates as a normal face orientation estimation unit 31, and for example, detects the position and shape of facial organs such as eyes, nose, mouth, and eyebrows from the face area detected in step S2, and detects the face. The orientation of the face is estimated based on the position and shape of the organ, and the process proceeds to step S10.
In step S10, the CPU 13 operates as a normal eye opening / closing detection unit 32, and for example, based on the position and shape of the facial organs obtained in step S9, particularly the position and shape of the feature points (eyelids, pupils) of the eyes. The open / closed state of the eyes, for example, whether the eyes are open or closed is detected, and the process proceeds to step S11.
 ステップS11では、CPU13は、通常の視線方向推定部33として動作し、例えば、ステップS9で求めた顔の向き、顔器官の位置や形状、特に目の特徴点(目尻、目頭、瞳孔)の位置や形状に基づいて、視線の方向を推定し、その後処理を終える。
 図5は、実施の形態に係るドライバモニタリング装置10における画像処理部12のCPU13が行う特定個人判定処理動作の一例を示すフローチャートである。本処理動作は、図4に示すステップS4における特定個人判定処理動作の一例であり、入力画像1枚(1フレーム)で判定する場合の処理動作例である。
In step S11, the CPU 13 operates as a normal line-of-sight direction estimation unit 33, and for example, the orientation of the face and the position and shape of the facial organs obtained in step S9, particularly the positions of the feature points of the eyes (outer corners of eyes, inner corners of eyes, pupils). The direction of the line of sight is estimated based on the shape and shape, and then the process is completed.
FIG. 5 is a flowchart showing an example of a specific individual determination processing operation performed by the CPU 13 of the image processing unit 12 in the driver monitoring device 10 according to the embodiment. This processing operation is an example of the specific individual determination processing operation in step S4 shown in FIG. 4, and is an example of the processing operation in the case of determining with one input image (1 frame).
 まず、ステップS21では、CPU13は、図4に示すステップS2、S3の顔検出処理で検出された顔領域から抽出された特徴量を読み込み、次のステップS22では、顔特徴量記憶部142から学習済みの特定個人の顔特徴量142aを読み込み、ステップS23に処理を進める。
 ステップS23では、CPU13は、ステップS21で読み込んだ顔領域から抽出された特徴量と、ステップS22で読み込んだ特定個人の顔特徴量142aとの相関係数を算出する処理を行い、ステップS24に処理を進める。
First, in step S21, the CPU 13 reads the feature amount extracted from the face area detected by the face detection processes of steps S2 and S3 shown in FIG. 4, and in the next step S22, learns from the face feature amount storage unit 142. The completed facial feature amount 142a of the specific individual is read, and the process proceeds to step S23.
In step S23, the CPU 13 performs a process of calculating the correlation coefficient between the feature amount extracted from the face area read in step S21 and the face feature amount 142a of the specific individual read in step S22, and processes in step S24. To proceed.
 ステップS24では、CPU13は、算出した相関係数が、特定個人か否かを判定するための所定の閾値より大きいか否かを判断し、相関係数が所定の閾値よりも大きい、換言すれば、顔領域から抽出された特徴量と、特定個人の顔特徴量142aとの相関性が高い(換言すれば、類似度が高い)と判断すれば、ステップS25に処理を進める。
 ステップS25では、CPU13は、顔領域に検出された顔が特定個人の顔であると判定し、その後処理を終える。
In step S24, the CPU 13 determines whether or not the calculated correlation coefficient is larger than a predetermined threshold value for determining whether or not the individual is a specific individual, and the correlation coefficient is larger than the predetermined threshold value, in other words. If it is determined that the feature amount extracted from the face area and the face feature amount 142a of the specific individual have a high correlation (in other words, the similarity is high), the process proceeds to step S25.
In step S25, the CPU 13 determines that the face detected in the face area is the face of a specific individual, and then ends the process.
 一方ステップS24において、相関係数が所定の閾値以下である、換言すれば、顔領域から抽出された特徴量と、特定個人の顔特徴量142aとの相関性が低い(換言すれば、類似度が低い)と判断すれば、ステップS26に処理を進める。
 ステップS26では、CPU13は、特定個人の顔ではない、換言すれば、通常の顔であると判定し、その後処理を終える。
On the other hand, in step S24, the correlation coefficient is equal to or less than a predetermined threshold value, in other words, the correlation between the feature amount extracted from the face region and the face feature amount 142a of the specific individual is low (in other words, the degree of similarity). Is low), the process proceeds to step S26.
In step S26, the CPU 13 determines that the face is not a specific individual's face, in other words, a normal face, and then ends the process.
 図6は、実施の形態に係るドライバモニタリング装置10における画像処理部12のCPU13が行う特定個人判定処理動作の一例を示すフローチャートである。
 本処理動作は、図4に示すステップS4における特定個人判定処理動作の別の一例であり、複数の入力画像(複数フレーム)で判定する場合の処理動作例である。
 まず、ステップS31では、CPU13は、特定個人の顔が検出された画像のカウンタ(numSp)を0にセットし、ステップS32では、判定する入力画像のカウンタ(i)を0にセットしてステップS33に処理を進める。
FIG. 6 is a flowchart showing an example of a specific individual determination processing operation performed by the CPU 13 of the image processing unit 12 in the driver monitoring device 10 according to the embodiment.
This processing operation is another example of the specific individual determination processing operation in step S4 shown in FIG. 4, and is a processing operation example in the case of determining with a plurality of input images (multiple frames).
First, in step S31, the CPU 13 sets the counter (numSp) of the image in which the face of a specific individual is detected to 0, and in step S32, the counter (i) of the input image to be determined is set to 0 and step S33. Proceed to processing.
 ステップS33では、CPU13は、入力画像1枚(1フレーム)に対する特定個人の判定処理(例えば、図5に示したステップS21~S26の処理)を行い、ステップS34に処理を進める。
 ステップS34では、CPU13は、ステップS33の判定処理の結果が、特定個人の顔であったか否かを判断し、特定個人の顔であったと判断すれば、ステップS35に処理を進める。
In step S33, the CPU 13 performs a specific individual determination process (for example, the processes of steps S21 to S26 shown in FIG. 5) for one input image (1 frame), and proceeds to step S34.
In step S34, the CPU 13 determines whether or not the result of the determination process in step S33 is the face of a specific individual, and if it is determined that the result is the face of a specific individual, the process proceeds to step S35.
 ステップS35では、CPU13は、特定個人の顔画像のカウンタ(numSp)に1を加算して、ステップS36に処理を進め、ステップS36では、入力画像のカウンタ(i)に1を加算して、ステップS37に処理を進める。
 一方ステップS34において、CPU13は、特定個人の顔ではなかったと判断すれば、ステップS36に処理を進め、ステップS36では、入力画像のカウンタ(i)に1を加算して、ステップS37に処理を進める。すなわち、この場合、特定個人の顔画像のカウンタ(numSp)に1は加算されない。
In step S35, the CPU 13 adds 1 to the counter (numSp) of the face image of the specific individual and proceeds with the process in step S36, and in step S36, adds 1 to the counter (i) of the input image and steps. The process proceeds to S37.
On the other hand, in step S34, if the CPU 13 determines that the face is not the face of a specific individual, the process proceeds to step S36, and in step S36, 1 is added to the counter (i) of the input image and the process proceeds to step S37. .. That is, in this case, 1 is not added to the counter (numSp) of the face image of the specific individual.
 ステップS37では、CPU13は、入力画像のカウンタ(i)が、所定の画像枚数N未満か否かを判断し、カウンタ(i)が、N未満である(換言すれば、所定の画像枚数(N枚)の判定を終えていない)と判定すれば、ステップS33に戻り、次の入力画像での特定個人の判定処理を繰り返す。
 一方ステップS37において、CPU13は、入力画像のカウンタ(i)が、N未満ではない(換言すれば、所定の画像枚数(N枚)の判定を終えた)と判定すれば、ステップS38に処理を進める。
In step S37, the CPU 13 determines whether or not the input image counter (i) is less than the predetermined number of images N, and the counter (i) is less than N (in other words, the predetermined number of images (N). If it is determined that the determination of (sheets) has not been completed), the process returns to step S33, and the determination process of the specific individual in the next input image is repeated.
On the other hand, in step S37, if the CPU 13 determines that the counter (i) of the input image is not less than N (in other words, the determination of the predetermined number of images (N) has been completed), the processing is performed in step S38. Proceed.
 ステップS38では、CPU13は、特定個人の顔画像のカウンタ(numSp)が、特定個人であると判定するための所定の閾値より大きいか否かを判断する。所定の閾値は、例えば、判定に使用する入力画像の数をN枚(Nフレーム)とした場合、N/2に設定してもよいし、N/2よりも大きい値に設定してもよい。
 ステップS38において、CPU13は、特定個人の顔画像のカウンタ(numSp)が、所定の閾値より大きいと判断すれば、ステップS39に処理を進め、ステップS39では、特定個人の顔であると判定し、特定個人の判定処理を終え、その後、特定個人用の顔画像処理を行う。
In step S38, the CPU 13 determines whether or not the counter (numSp) of the face image of the specific individual is larger than a predetermined threshold value for determining the specific individual. The predetermined threshold value may be set to N / 2 or a value larger than N / 2 when the number of input images used for determination is N (N frames), for example. ..
In step S38, if the CPU 13 determines that the face image counter (numSp) of the specific individual is larger than the predetermined threshold value, the process proceeds to step S39, and in step S39, it is determined that the face of the specific individual is the face of the specific individual. After finishing the determination process of the specific individual, the face image processing for the specific individual is performed.
 一方ステップS38において、CPU13は、特定個人顔画像のカウンタ(numSp)が、所定の閾値以下であると判断すれば、ステップS40に処理を進め、ステップS40では、特定個人の顔ではない(換言すれば、通常の顔である)と判定し、特定個人の判定処理を終え、その後、通常の顔画像処理を行う。
 上記した実施の形態に係るドライバモニタリング装置10によれば、顔特徴量記憶部142に学習済みの顔特徴量として、特定個人の顔特徴量142aと、通常の顔特徴量142bとが記憶され、特定個人判定部25により、顔検出部22で検出された顔領域の特徴量と、特定個人の顔特徴量142aとを用いて、顔領域の顔が特定個人の顔であるか否かが判定される。したがって、特定個人の顔特徴量142aを用いることにより、特定個人の顔であるか否かを精度良く判定することができるとともに、該判定処理にかかる負荷を低減することができる。
On the other hand, in step S38, if the CPU 13 determines that the counter (numSp) of the specific individual face image is equal to or less than a predetermined threshold value, the process proceeds to step S40, and in step S40, the face is not the face of the specific individual (in other words, For example, it is a normal face), the judgment process of a specific individual is completed, and then the normal face image processing is performed.
According to the driver monitoring device 10 according to the above-described embodiment, the face feature amount 142a of a specific individual and the normal face feature amount 142b are stored as the learned face feature amount in the face feature amount storage unit 142. The specific individual determination unit 25 determines whether or not the face in the face region is the face of a specific individual by using the feature amount of the face region detected by the face detection unit 22 and the face feature amount 142a of the specific individual. Will be done. Therefore, by using the face feature amount 142a of the specific individual, it is possible to accurately determine whether or not the face is the face of the specific individual, and it is possible to reduce the load on the determination process.
 また、特定個人判定部25により特定個人の顔であると判定された場合、第1顔画像処理部26により特定個人の顔画像処理を精度良く実施することができる。一方、特定個人判定部25により特定個人の顔ではない、換言すれば、通常の顔(特定個人以外の人の顔)であると判定された場合、第2顔画像処理部30により通常の顔画像処理を精度良く実施することができる。したがって、ドライバ3が、特定個人であっても、特定個人以外の通常の人であっても、それぞれの顔のセンシングを精度良く実施することができる。 Further, when the specific individual determination unit 25 determines that the face is a specific individual, the first face image processing unit 26 can accurately perform the face image processing of the specific individual. On the other hand, when the specific individual determination unit 25 determines that the face is not a specific individual's face, in other words, a normal face (a face of a person other than the specific individual), the second face image processing unit 30 determines that the face is a normal face. Image processing can be performed with high accuracy. Therefore, whether the driver 3 is a specific individual or an ordinary person other than the specific individual, it is possible to accurately perform sensing of each face.
 また、特定個人判定部25が、顔領域から抽出された特徴量と特定個人の顔特徴量142aとの相関を示す指標として、相関係数を算出し、算出した前記相関係数に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する。これにより、前記相関係数に基づいて前記顔領域の顔が前記特定個人の顔であるか否かを効率良く判定することができ、また、前記相関係数と所定の閾値とを比較する処理により、前記判定の処理効率を更に高めることができる。 Further, the specific individual determination unit 25 calculates a correlation coefficient as an index showing the correlation between the feature amount extracted from the face region and the face feature amount 142a of the specific individual, and based on the calculated correlation coefficient, It is determined whether or not the face in the face region is the face of the specific individual. As a result, it is possible to efficiently determine whether or not the face in the face region is the face of the specific individual based on the correlation coefficient, and a process of comparing the correlation coefficient with a predetermined threshold value. Therefore, the processing efficiency of the determination can be further improved.
 また、車載システム1が、ドライバモニタリング装置10と、ドライバモニタリング装置10から出力されるモニタリングの結果に基づいて、所定の処理を実行する1以上のECU40とを備えている。したがって、前記モニタリングの結果に基づいて、ECU40に所定の制御を適切に実行させることが可能となる。これにより、特定個人であっても安心して運転することができる安全性の高い車載システムを構築することが可能となる。 Further, the in-vehicle system 1 includes a driver monitoring device 10 and one or more ECUs 40 that execute a predetermined process based on the monitoring result output from the driver monitoring device 10. Therefore, based on the result of the monitoring, the ECU 40 can appropriately execute a predetermined control. This makes it possible to construct a highly safe in-vehicle system that allows even a specific individual to drive with peace of mind.
 以上、本発明の実施の形態を詳細に説明したが、前述までの説明はあらゆる点において本発明の例示に過ぎない。本発明の範囲を逸脱することなく、種々の改良や変更を行うことができることは言うまでもない。
 上記実施の形態では、本発明に係る画像処理装置をドライバモニタリング装置10に適用した場合について説明したが、適用例はこれに限定されない。例えば、工場内の機械や装置などの各種設備を操作したり、監視したり、所定の作業をしたりする人などをモニタリングする装置やシステムなどにおいて、モニタリング対象者に上記した特定個人が含まれる場合に、本発明に係る画像処理装置を適用可能である。
Although the embodiments of the present invention have been described in detail above, the above description is merely an example of the present invention in all respects. Needless to say, various improvements and changes can be made without departing from the scope of the present invention.
In the above embodiment, the case where the image processing device according to the present invention is applied to the driver monitoring device 10 has been described, but the application example is not limited to this. For example, in a device or system for monitoring a person who operates, monitors, or performs a predetermined work of various facilities such as machines and devices in a factory, the above-mentioned specific individual is included in the monitoring target person. In some cases, the image processing apparatus according to the present invention can be applied.
[付記]
 本発明の実施の形態は、以下の付記の様にも記載され得るが、これらに限定されない。
(付記1)
 撮像部(11)から入力される画像を処理する画像処理装置(12)であって、
 前記画像から顔を検出するための学習を行った学習済みの顔特徴量として、特定個人の顔特徴量(142a)と、通常の顔特徴量(142b)とが記憶される顔特徴量記憶部(142)と、
 前記画像から顔を検出するための特徴量を抽出しながら顔領域を検出する顔検出部(22)と、
 検出された前記顔領域の前記特徴量と、前記特定個人の顔特徴量(142a)とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定部(25)と、
 該特定個人判定部(25)により前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理部(26)と、
 前記特定個人判定部(25)により前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理部(30)とを備えていることを特徴とする画像処理装置。
[Additional Notes]
Embodiments of the present invention may also be described as, but are not limited to, the following appendices.
(Appendix 1)
An image processing device (12) that processes an image input from the image pickup unit (11).
A face feature storage unit that stores a specific individual's face feature (142a) and a normal face feature (142b) as learned face features that have been learned to detect a face from the image. (142) and
A face detection unit (22) that detects a face region while extracting a feature amount for detecting a face from the image, and a face detection unit (22).
A specific individual determination unit that determines whether or not the face in the face region is the face of the specific individual by using the detected feature amount of the face region and the face feature amount (142a) of the specific individual. (25) and
When the specific individual determination unit (25) determines that the face is the specific individual's face, the first face image processing unit (26) that performs face image processing for the specific individual and
Image processing characterized by including a second face image processing unit (30) that performs normal face image processing when the specific individual determination unit (25) determines that the face is not the face of the specific individual. apparatus.
(付記2)
 撮像部(11)と、記憶部(14)と、
 少なくとも1以上のプロセッサ(13)と、を備えた装置を用い、
 前記撮像部(11)から入力される画像を処理する画像処理方法であって、
 前記記憶部(14)が、
 前記画像から顔を検出するための学習を行った学習済みの顔特徴量として、特定個人の顔特徴量(142a)と、通常の顔特徴量(142b)とが記憶される顔特徴量記憶部(142)を備え、
 前記プロセッサ(13)が、
 前記画像から顔の特徴量を抽出しながら顔領域を検出する顔検出ステップ(S2、S3)と、
 該顔検出ステップ(S2、S3)により検出された前記顔領域の前記特徴量と、前記顔特徴量記憶部(142)から読み込んだ前記特定個人の顔特徴量(142a)とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定ステップ(S4)と、
 該特定個人判定ステップ(S4)により前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理ステップ(S6、S7、S8)と、
 前記特定個人判定ステップ(S4)により前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理ステップ(S9、S10、S11)とを含むことを特徴とする画像処理方法。
(Appendix 2)
The imaging unit (11), the storage unit (14),
Using a device with at least one or more processors (13)
An image processing method for processing an image input from the imaging unit (11).
The storage unit (14)
A face feature storage unit that stores a specific individual's face feature (142a) and a normal face feature (142b) as learned face features that have been learned to detect a face from the image. With (142)
The processor (13)
Face detection steps (S2, S3) for detecting a face region while extracting facial features from the image, and
Using the feature amount of the face region detected by the face detection steps (S2, S3) and the face feature amount (142a) of the specific individual read from the face feature amount storage unit (142), the said A specific individual determination step (S4) for determining whether or not the face in the face region is the face of the specific individual, and
When the face of the specific individual is determined by the specific individual determination step (S4), the first face image processing step (S6, S7, S8) for performing the face image processing for the specific individual and
When it is determined by the specific individual determination step (S4) that the face is not the specific individual's face, it includes a second face image processing step (S9, S10, S11) for performing normal face image processing. Image processing method.
 1   車載システム
 2   車両
 3   ドライバ
10   ドライバモニタリング装置
11   カメラ
12   画像処理部
13   CPU
14   ROM
141  プログラム記憶部
142  顔特徴量記憶部
142a 特定個人の顔特徴量
142b 通常の顔特徴量
15   RAM
151  画像メモリ
16   通信部
21   画像入力部
22   顔検出部
23   特定個人の顔検出部
24   通常の顔検出部
25   特定個人判定部
26   第1顔画像処理部
27   特定個人の顔向き推定部
28   特定個人の目開閉検出部
29   特定個人の視線方向推定部
30   第2顔画像処理部
31   通常の顔向き推定部
32   通常の目開閉検出部
33   通常の視線方向推定部
34   出力部
40   ECU
41   センサ
42   アクチュエータ
43   通信バス
1 In-vehicle system 2 Vehicle 3 Driver 10 Driver monitoring device 11 Camera 12 Image processing unit 13 CPU
14 ROM
141 Program storage unit 142 Face feature amount Storage unit 142a Specific individual face feature amount 142b Normal face feature amount 15 RAM
151 Image memory 16 Communication unit 21 Image input unit 22 Face detection unit 23 Specific individual face detection unit 24 Normal face detection unit 25 Specific individual judgment unit 26 First face image processing unit 27 Specific individual face orientation estimation unit 28 Specific individual Eye opening / closing detection unit 29 Line-of-sight direction estimation unit 30 of a specific individual Second face image processing unit 31 Normal face orientation estimation unit 32 Normal eye opening / closing detection unit 33 Normal line-of-sight direction estimation unit 34 Output unit 40 ECU
41 Sensor 42 Actuator 43 Communication bus

Claims (11)

  1.  撮像部から入力される画像を処理する画像処理装置であって、
     前記画像から顔を検出するための学習を行った学習済みの顔特徴量として、特定個人の顔特徴量と、通常の顔特徴量とが記憶される顔特徴量記憶部と、
     前記画像から顔を検出するための特徴量を抽出しながら顔領域を検出する顔検出部と、
     検出された前記顔領域の前記特徴量と、前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定部と、
     該特定個人判定部により前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理部と、
     前記特定個人判定部により前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理部と、
    を備えている画像処理装置。
    An image processing device that processes images input from the image pickup unit.
    As the learned facial features that have been learned to detect the face from the image, a facial feature storage unit that stores the facial features of a specific individual and the normal facial features, and
    A face detection unit that detects a face region while extracting a feature amount for detecting a face from the image, and a face detection unit.
    Using the detected feature amount of the face region and the face feature amount of the specific individual, a specific individual determination unit for determining whether or not the face in the face region is the face of the specific individual.
    When the specific individual determination unit determines that the face is the face of the specific individual, the first face image processing unit that performs face image processing for the specific individual and the first face image processing unit
    When the specific individual determination unit determines that the face is not the specific individual's face, the second face image processing unit that performs normal face image processing and
    An image processing device equipped with.
  2.  前記特定個人判定部が、
     前記顔領域から抽出された前記特徴量と前記特定個人の顔特徴量との相関を示す指標を算出し、
     算出した前記指標に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する、
    請求項1記載の画像処理装置。
    The specific individual judgment unit
    An index showing the correlation between the feature amount extracted from the face region and the facial feature amount of the specific individual was calculated.
    Based on the calculated index, it is determined whether or not the face in the face region is the face of the specific individual.
    The image processing apparatus according to claim 1.
  3.  前記特定個人判定部が、
     前記指標が所定の閾値より大きい場合、前記顔領域の顔が前記特定個人の顔であると判定し、
     前記指標が前記所定の閾値以下の場合、前記顔領域の顔が前記特定個人の顔ではないと判定する、
    請求項2記載の画像処理装置。
    The specific individual judgment unit
    When the index is larger than a predetermined threshold value, it is determined that the face in the face region is the face of the specific individual.
    When the index is equal to or less than the predetermined threshold value, it is determined that the face in the face region is not the face of the specific individual.
    The image processing apparatus according to claim 2.
  4.  前記特定個人判定部が、
     前記画像の1フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する、
    請求項1~3のいずれかの項に記載の画像処理装置。
    The specific individual judgment unit
    Based on the result of the determination for one frame of the image, it is determined whether or not the face in the face region is the face of the specific individual.
    The image processing apparatus according to any one of claims 1 to 3.
  5.  前記特定個人判定部が、
     前記画像の複数フレームに対する判定の結果に基づいて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する、
    請求項1~3のいずれかの項に記載の画像処理装置。
    The specific individual judgment unit
    Based on the result of determination for a plurality of frames of the image, it is determined whether or not the face in the face region is the face of the specific individual.
    The image processing apparatus according to any one of claims 1 to 3.
  6.  前記顔画像処理には、顔検出処理、顔向き推定処理、視線方向推定処理、及び目開閉検出処理のうちの少なくともいずれかが含まれている、
    請求項1~5のいずれかの項に記載の画像処理装置。
    The face image processing includes at least one of a face detection process, a face orientation estimation process, a line-of-sight direction estimation process, and an eye opening / closing detection process.
    The image processing apparatus according to any one of claims 1 to 5.
  7.  請求項1~6のいずれかの項に記載の画像処理装置と、
     該画像処理装置に入力する画像を撮像する撮像部と、
     前記画像処理装置による画像処理に基づく情報を出力する出力部と、
    を備えているモニタリング装置。
    The image processing apparatus according to any one of claims 1 to 6.
    An image pickup unit that captures an image to be input to the image processing device,
    An output unit that outputs information based on image processing by the image processing device, and
    A monitoring device equipped with.
  8.  請求項7記載のモニタリング装置と、
     該モニタリング装置と通信可能に接続され、該モニタリング装置から出力される前記情報に基づいて、所定の処理を実行する1以上の制御装置とを備えている、
    制御システム。
    The monitoring device according to claim 7 and
    It includes one or more control devices that are communicably connected to the monitoring device and execute a predetermined process based on the information output from the monitoring device.
    Control system.
  9.  前記モニタリング装置が、車両のドライバをモニタリングするための装置であり、
     前記制御装置が、前記車両に搭載される電子制御ユニットを含む、
    請求項8記載の制御システム。
    The monitoring device is a device for monitoring the driver of the vehicle.
    The control device includes an electronic control unit mounted on the vehicle.
    The control system according to claim 8.
  10.  撮像部から入力される画像を処理する画像処理方法であって、
     前記画像から顔の特徴量を抽出しながら顔領域を検出する顔検出ステップと、
     該顔検出ステップにより検出された前記顔領域の前記特徴量と、特定個人の顔を検出するための学習を行った学習済みの前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定ステップと、
     該特定個人判定ステップにより前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理ステップと、
     前記特定個人判定ステップにより前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理ステップとを含む、
    画像処理方法。
    It is an image processing method that processes an image input from an imaging unit.
    A face detection step of detecting a face region while extracting facial features from the image, and
    The face of the face region is used by using the feature amount of the face region detected by the face detection step and the learned face feature amount of the specific individual who has been trained to detect the face of the specific individual. A specific individual determination step for determining whether or not is the face of the specific individual, and
    When the face of the specific individual is determined by the specific individual determination step, the first face image processing step of performing the face image processing for the specific individual and the first face image processing step
    When it is determined by the specific individual determination step that it is not the face of the specific individual, it includes a second face image processing step of performing normal face image processing.
    Image processing method.
  11.  撮像部から入力される画像の処理を少なくとも1以上のコンピュータに実行させるためのプログラムであって、
     前記少なくとも1以上のコンピュータに、
     前記画像から顔の特徴量を抽出しながら顔領域を検出する顔検出ステップと、
     該顔検出ステップにより検出された前記顔領域の前記特徴量と、特定個人の顔を検出するための学習を行った学習済みの前記特定個人の顔特徴量とを用いて、前記顔領域の顔が前記特定個人の顔であるか否かを判定する特定個人判定ステップと、
     該特定個人判定ステップにより前記特定個人の顔であると判定された場合、特定個人用の顔画像処理を行う第1顔画像処理ステップと、
     前記特定個人判定ステップにより前記特定個人の顔ではないと判定された場合、通常の顔画像処理を行う第2顔画像処理ステップとを実行させるためのプログラム。
    A program for causing at least one or more computers to process an image input from an imaging unit.
    To at least one of the above computers
    A face detection step of detecting a face region while extracting facial features from the image, and
    The face of the face region is used by using the feature amount of the face region detected by the face detection step and the learned face feature amount of the specific individual who has been trained to detect the face of the specific individual. A specific individual determination step for determining whether or not is the face of the specific individual, and
    When the face of the specific individual is determined by the specific individual determination step, the first face image processing step of performing the face image processing for the specific individual and the first face image processing step
    A program for executing a second face image processing step of performing normal face image processing when it is determined by the specific individual determination step that the face is not the face of the specific individual.
PCT/JP2020/019848 2019-06-24 2020-05-20 Image processing device, monitoring device, control system, image processing method, and program WO2020261820A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2019-116544 2019-06-24
JP2019116544A JP7318352B2 (en) 2019-06-24 2019-06-24 Image processing device, monitoring device, control system, image processing method, and program

Publications (1)

Publication Number Publication Date
WO2020261820A1 true WO2020261820A1 (en) 2020-12-30

Family

ID=73995278

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/019848 WO2020261820A1 (en) 2019-06-24 2020-05-20 Image processing device, monitoring device, control system, image processing method, and program

Country Status (2)

Country Link
JP (1) JP7318352B2 (en)
WO (1) WO2020261820A1 (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010198313A (en) * 2009-02-25 2010-09-09 Denso Corp Device for specifying degree of eye opening

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010198313A (en) * 2009-02-25 2010-09-09 Denso Corp Device for specifying degree of eye opening

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
IWAI, YOSHIO. .: "4.Recognition Using Moving Images", A SURVEY ON FACE DETECTION AND FACE RECOGNITION, no. 38, 2005, pages 343 - 368 *

Also Published As

Publication number Publication date
JP7318352B2 (en) 2023-08-01
JP2021002281A (en) 2021-01-07

Similar Documents

Publication Publication Date Title
Bila et al. Vehicles of the future: A survey of research on safety issues
Alioua et al. Driver head pose estimation using efficient descriptor fusion
Trivedi et al. Looking-in and looking-out of a vehicle: Computer-vision-based enhanced vehicle safety
Rezaei et al. Simultaneous analysis of driver behaviour and road condition for driver distraction detection
US20220222466A1 (en) Material spectroscopy
Amanatiadis et al. ViPED: On-road vehicle passenger detection for autonomous vehicles
Sun et al. Driver fatigue detection system based on colored and infrared eye features fusion
Panicker et al. Open-eye detection using iris–sclera pattern analysis for driver drowsiness detection
CN114821696A (en) Material spectrometry
Rani et al. Development of an Automated Tool for Driver Drowsiness Detection
WO2021024905A1 (en) Image processing device, monitoring device, control system, image processing method, computer program, and recording medium
Shirpour et al. A probabilistic model for visual driver gaze approximation from head pose estimation
JP2004334786A (en) State detection device and state detection system
CN116012822B (en) Fatigue driving identification method and device and electronic equipment
WO2020261820A1 (en) Image processing device, monitoring device, control system, image processing method, and program
Llorca et al. Stereo-based pedestrian detection in crosswalks for pedestrian behavioural modelling assessment
WO2020261832A1 (en) Image processing device, monitoring device, control system, image processing method, and program
JP2021009503A (en) Personal data acquisition system, personal data acquisition method, face sensing parameter adjustment method for image processing device and computer program
CN114821697A (en) Material spectrum
CN114821694A (en) Material spectrometry
CN113128540B (en) Method and device for detecting vehicle theft behavior of non-motor vehicle and electronic equipment
Nowosielski Vision-based solutions for driver assistance
Kim et al. Driving environment assessment using fusion of in-and out-of-vehicle vision systems
CN112348718A (en) Intelligent auxiliary driving guidance method, device and computer storage medium
Öztürk et al. Drowsiness detection system based on machine learning using eye state

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20833155

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20833155

Country of ref document: EP

Kind code of ref document: A1