WO2018168038A1 - Driver seating determination device - Google Patents

Driver seating determination device Download PDF

Info

Publication number
WO2018168038A1
WO2018168038A1 PCT/JP2017/036276 JP2017036276W WO2018168038A1 WO 2018168038 A1 WO2018168038 A1 WO 2018168038A1 JP 2017036276 W JP2017036276 W JP 2017036276W WO 2018168038 A1 WO2018168038 A1 WO 2018168038A1
Authority
WO
WIPO (PCT)
Prior art keywords
driver
seat
seating determination
captured image
seated
Prior art date
Application number
PCT/JP2017/036276
Other languages
French (fr)
Japanese (ja)
Inventor
初美 青位
相澤 知禎
匡史 日向
Original Assignee
オムロン株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by オムロン株式会社 filed Critical オムロン株式会社
Publication of WO2018168038A1 publication Critical patent/WO2018168038A1/en

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R21/00Arrangements or fittings on vehicles for protecting or preventing injuries to occupants or pedestrians in case of accidents or other traffic risks
    • B60R21/01Electrical circuits for triggering passive safety arrangements, e.g. airbags, safety belt tighteners, in case of vehicle accidents or impending vehicle accidents
    • B60R21/015Electrical circuits for triggering passive safety arrangements, e.g. airbags, safety belt tighteners, in case of vehicle accidents or impending vehicle accidents including means for detecting the presence or position of passengers, passenger seats or child seats, and the related safety parameters therefor, e.g. speed or timing of airbag inflation in relation to occupant position or seat belt use

Definitions

  • the present invention relates to a driver seating determination device, a driver seating method, and a driver seating program.
  • a technique for determining whether a driver is seated in a driver's seat of a car has been proposed. Such a technique is used, for example, to issue a warning when a driver is seated in a driver's seat but does not wear a seat belt.
  • a pressure-sensitive sensor is provided in the driver's seat so that it can be determined whether or not the driver is seated in the driver's seat.
  • the present invention has been made to solve this problem, and it is possible to accurately determine whether or not a driver is seated in a driver's seat of an automobile.
  • An object is to provide a seating method and a driver's seating program.
  • a driver's seating determination apparatus is a driver's seating determination apparatus connected to at least one camera that captures a driver's seat of an automobile, and acquires an image acquired by the camera. And an analysis unit that determines whether or not a driver is seated in the driver's seat from the captured image.
  • a captured image obtained by capturing the driver's seat by the camera is acquired, and whether or not the driver is seated in the driver's seat is analyzed by analyzing whether or not the driver is included in the captured image. I am trying to judge. Therefore, it can be reliably determined that the driver is seated in the driver's seat.
  • the seating determination apparatus may further include an observation information acquisition unit that acquires observation information of the driver including face behavior information regarding the behavior of the driver's face, and the analysis unit includes the driver's observation information. Learning the seating information as to whether or not the driver is seated by inputting the captured image and the observation information into a learned learning device that has learned to determine seating in the driver's seat And a driver state estimation unit that is obtained from the vessel.
  • the observation information acquisition unit performs predetermined image analysis on the acquired captured image, thereby detecting whether or not the driver's face can be detected, the position of the face, the direction of the face, and the movement of the face.
  • Information on at least one of eye gaze direction, face organ position, and eye opening / closing can be acquired as the face behavior information.
  • the analysis unit may further include a resolution conversion unit that reduces the resolution of the acquired captured image, and the driver state estimation unit displays the captured image with the resolution reduced. It can be input to the learning device.
  • the analysis unit can determine the seating of the driver by various methods.For example, the analysis unit detects the driver's face from the captured image, and It can be determined that the driver is seated in the driver seat.
  • the driver's face can be detected by various methods.
  • the analysis unit detects the organ of the person's face from the image included in the captured image, thereby It can be determined that the driver is seated in the driver seat.
  • the analysis unit can determine the seating of the driver by various methods. For example, the analysis unit has learned learning that has been performed to detect a human face. It is possible to provide a learning device that takes a captured image including a driver's seat as an input and outputs whether or not a human face is included in the captured image.
  • Each of the seating determination devices may further include a warning unit that issues a warning when it is determined that the driver is not seated in the driver seat.
  • the above seating determination devices are particularly effective when the automobile has an automatic driving function.
  • the analysis unit can be configured to be able to determine the seating of the driver during the operation of the automatic driving function.
  • the method for determining whether a driver is seated according to the present invention includes a step of photographing a driver's seat of an automobile with at least one camera, and whether or not the driver is seated in the driver's seat from a photographed image photographed by the camera. And a step of judging.
  • the seating determination method it is possible to determine that the driver is seated in the driver seat by detecting the driver's face from the captured image.
  • the seating determination method it is possible to determine that the driver is seated in the driver seat by detecting a human facial organ from the image included in the captured image.
  • the driver seating determination program includes a step of photographing a driver's seat of a car with at least one camera on a car computer, and a driver seated in the driver's seat from a photographed image taken by the camera. And a step of determining whether or not.
  • the seating determination program it is possible to determine that the driver is seated in the driver seat by detecting the driver's face from the captured image.
  • the seating determination program it is possible to determine that the driver is seated in the driver seat by detecting a human facial organ from the image included in the photographed image.
  • FIG. 1 is a partial schematic configuration diagram of an automobile to which a seating determination device is attached
  • FIG. 2 is a diagram illustrating a schematic configuration of a seating determination system.
  • the driver's seat 900 is photographed by the camera 3 disposed in front of the driver's seat 900 to obtain a photographed image, and facial organs (eyes, nose, mouth) are obtained from the photographed image. ) Is detected, the driver 800 is determined to be seated in the driver's seat 900.
  • the seating determination system includes a seating determination device 1, a learning device 2, and a camera 3.
  • the seating determination device 1 can acquire a learned learning device created by the learning device 2 via the network 10, for example.
  • the type of the network 10 may be appropriately selected from, for example, the Internet, a wireless communication network, a mobile communication network, a telephone network, a dedicated network, and the like.
  • the learning device can be transmitted by directly connecting the seating determination device 1 and the learning device 2.
  • the learning device learned by the learning device 2 is stored in a storage medium such as a CD-ROM without connecting the seating determination device 1 and the learning device 2, and the learning device stored in the storage medium Can also be stored in the seating determination apparatus 1.
  • a storage medium such as a CD-ROM
  • a moving image the captured image is transmitted to the seating determination device 1 for each frame, and seating determination is performed.
  • FIG. 3 is a block diagram showing a seating determination apparatus according to the present embodiment.
  • the seating determination apparatus 1 according to the present embodiment is electrically connected to the control unit 11, the storage unit 12, the external interface 13, the input device 14, the output device 15, the communication interface 13, and the drive 17.
  • the communication interface and the external interface are described as “communication I / F” and “external I / F”, respectively.
  • the control unit 11 includes a CPU (Central Processing Unit), a RAM (Random Access Memory), a ROM (Read Only Memory), and the like, and controls each component according to information processing.
  • the storage unit 12 is, for example, an auxiliary storage device such as a hard disk drive or a solid state drive, and stores a seating determination program 121 executed by the control unit 11, learning result data 122 indicating information related to a learned learning device, and the like. .
  • the seating determination program 121 is a program for causing the seating determination apparatus 1 to execute a process as to whether or not a human face organ can be detected from a captured image.
  • the learning result data 122 is data for setting a learned learner. Details will be described later.
  • the communication interface 16 is, for example, a wired LAN (Local Area Network) module, a wireless LAN module, or the like, and is an interface for performing wired or wireless communication via a network.
  • the input device 14 is a device for performing input using, for example, a mouse or a keyboard.
  • the output device 15 is a device for outputting, for example, a display or a speaker.
  • the external interface 13 is a USB (Universal Serial Bus) port or the like, and is an interface for connecting to an external device such as the camera 3, a speaker in a vehicle, a display, or a device for controlling the speed.
  • various displays such as a display for car navigation provided on a dashboard can be used as the display in the vehicle.
  • the external device connected to the external interface 13 may not be limited to each of the above devices, and may be appropriately selected according to the embodiment. Therefore, the external interface 13 may be provided for each external device to be connected, and the number thereof can be selected as appropriate according to the embodiment.
  • the drive 17 is, for example, a CD (Compact Disk) drive, a DVD (Digital Versatile Disk) drive, or the like, and is a device for reading a program stored in the storage medium 91.
  • the type of the drive 17 may be appropriately selected according to the type of the storage medium 91.
  • the seating determination program 121 and / or the learning result data 122 may be stored in the storage medium 91.
  • the storage medium 91 stores information such as a program by an electrical, magnetic, optical, mechanical, or chemical action so that the information such as a program recorded by a computer or other device or machine can be read. It is a medium to do.
  • the seating determination apparatus 1 may acquire the seating determination program 121 and / or the learning result data 122 from the storage medium 91.
  • a disk-type storage medium such as a CD or a DVD is illustrated.
  • the type of the storage medium 91 is not limited to the disk type and may be other than the disk type.
  • Examples of the storage medium other than the disk type include a semiconductor memory such as a flash memory.
  • the control unit 11 may include a plurality of processors.
  • the seating determination device 1 may be composed of a plurality of information processing devices.
  • FIG. 4 is a block diagram illustrating the learning device according to the present embodiment.
  • the learning device 2 according to the present embodiment is for learning the learning device included in the second detection unit 102, and includes a control unit 21, a storage unit 22, a communication interface 23, and an input.
  • a computer in which the device 24, the output device 25, the external interface 26, and the drive 27 are electrically connected.
  • the communication interface and the external interface are described as “communication I / F” and “external I / F”, respectively.
  • the control unit 21 to the drive 27 and the storage medium 92 are the same as the control unit 11 to the drive 17 and the storage medium 91 of the seating determination device 1, respectively.
  • the storage unit 22 of the learning device 2 stores a learning program 221 executed by the control unit 21, learning data 222 used for learning of the learning device, learning result data 122 created by executing the learning program 221, and the like. . *
  • the learning program 221 is a program for causing the learning device 2 to execute a neural network learning process (FIG. 8) described later.
  • the learning data 222 is data for performing learning of a learning device in order to detect a human facial organ from a captured image. Details will be described later.
  • the learning program 221 and / or the learning data 222 may be stored in the storage medium 92 as in the seating determination apparatus 1.
  • the learning device 2 may acquire the learning program 221 and / or the learning data 222 to be used from the storage medium 92.
  • the learning device 2 may be a general-purpose server device, a desktop PC, or the like, in addition to an information processing device designed exclusively for the provided service.
  • FIG. 5 schematically illustrates an example of a functional configuration of the seating determination apparatus 1 according to the present embodiment.
  • the seating determination apparatus 1 functions as a computer including an image acquisition unit 111, an analysis unit 116, and a warning unit 117.
  • the image acquisition unit 111 acquires the captured image 123 generated by the camera 3. Further, the analysis unit 116 determines whether or not the driver is seated in the driver's seat from the captured image 123. When the analysis unit 116 determines that the driver is not seated in the driver's seat, the warning unit 117 is configured to issue a warning.
  • these functional configurations will be described in detail.
  • the analysis unit 116 uses the photographed image 123 as an input of a learning device learned to detect a facial organ. An output value is obtained from the learning device by the arithmetic processing of the learning device. Then, the analysis unit 116 determines whether or not a human face organ in the captured image 123 exists based on the output value obtained from the learning device.
  • the facial organ includes eyes, nose, mouth, and the like, and at least one of these feature points can be detected. However, depending on the type of camera, the eyes may not be detected when the driver is wearing sunglasses. For example, feature points of the nose and mouth can be detected. In addition, when the driver wears a mask, the nose and mouth cannot be detected, so that, for example, the eye feature point can be detected.
  • the seating determination apparatus 1 uses, as an example, a learning device that learns about the presence or absence of a facial organ in the captured image 123.
  • the learning device 7 is composed of a neural network. Specifically, it is a neural network having a multilayer structure used for so-called deep learning as shown in FIG. 4, and includes an input layer 71, an intermediate layer (hidden layer) 72, and an output layer 73 in order from the input. .
  • the neural network 7 includes one intermediate layer 72, the output of the input layer 71 is the input of the intermediate layer 72, and the output of the intermediate layer 72 is the input of the output layer 73.
  • the number of intermediate layers 72 is not limited to one, and the neural network 7 may include two or more intermediate layers 72.
  • Each layer 71 to 73 includes one or a plurality of neurons.
  • the number of neurons in the input layer 71 can be set according to the number of pixels in each captured image 123.
  • the number of neurons in the intermediate layer 72 can be set as appropriate according to the embodiment.
  • the output layer 73 can be set according to the determination of the presence or absence of a facial organ.
  • Adjacent layers of neurons are appropriately connected to each other, and a weight (connection load) is set for each connection.
  • each neuron is connected to all neurons in the adjacent layers, but the neuron connection is not limited to such an example, and is appropriately set according to the embodiment. It's okay.
  • a threshold is set for each neuron, and basically, the output of each neuron is determined by whether or not the sum of products of each input and each weight exceeds the threshold.
  • the seating determination apparatus 1 determines whether or not the driver is seated in the driver's seat based on the output value obtained from the output layer 73 by inputting the respective captured images to the input layer 71 of the neural network 7. Determine.
  • the configuration of the neural network 7 (for example, the number of layers of the neural network 7, the number of neurons in each layer, the connection relationship between neurons, the transfer function of each neuron), the weight of connection between each neuron, and each neuron
  • the information indicating the threshold value is included in the learning result data 122.
  • the seating determination device 1 refers to the learning result data 122 and sets the learned learning device 7 used for processing for determining whether or not the driver is seated in the driver's seat. This also applies to a second embodiment described later.
  • Warning section> When the analysis unit 116 determines that the driver is not seated in the driver's seat, the warning unit 117 drives a display, a speaker, and the like in the vehicle through the external interface 16 to give a warning. That is, it is displayed on the display that the driver is not seated, or the vehicle is informed through the speaker that the driver is not seated. In addition, a warning can also be given by reducing the speed of a running car, such as driving a brake, or stopping.
  • FIG. 6 schematically illustrates an example of a functional configuration of the learning device 2 according to the present embodiment.
  • the control unit 21 of the learning device 2 expands the learning program 221 stored in the storage unit 22 in the RAM. Then, the control unit 21 interprets and executes the learning program 221 expanded in the RAM, and controls each component. Accordingly, as illustrated in FIG. 6, the learning device 2 according to the present embodiment functions as a computer including the learning data acquisition unit 211 and the learning processing unit 212.
  • the learning data acquisition unit 211 includes, as learning data 222, a captured image 223 captured by the camera 3 and seating information 2241 indicating whether or not a facial organ is indicated in the captured image 223. Get a pair.
  • the captured image 223 and the seating information 2241 correspond to the teacher data of the neural network 8.
  • the learning processing unit 212 causes the neural network 8 to learn so as to output an output value corresponding to the seating information 2241.
  • a neural network 8 as an example of a learning device includes an input layer 81, an intermediate layer (hidden layer) 82, and an output layer 83, and is configured in the same manner as the neural network 7.
  • the layers 81 to 83 are the same as the layers 71 to 73 described above.
  • the learning processing unit 212 constructs the neural network 8 that outputs an output value corresponding to the seating information 2241 when the captured image 223 is input by the learning processing of the neural network.
  • the learning processing unit 212 stores information indicating the configuration of the constructed neural network 8, the weight of the connection between the neurons, and the threshold value of each neuron as the learning result data 122 in the storage unit 22.
  • the learning result data 122 is transmitted to the seating determination apparatus 1 by the various methods described above. Further, such learning result data 122 may be periodically updated.
  • the control part 21 may update the learning result data 122 which the seating determination apparatus 1 hold
  • each function of the seating determination device 1 and the learning device 2 will be described in detail in an operation example described later.
  • an example in which each function of the seating determination device 1 and the learning device 2 is realized by a general-purpose CPU is described.
  • part or all of the above functions may be realized by one or a plurality of dedicated processors.
  • functions may be omitted, replaced, and added as appropriate according to the embodiment.
  • FIG. 7 is a flowchart illustrating an example of a processing procedure of the seating determination apparatus 1. Note that the processing procedure described below is merely an example, and each processing may be changed as much as possible. Further, in the processing procedure described below, steps can be omitted, replaced, and added as appropriate according to the embodiment.
  • the user activates the seating determination apparatus 1 and causes the activated seating determination apparatus 1 to execute the seating determination program 121.
  • the control unit 11 of the seating determination apparatus 1 refers to the learning result data 122 and sets the structure of the neural network 7, the weight of connection between neurons, and the threshold value of each neuron. And the control part 11 determines whether the driver
  • the control unit 11 functions as the image acquisition unit 111 and photographs the driver's seat from the front from the camera 3 connected via the external interface 16.
  • the acquired captured image 123 is acquired (step S102).
  • the captured image 123 may be a still image, or in the case of a moving image, a captured image is acquired for each frame.
  • control unit 11 functions as the analysis unit 116 and determines whether or not a facial organ is included in each captured image 123 acquired in step S102 (step S103). If a facial organ is detected in the captured image 123, it is determined that the driver is seated in the driver's seat (YES in step S103). Thereafter, if driving is being performed (YES in step S101), the captured image 123 is continuously acquired (step S102), and the driver's seating is determined (step S103). On the other hand, if the facial organ cannot be detected in the captured image 123, it is determined that the driver is not seated in the driver's seat (NO in step S101), and a warning is transmitted (step S104).
  • control unit 11 functions as the warning unit 117 and notifies the vehicle that the driver is not seated in the driver's seat using the display or speaker in the vehicle.
  • the automobile can be decelerated or stopped. Thereafter, if driving is being performed (YES in step S101), the captured image 123 is continuously acquired (step S102), and the driver's seating is determined (step S103). On the other hand, when the operation is not performed, the process is stopped.
  • a warning may be issued immediately as described above. (Alternatively, a predetermined number of frames), a warning can be issued if a facial organ cannot be detected.
  • the above processing may be performed immediately after the ignition power of the automobile is turned on. For example, when the automobile can be switched between the manual operation mode and the automatic operation mode, the automobile has shifted to the automatic operation mode. You may only do that.
  • the captured image 123 obtained by capturing the driver's seat with the camera 3 is acquired, and by analyzing whether or not the captured image 123 includes a human facial organ, It is determined whether or not the driver is seated in the driver's seat. Therefore, it can be reliably determined that the driver is seated in the driver's seat.
  • the determination is performed by the learning device 7 configured by a neural network. That is, since the learning device 7 has learned to detect a facial organ from many captured images 123, it can make a highly accurate determination.
  • the seating determination system according to the present embodiment includes the seating determination device 1 and the learning device 2 as in the first embodiment.
  • the seating determination device 1 captures a captured image from a camera 3 that is arranged to capture a driver 800 that has arrived at the driver's seat of the vehicle. get.
  • the seating determination apparatus 1 acquires driver observation information including face behavior information related to the behavior of the driver 800 face.
  • the seating determination device 1 uses the acquired captured image and observation to a learned learning device (a neural network described later) that has performed learning to determine whether or not the driver 800 is seated in the driver's seat 900. By inputting information, it is determined whether or not the driver 800 is seated in the driver's seat 900.
  • the learning device 2 constructs a learning device to be used in the seating determination device 1, that is, whether or not the driver 800 is seated in the driver's seat according to the input of the captured image and the observation information. It is a computer that performs machine learning of a learning device so as to output seating information indicating the above. Specifically, the learning device 2 acquires a set of the above-described captured image, observation information, and seating information as learning data. Then, the learning device 2 causes the learning device (a neural network 6 described later) to learn so as to output an output value corresponding to the seating information when the captured image and the observation information are input. As a result, a learned learning device used in the seating determination apparatus 1 is created. The connection between the seating determination device 1 and the learning device 2 is the same as that in the first embodiment.
  • a learned learning device that has performed learning for estimating the seating of the driver is used.
  • the driver who took the driver's seat is photographed.
  • a photographed image obtained from the camera 3 arranged in the above is used. Therefore, not only the behavior of the driver 800's face but also the state of the driver's 800 body (for example, body orientation, posture, etc.) can be analyzed from the captured image. Therefore, according to the present embodiment, it is possible to determine whether or not the driver 800 is seated in the driver's seat 900, reflecting various states that the driver 800 can take. Details will be described below.
  • FIG. 8 is a block diagram of the seating determination apparatus according to the present embodiment.
  • the hardware configuration of the seating determination device according to the present embodiment is substantially the same as that of the first embodiment, and the devices connected to the external I / F are different. Therefore, hereinafter, only differences from the first embodiment will be described, and the same components will be denoted by the same reference numerals and description thereof will be omitted.
  • the external interface 13 is connected to the navigation device 30, the biosensor 32, and the speaker 33 in addition to the above-described camera 3 via, for example, CAN (Controller (Area Network).
  • the navigation device 30 is a computer that provides route guidance when the vehicle is traveling.
  • a known car navigation device may be used as the navigation device 30.
  • the navigation device 30 is configured to measure the position of the vehicle based on a GPS (Global Positioning System) signal, and to perform route guidance using map information and surrounding information on surrounding buildings and the like.
  • GPS information information indicating the vehicle position measured based on the GPS signal.
  • the biological sensor 32 is configured to measure the biological information of the driver 800.
  • the biological information to be measured is not particularly limited, and may be, for example, an electroencephalogram, a heart rate, or the like.
  • the biological sensor 32 is not particularly limited as long as biological information to be measured can be measured.
  • a known brain wave sensor, pulse sensor, or the like may be used.
  • the biosensor 32 is attached to the body part of the driver 800 corresponding to the biometric information to be measured.
  • the speaker 33 is configured to output sound.
  • the speaker 33 is used to warn the driver 800 to take a state suitable for driving the vehicle when the driver 800 is not in a state suitable for driving the vehicle while the vehicle is running. Is done. Details will be described later.
  • FIG. 9 schematically illustrates an example of a functional configuration of the seating determination apparatus 1 according to the present embodiment.
  • the control unit 11 of the seating determination apparatus 1 expands the program 121 stored in the storage unit 12 in the RAM.
  • the control unit 11 interprets and executes the program 121 expanded in the RAM by the CPU and controls each component.
  • the seating determination apparatus 1 includes a computer including an image acquisition unit 111, an observation information acquisition unit 112, a resolution conversion unit 113, a driving state estimation unit 114, and a warning unit 115. Function as. Among these, the resolution conversion unit 113 and the operation state estimation unit 114 correspond to the analysis unit of the present invention.
  • the image acquisition unit 111 acquires the captured image 123 from the camera 31 that is arranged so as to capture the driver 800 seated in the driver's seat of the vehicle.
  • the observation information acquisition unit 112 acquires observation information 124 including face behavior information 1241 related to the behavior of the face of the driver 800 and biological information 1242 measured by the biological sensor 32.
  • the face behavior information 1241 is obtained by image analysis of the captured image 123.
  • the observation information 124 may not be limited to such an example, and the biological information 1242 may be omitted. In this case, the biosensor 32 may be omitted. That is, only the face behavior information 1241 acquired from the captured image 123 can be used as the observation information 124.
  • the resolution conversion unit 113 reduces the resolution of the captured image 123 acquired by the image acquisition unit 111. Thereby, the resolution conversion unit 113 forms a low-resolution captured image 1231.
  • the driving state estimation unit 114 reduces the resolution of the captured image 123 to a learned learning device (neural network 5) that has performed learning for reversing whether or not the driver is seated in the driver's seat.
  • the low-resolution captured image 1231 and the observation information 124 obtained in the above are input. Accordingly, the driving state estimation unit 114 acquires the sitting information 125 related to the sitting of the driver 800 from the learning device. Note that the resolution reduction process may be omitted. In this case, the driving state estimation unit 114 may input the captured image 123 to the learning device.
  • the driver 800 in determining whether or not the driver 800 is seated in the driver's seat 900, there is a possibility that an erroneous determination is made. For example, when the passenger in the front passenger seat has a body with a face on the driver's seat 900, or when the passenger in the rear seat has a body with a face on the driver's seat 900, the driver can enter the driver's seat 900. There is a risk that 800 is seated. Further, there is a possibility that it is determined that the driver is not seated even though the driver seat 900 is seated. For example, when the driver 800 is depressed or facing backward, the facial organ cannot be detected as will be described later, and it may be determined that the driver 800 is not seated.
  • the seating determination device 1 in addition to using information on facial organs included in the observation information 124 as a material for determination, the state of the body of the person shown in the driver's seat from the low-resolution captured image 1231 (For example, body orientation, posture, etc.) is used as a material for determination. From such a physical state, it can be determined whether the person in the driver's seat is a person sitting in the driver's seat, a passenger seat or a rear seat, or a child.
  • the state of the body of the person shown in the driver's seat from the low-resolution captured image 1231 (For example, body orientation, posture, etc.) is used as a material for determination. From such a physical state, it can be determined whether the person in the driver's seat is a person sitting in the driver's seat, a passenger seat or a rear seat, or a child.
  • the warning unit 115 is the same as that of the first embodiment, and when it is determined that the driver 800 is not seated in the driver's seat 900, the display unit and the speaker in the vehicle are driven through the external interface 16 to give a warning.
  • the seating determination device 1 is a neural network as a learned learner that has performed learning for determining whether or not the driver 800 is seated in the driver seat 900. 5 is used.
  • the neural network 5 according to the present embodiment is configured by combining a plurality of types of neural networks.
  • the neural network 5 is divided into four parts: a fully connected neural network 51, a convolutional neural network 52, a connected layer 53, and an LSTM network 54.
  • the fully connected neural network 51 and the convolutional neural network 52 are arranged in parallel on the input side.
  • Observation information 124 is input to the fully connected neural network 51, and a low-resolution captured image 1231 is input to the convolutional neural network 52.
  • the connection layer 53 combines the outputs of the fully connected neural network 51 and the convolutional neural network 52.
  • the LSTM network 54 receives the output from the coupling layer 53 and outputs the seating information 125.
  • the fully connected neural network 51 is a so-called multilayered neural network, and includes an input layer 511, an intermediate layer (hidden layer) 512, and an output layer 513 in order from the input side.
  • the number of layers of the fully connected neural network 51 may not be limited to such an example, and may be appropriately selected according to the embodiment.
  • Each layer 511 to 513 includes one or a plurality of neurons (nodes).
  • the number of neurons included in each of the layers 511 to 513 may be set as appropriate according to the embodiment.
  • the all-connected neural network 51 is configured by connecting each neuron included in each layer 511 to 513 to all the neurons included in the adjacent layers.
  • a weight (coupling load) is appropriately set for each coupling.
  • the convolutional neural network 52 is a forward propagation neural network having a structure in which convolutional layers 521 and pooling layers 522 are alternately connected.
  • a plurality of convolutional layers 521 and pooling layers 522 are alternately arranged on the input side. Then, the output of the pooling layer 522 arranged on the most output side is input to the total coupling layer 523, and the output of the total coupling layer 523 is input to the output layer 524.
  • the convolution layer 521 is a layer that performs an image convolution operation.
  • Image convolution corresponds to processing for calculating the correlation between an image and a predetermined filter. Therefore, by performing image convolution, for example, a shading pattern similar to the shading pattern of the filter can be detected from the input image.
  • the pooling layer 522 is a layer that performs a pooling process.
  • the pooling process discards a part of the information of the position where the response to the image filter is strong, and realizes the invariance of the response to the minute position change of the feature appearing in the image.
  • the total connection layer 523 is a layer in which all neurons between adjacent layers are connected. That is, each neuron included in all connection layers 523 is connected to all neurons included in adjacent layers.
  • the convolutional neural network 52 may include two or more fully connected layers 523. Further, the number of neurons included in all connection layers 423 may be set as appropriate according to the embodiment.
  • the output layer 524 is a layer arranged on the most output side of the convolutional neural network 52.
  • the number of neurons included in the output layer 524 may be appropriately set according to the embodiment.
  • the configuration of the convolutional neural network 52 is not limited to such an example, and may be appropriately set according to the embodiment.
  • connection layer 53 is disposed between the fully connected neural network 51 and the convolutional neural network 52 and the LSTM network 54.
  • the connection layer 53 combines the output from the output layer 513 of the fully connected neural network 51 and the output from the output layer 524 of the convolutional neural network 52.
  • the number of neurons included in the connection layer 53 may be appropriately set according to the number of outputs of the fully connected neural network 51 and the convolutional neural network 52.
  • the LSTM network 54 is a recurrent neural network that includes an LSTM block 542.
  • a recursive neural network is a neural network having a loop inside, such as a path from an intermediate layer to an input layer.
  • the LSTM network 54 has a structure in which an intermediate layer of a general recurrent neural network is replaced with an LSTM block 542.
  • the LSTM network 54 includes an input layer 541, an LSTM block 542, and an output layer 543 in order from the input side.
  • a path returning from the LSTM block 542 to the input layer 541 is provided. Have.
  • the number of neurons included in the input layer 541 and the output layer 543 may be set as appropriate according to the embodiment.
  • the LSTM block 542 includes an input gate and an output gate, and is configured to be able to learn information storage and output timing (S. Hochreiter and J.Schmidhuber, "Long short-term memory” Neural Computation, 9). (8): 1735-1780, November 15, 1997).
  • the LSTM block 542 may also include a forgetting gate that adjusts the timing of forgetting information (FelixFA. Gers, Jurgen Schmidhuber and Fred Cummins, "Learning to Forget: Continual Prediction with LSTM” Neural Computation, pages 2451- 2471, “October” 2000).
  • the configuration of the LSTM network 54 can be set as appropriate according to the embodiment.
  • (E) Summary A threshold is set for each neuron, and basically, the output of each neuron is determined by whether or not the sum of products of each input and each weight exceeds the threshold.
  • the seating determination apparatus 1 inputs observation information 124 to the fully connected neural network 51 and inputs a low-resolution captured image 1231 to the convolutional neural network 52. And the seating determination apparatus 1 performs the firing determination of each neuron included in each layer in order from the input side. Thereby, the seating determination apparatus 1 acquires an output value corresponding to the seating information 125 from the output layer 543 of the neural network 5.
  • FIG. 10 schematically illustrates an example of a functional configuration of the learning device 2 according to the present embodiment.
  • the control unit 21 of the learning device 2 develops the learning program 221 stored in the storage unit 22 in the RAM. Then, the control unit 21 interprets and executes the learning program 221 expanded in the RAM, and controls each component. Accordingly, as illustrated in FIG. 10, the learning device 2 according to the present embodiment functions as a computer including the learning data acquisition unit 211 and the learning processing unit 212.
  • the learning data acquisition unit 211 includes a captured image acquired from an imaging device arranged to capture a driver who has arrived at the driver's seat of the vehicle, and facial behavior information regarding the behavior of the driver's face.
  • a set of observation information and seating information related to the driver's seating is acquired as learning data.
  • the learning data acquisition unit 211 acquires a set of the low-resolution captured image 223, the observation information 224, and the seating information 225 as learning data 222.
  • the low-resolution captured image 223 and the observation information 224 correspond to the low-resolution captured image 1231 and the observation information 124, respectively.
  • the seating information 225 corresponds to the seating information 125.
  • the learning processing unit 212 When the learning processing unit 212 inputs the low-resolution captured image 223 and the observation information 224, the learning processing unit 212 learns the learning device so as to output an output value corresponding to the seating information 225. Thereby, in this learning apparatus 2, learning for avoiding the erroneous determination described above is performed.
  • the learning device to be learned is a neural network 6.
  • the neural network 6 includes a fully connected neural network 61, a convolutional neural network 62, a connected layer 63, and an LSTM network 64, and is configured in the same manner as the neural network 5.
  • the fully connected neural network 61, the convolutional neural network 62, the connection layer 63, and the LSTM network 64 are the same as the above-described all connection neural network 51, the convolutional neural network 52, the connection layer 53, and the LSTM network 54, respectively.
  • the learning processing unit 212 inputs the observation information 224 to the fully connected neural network 61 and inputs the low-resolution captured image 223 to the convolutional neural network 62 by the learning processing of the neural network, and outputs an output value corresponding to the seating information 225 to the LSTM.
  • the neural network 6 that outputs from the network 64 is constructed.
  • the learning processing unit 212 stores information indicating the configuration of the constructed neural network 6, the weight of the connection between the neurons, and the threshold value of each neuron as the learning result data 122 in the storage unit 22.
  • FIG. 11 is a flowchart illustrating an example of a processing procedure of the seating determination apparatus 1.
  • the processing procedure described below is merely an example, and each processing may be changed as much as possible. Further, in the processing procedure described below, steps can be omitted, replaced, and added as appropriate according to the embodiment.
  • the driver 800 activates the seating determination device 1 by turning on the ignition power of the vehicle, and causes the activated seating determination device 1 to execute the program 121.
  • the timing which the seating determination apparatus 1 starts is not restricted to this.
  • the timing at which the seating determination device 1 is activated may be the timing at which the automatic operation mode is activated.
  • the control unit 11 of the seating determination device 1 acquires map information, peripheral information, and GPS information from the navigation device 30 and starts automatic driving of the vehicle based on the acquired map information, peripheral information, and GPS information.
  • a control method for automatic operation a known control method can be used.
  • the control unit 11 monitors the state of the driver 800 according to the following processing procedure.
  • Step S202 In step S ⁇ b> 101, the control unit 11 functions as the image acquisition unit 111, and acquires the captured image 123 from the camera 31 arranged to capture the driver 800 attached to the driver's seat of the vehicle.
  • the captured image 123 to be acquired may be a moving image or a still image.
  • the control unit 11 advances the processing to the next step S203.
  • step S203 In step S ⁇ b> 203, the control unit 11 functions as the observation information acquisition unit 112, and acquires observation information 124 including face behavior information 1241 and biological information 1242 that behave on the face of the driver 800.
  • the control unit 11 advances the processing to the next step S204.
  • the face behavior information 1241 may be acquired as appropriate.
  • the control unit 11 performs predetermined image analysis on the captured image 123 acquired in step S202, thereby determining whether the driver 800 can detect the face, the position of the face, the direction of the face, the movement of the face, and the line of sight.
  • Information regarding at least one of the direction, the position of the facial organ, and the opening and closing of the eyes can be acquired as the face behavior information 1241.
  • the control unit 11 detects the face of the driver 800 from the photographed image 123 and specifies the position of the detected face. Thereby, the control part 11 can acquire the information regarding the detectability and position of a face. Moreover, the control part 11 can acquire the information regarding a motion of a face by detecting a face continuously. Next, the control unit 11 detects each organ (eye, mouth, nose, ear, etc.) included in the face of the driver 800 in the detected face image. Thereby, the control part 11 can acquire the information regarding the position of the facial organ.
  • control part 11 can acquire the information regarding the direction of a face, the direction of eyes
  • a known image analysis method may be used for face detection, organ detection, and organ state analysis.
  • the control unit 11 When the captured image 123 to be acquired is a moving image or a plurality of still images arranged in time series, the control unit 11 performs these image analyzes on each frame of the captured image 123 so that the acquired images are arranged in time series. Various information can be acquired. Thereby, the control part 11 can acquire the various information represented by the histogram or the statistic (an average value, a variance value, etc.) with time series data.
  • control unit 11 acquires biological information (for example, brain waves, heart rate, etc.) 1242 from the biological sensor 32.
  • biological information 1242 may be represented by a histogram or a statistic (average value, variance value, etc.). Similar to the face behavior information 1241, the control unit 11 can obtain the biological information 1242 as time-series data by continuously accessing the biological sensor 32. As described above, the biological information 1242 is not necessarily required, and the control unit 11 can generate the observation information 124 using only the face behavior information 1241.
  • Step S204 the control unit 11 functions as the resolution conversion unit 113, and reduces the resolution of the captured image 123 acquired in step S202. Thereby, the control unit 11 forms a low-resolution captured image 1231.
  • the processing method for reducing the resolution is not particularly limited, and may be appropriately selected according to the embodiment.
  • the control unit 11 can form the low-resolution captured image 1231 by the nearest neighbor method, the bilinear interpolation method, the bicubic method, or the like.
  • the control unit 11 advances the processing to the next step S104. Note that this step S103 may be omitted. That is, the control unit 11 can input the captured image 123 to the learning device 5 without reducing the resolution of the captured image 123.
  • step S ⁇ b> 205 the control unit 11 functions as the driving state estimation unit 114, and executes arithmetic processing of the neural network 5 using the acquired observation information 124 and the low-resolution captured image 1231 as inputs of the neural network 5. Thereby, in step S ⁇ b> 105, the control unit 11 obtains an output value corresponding to each of the seating information 125 from the neural network 5.
  • control unit 11 inputs the observation information 124 acquired in step S203 to the input layer 511 of the fully connected neural network 51, and the low-resolution captured image 1231 acquired in step S204 is the most in the convolutional neural network 52. It inputs into the convolution layer 521 arrange
  • step S207 the control unit 11 functions as the warning unit 115, and determines whether or not the driver 800 is seated in the driver's seat based on the seating information 125 acquired in step S206. And when it determines with the driver
  • the seating determination apparatus 1 performs the processing from step S202 to step S204, the observation information 124 including the face behavior information 1241 of the driver 800 and the driver who has arrived at the driver's seat of the vehicle. And a captured image (low-resolution captured image 1231) obtained from the camera 3 arranged so as to capture the image.
  • the seating determination device 1 uses the acquired observation information 124 and the low-resolution captured image 1231 as inputs of the learned neural network (the neural network 5) in steps S205 and S206, so that the driver 800 It is determined whether the user is seated at 900.
  • the learned neural network is created by the learning device 2 using learning data including the low-resolution captured image 223, the observation information 224, and the seating information 225. Therefore, in the present embodiment, in the process of determining the driver's seating, not only the behavior of the driver 800's face but also the state of the driver's 800 body (for example, the body orientation, Attitude).
  • the captured image 123 it is determined from the state of the body of the person shown in the driver's seat 900 whether this is the driver 800 or whether the person in the front passenger seat or the rear seat is riding on the driver's seat 900. Can be determined. Even when the driver 800 is seated, the facial organ cannot be accurately detected when the driver 800 is depressed or turned backward. Even in such a case, the body of the driver 800 is not detected. By detecting the state, it can be determined that the driver 800 is seated in the driver's seat 900. Therefore, it can be accurately determined whether or not the driver 800 is seated on the driver's seat 900.
  • observation information (124, 224) including driver's face behavior information is used as an input to the neural network (5, 6). Therefore, the captured image to be input to the neural network (5, 6) does not have to have a high resolution so that the behavior of the driver's face can be determined. Therefore, in this embodiment, low-resolution captured images (1231, 223) obtained by reducing the resolution of the captured image obtained by the camera 31 may be used as the input of the neural network (5, 6). Thereby, the calculation amount of the arithmetic processing of the neural network (5, 6) can be reduced, and the load on the processor can be reduced.
  • the resolution of the low-resolution captured image (1231, 223) is preferably such that the behavior of the driver's face cannot be discriminated, but the feature relating to the driver's posture can be extracted. Note that, as described above, it is not always necessary to reduce the resolution of the captured image 123.
  • the captured image 123 can be used as the input of the learning device 5 without considering the processing load.
  • the neural network 5 includes a fully connected neural network 51 and a convolutional neural network 52 on the input side.
  • the observation information 124 is input to the fully connected neural network 51
  • the low-resolution captured image 1231 is input to the convolutional neural network 52.
  • analysis suitable for each input can be performed.
  • the neural network 5 according to this embodiment includes an LSTM network 54.
  • the time series data is used for the observation information 124 and the low-resolution captured image 1231, and the seating of the driver 800 can be determined in consideration of not only short-term dependency but also long-term dependency. . Therefore, according to the present embodiment, the seating determination accuracy of the driver 800 can be increased.
  • each neural network (7, 8) a general forward propagation type neural network having a multilayer structure is used as each neural network (7, 8).
  • the type of each neural network (7, 8) may not be limited to such an example, and may be appropriately selected according to the embodiment.
  • each neural network (7, 8) may be a convolutional neural network that uses the input layer 71 and the intermediate layer 72 as a convolution layer and a pooling layer.
  • each neural network (7, 8) may be a recursive neural network having a connection that recurs from the output side to the input side, such as the intermediate layer 72 to the input layer 71.
  • the number of layers in each neural network (7, 8), the number of neurons in each layer, the connection relationship between neurons, and the transfer function of each neuron may be determined as appropriate according to the embodiment.
  • the seating determination device 1 and the learning device 2 that learns the learning device (neural network) 7 are configured by separate computers.
  • the configuration of the seating determination device 1 and the learning device 2 may not be limited to such an example, and a system having both functions of the seating determination device 1 and the learning device 2 is realized by one or a plurality of computers. May be.
  • the learning device 2 can also be used by being incorporated in the seating determination device 1.
  • the learning device is configured by a neural network.
  • the type of learning device is not limited to the neural network as long as the captured image 123 captured by the camera 3 can be used as an input, and may be appropriately selected according to the embodiment.
  • a learning device capable of inputting a plurality of captured images 123 for example, a learning device configured by a learning device that learns by support vector machine, self-organizing map, or reinforcement learning in addition to the neural network can be cited. .
  • the seating determination device 1 is mounted on a vehicle as a single device.
  • the seating determination program can be installed in a computer of the vehicle to perform seating determination.
  • the method for detecting the facial organ can be used other than the learning device as described above.
  • a method there are various methods, for example, known pattern matching can be used.
  • there is a method of extracting feature points using a three-dimensional model Specifically, for example, a method described in International Publication No. 2006/051607, Japanese Patent Application Laid-Open No. 2007-249280, or the like is adopted. Can do.
  • seating can be determined by acquiring a three-dimensional shape of an object located in the driver's seat by the visual volume intersection method and determining whether the three-dimensional shape is a person.
  • a plurality of cameras are provided in the vehicle, and the driver's seat is photographed from a plurality of angles with the plurality of cameras, and a plurality of photographed images are acquired.
  • the three-dimensional shape of the object seated in a driver's seat is acquired from a some picked-up image by a visual volume intersection method. Then, it is determined whether or not the three-dimensional shape is a person. If the person is a person, it can be determined that the driver is seated in the driver's seat. On the other hand, when the three-dimensional shape cannot be obtained, or when it is determined that the three-dimensional shape is not a person, a warning can be given as described above.
  • the observation information 124 includes biological information 1242 in addition to the face behavior information 1241.
  • the configuration of the observation information 124 is not limited to such an example, and may be appropriately selected according to the embodiment.
  • the biological information 1242 may be omitted.
  • the observation information 124 may include information other than the biological information 1242.
  • a driver's seating determination device connected to at least one camera for photographing a driver's seat of a car, Comprising at least one hardware processor;
  • the hardware processor is Obtaining a captured image captured by the camera,
  • a driver seating determination device that determines whether a driver is seated in the driver seat from the captured image.

Landscapes

  • Engineering & Computer Science (AREA)
  • Mechanical Engineering (AREA)
  • Traffic Control Systems (AREA)
  • Image Analysis (AREA)

Abstract

The driver seating determination device according to the present invention is connected to at least one camera for photographing a driver's seat of an automobile, wherein the driver seating determination device comprises an acquisition unit that acquires a photograph image photographed by the camera, and an analyzing unit that from the photograph image assesses whether or not a driver is sitting in the driver's seat.

Description

運転者の着座判定装置Driver seating determination device
 本発明は、運転者の着座判定装置、運転者の着座方法、及び運転者の着座プログラムに関する。 The present invention relates to a driver seating determination device, a driver seating method, and a driver seating program.
 自動車の運転席に運転者が着座しているか否かを判断するための種々の技術が提案されている。このような技術は、例えば、運転者が運転席に着座しているにもかかわらず、シートベルトを着用していない場合に、警告を発することに利用される。例えば、特許文献1には、運転席に感圧センサを設け、これによって運転席に運転者が着座したか否かを判断するようにしている。 Various techniques for determining whether a driver is seated in a driver's seat of a car have been proposed. Such a technique is used, for example, to issue a warning when a driver is seated in a driver's seat but does not wear a seat belt. For example, in Patent Document 1, a pressure-sensitive sensor is provided in the driver's seat so that it can be determined whether or not the driver is seated in the driver's seat.
特開2011-213342号公報JP 2011-213342 A
 しかしながら、上記技術では、例えば、運転席に重量の大きい荷物が置かれた場合には、これを運転者と認識して誤作動を引き起こす可能性がある。また、近年の自動運転技術では、自動運転中であっても、運転者が運転席に着座していることが求められることがあるため、このような場合に、運転者の着座を正確に判断できないと、運転に支障を来すおそれがある。本発明は、この問題を解決するためになされたものであり、自動車の運転席に運転者が着座しているか否かを正確に判断することができる、運転者の着座判定装置、運転者の着座方法、及び運転者の着座プログラムを提供することを目的とする。 However, with the above technology, for example, when a heavy load is placed in the driver's seat, it may be recognized as a driver and cause a malfunction. Also, in recent automatic driving technology, it is sometimes required that the driver is seated in the driver's seat even during automatic driving. In such a case, the driver's seating is accurately determined. Failure to do so may interfere with driving. The present invention has been made to solve this problem, and it is possible to accurately determine whether or not a driver is seated in a driver's seat of an automobile. An object is to provide a seating method and a driver's seating program.
 本発明に係る運転者の着座判定装置は、自動車の運転席を撮影する少なくとも1つのカメラと接続される運転者の着座判定装置であって、前記カメラによって撮影された撮影画像を取得する画像取得部と、前記撮影画像から前記運転席に運転者が着座しているか否かを判断する解析部と、を備えている。 A driver's seating determination apparatus according to the present invention is a driver's seating determination apparatus connected to at least one camera that captures a driver's seat of an automobile, and acquires an image acquired by the camera. And an analysis unit that determines whether or not a driver is seated in the driver's seat from the captured image.
 この構成によれば、カメラによって運転席を撮影した撮影画像を取得し、この撮影画像に運転者が含まれているか否かを解析することで、運転者が運転席に着座しているか否かを判断するようにしている。したがって、運転者が運転席に着座していることを確実に判断することができる。 According to this configuration, a captured image obtained by capturing the driver's seat by the camera is acquired, and whether or not the driver is seated in the driver's seat is analyzed by analyzing whether or not the driver is included in the captured image. I am trying to judge. Therefore, it can be reliably determined that the driver is seated in the driver's seat.
 上記着座判定装置においては、前記運転者の顔の挙動に関する顔挙動情報を含む当該運転者の観測情報を取得する観測情報取得部をさらに備えることができ、前記解析部は、前記運転者の前記運転席への着座を判定するための学習を行った学習済みの学習器に、前記撮影画像及び前記観測情報を入力することで、前記運転者が着座しているか否かの着座情報を当該学習器から取得する運転者状態推定部と、を備えることができる。 The seating determination apparatus may further include an observation information acquisition unit that acquires observation information of the driver including face behavior information regarding the behavior of the driver's face, and the analysis unit includes the driver's observation information. Learning the seating information as to whether or not the driver is seated by inputting the captured image and the observation information into a learned learning device that has learned to determine seating in the driver's seat And a driver state estimation unit that is obtained from the vessel.
 上記着座判定装置において、前記観測情報取得部は、取得した前記撮影画像に対して所定の画像解析を行うことで、前記運転者の顔の検出可否、顔の位置、顔の向き、顔の動き、視線の方向、顔の器官の位置、及び目の開閉の少なくともいずれか1つに関する情報を前記顔挙動情報として取得することができる。 In the seating determination apparatus, the observation information acquisition unit performs predetermined image analysis on the acquired captured image, thereby detecting whether or not the driver's face can be detected, the position of the face, the direction of the face, and the movement of the face. Information on at least one of eye gaze direction, face organ position, and eye opening / closing can be acquired as the face behavior information.
 上記各着座判定装置において、前記解析部は、取得した前記撮影画像の解像度を低下させる解像度変換部を更に備えることができ、前記運転者状態推定部は、解像度を低下させた前記撮影画像を前記学習器に入力することができる。 In each of the seating determination devices, the analysis unit may further include a resolution conversion unit that reduces the resolution of the acquired captured image, and the driver state estimation unit displays the captured image with the resolution reduced. It can be input to the learning device.
 上記着座判定装置において、解析部は、種々の方法で、運転者の着座を判断することができるが、例えば、前記解析部は、前記撮影画像から前記運転者の顔を検出することで、前記運転席に運転者が着座していると判断することができる。 In the seating determination apparatus, the analysis unit can determine the seating of the driver by various methods.For example, the analysis unit detects the driver's face from the captured image, and It can be determined that the driver is seated in the driver seat.
 上記着座判定装置においては、種々の方法で運転者の顔を検出することができるが、例えば、前記解析部は、前記撮影画像に含まれる像から人の顔の器官を検出することで、前記運転席に運転者が着座していると判断することができる。 In the seating determination apparatus, the driver's face can be detected by various methods.For example, the analysis unit detects the organ of the person's face from the image included in the captured image, thereby It can be determined that the driver is seated in the driver seat.
 上記着座判定装置において、解析部は、種々の方法で、運転者の着座を判断することができるが、例えば、前記解析部は、人の顔を検出するための学習を行った学習済みの学習器であって、運転席を含む撮影画像を入力とし、当該撮影画像に人の顔が含まれるか否かを出力とする学習器を備えることができる。 In the seating determination apparatus, the analysis unit can determine the seating of the driver by various methods. For example, the analysis unit has learned learning that has been performed to detect a human face. It is possible to provide a learning device that takes a captured image including a driver's seat as an input and outputs whether or not a human face is included in the captured image.
 上記各着座判定装置においては、前記運転席に運転者が着座していないと判断した場合に、警告を発する警告部をさらに備えることができる。 Each of the seating determination devices may further include a warning unit that issues a warning when it is determined that the driver is not seated in the driver seat.
 上記各着座判定装置においては、特に、自動車が自動運転機能を有するものであるときに有効である。この場合、前記解析部は、前記自動運転機能の作動中に、運転者の着座を判断することができるように構成できる。 The above seating determination devices are particularly effective when the automobile has an automatic driving function. In this case, the analysis unit can be configured to be able to determine the seating of the driver during the operation of the automatic driving function.
 本発明に係る運転者の着座判定方法は、自動車の運転席を、少なくとも1つのカメラで撮影するステップと、前記カメラによって撮影された撮影画像から前記運転席に運転者が着座しているか否かを判断するステップと、を備えている。 The method for determining whether a driver is seated according to the present invention includes a step of photographing a driver's seat of an automobile with at least one camera, and whether or not the driver is seated in the driver's seat from a photographed image photographed by the camera. And a step of judging.
 上記着座判定方法においては、前記撮影画像から前記運転者の顔を検出することで、前記運転席に運転者が着座していると判断することができる。 In the seating determination method, it is possible to determine that the driver is seated in the driver seat by detecting the driver's face from the captured image.
 上記着座判定方法においては、前記撮影画像に含まれる像から人の顔の器官を検出することで、前記運転席に運転者が着座していると判断することができる。 In the seating determination method, it is possible to determine that the driver is seated in the driver seat by detecting a human facial organ from the image included in the captured image.
 本発明に係る運転者の着座判定プログラムは、自動車のコンピュータに、自動車の運転席を、少なくとも1つのカメラで撮影するステップと、前記カメラによって撮影された撮影画像から前記運転席に運転者が着座しているか否かを判断するステップと、を実行させる。 The driver seating determination program according to the present invention includes a step of photographing a driver's seat of a car with at least one camera on a car computer, and a driver seated in the driver's seat from a photographed image taken by the camera. And a step of determining whether or not.
 上記着座判定プログラムにおいては、前記撮影画像から前記運転者の顔を検出することで、前記運転席に運転者が着座していると判断することかできる。 In the seating determination program, it is possible to determine that the driver is seated in the driver seat by detecting the driver's face from the captured image.
 上記着座判定プログラムにおいては、前記撮影画像に含まれる像から人の顔の器官を検出することで、前記運転席に運転者が着座していると判断することができる。 In the seating determination program, it is possible to determine that the driver is seated in the driver seat by detecting a human facial organ from the image included in the photographed image.
 本発明によれば、自動車の運転席に運転者が着座しているか否かを正確に判断することができる。 According to the present invention, it is possible to accurately determine whether or not the driver is seated in the driver's seat of the automobile.
本発明の第1実施形態に係る着座判定装置が取り付けられる自動車の一部概略構成図である。It is a partial schematic block diagram of the motor vehicle with which the seating determination apparatus which concerns on 1st Embodiment of this invention is attached. 本発明の第1実施形態に係る着座判定装置が含まれる着座判定システムの概要を示す図である。It is a figure which shows the outline | summary of the seating determination system in which the seating determination apparatus which concerns on 1st Embodiment of this invention is included. 図2の着座判定装置のハードウェア構成の一例を示すブロック図である。It is a block diagram which shows an example of the hardware constitutions of the seating determination apparatus of FIG. 図2の学習装置のハードウェア構成の一例を示すブロック図である。It is a block diagram which shows an example of the hardware constitutions of the learning apparatus of FIG. 図2の着座判定装置の機能構成の一例を示すブロック図である。It is a block diagram which shows an example of a function structure of the seating determination apparatus of FIG. 図2の学習装置の機能構成の一例を示すブロック図である。It is a block diagram which shows an example of a function structure of the learning apparatus of FIG. 図2の着座判定装置の動作を示すフローチャートである。It is a flowchart which shows operation | movement of the seating determination apparatus of FIG. 本発明の第2実施形態に係る着座判定装置のハードウェア構成の一例を示すブロック図である。It is a block diagram which shows an example of the hardware constitutions of the seating determination apparatus which concerns on 2nd Embodiment of this invention. 第2実施形態に係る着座判定装置の機能構成の一例を示すブロック図である。It is a block diagram which shows an example of a function structure of the seating determination apparatus which concerns on 2nd Embodiment. 第2実施形態に係る学習装置の機能構成の一例を示すブロック図である。It is a block diagram which shows an example of a function structure of the learning apparatus which concerns on 2nd Embodiment. 第2実施形態に係る着座判定装置の動作を示すフローチャートである。It is a flowchart which shows operation | movement of the seating determination apparatus which concerns on 2nd Embodiment.
 <A.第1実施形態>
 以下、本発明に係る運転者の着座判定装置、着座判定方法、及び着座判定プログラムの第1実施形態について、図面を参照しつつ説明する。ただし、以下で説明する本実施形態は、あらゆる点において本発明の例示に過ぎない。本発明の範囲を逸脱することなく種々の改良や変形を行うことができることは言うまでもない。つまり、本発明の実施にあたって、実施形態に応じた具体的構成が適宜採用されてもよい。なお、本実施形態において登場するデータを自然言語により説明しているが、より具体的には、コンピュータが認識可能な疑似言語、コマンド、パラメータ、マシン語等で指定される。これらの点は、後述する第2実施形態についても同様である。
<A. First Embodiment>
Hereinafter, a driver's seating determination device, a seating determination method, and a seating determination program according to a first embodiment of the present invention will be described with reference to the drawings. However, this embodiment described below is only an illustration of the present invention in all respects. It goes without saying that various improvements and modifications can be made without departing from the scope of the present invention. That is, in implementing the present invention, a specific configuration according to the embodiment may be adopted as appropriate. Although data appearing in this embodiment is described in a natural language, more specifically, it is specified by a pseudo language, a command, a parameter, a machine language, or the like that can be recognized by a computer. These points are the same in the second embodiment described later.
 <1.着座判定システムの概要>
 まず、本実施形態に係る着座判定装置が含まれる着座判定システムについて、説明する。図1は着座判定装置が取り付けられる自動車の一部概略構成図であり、図2は着座判定システムの概略構成を示す図である。図1に示すように、このシステムでは、運転席900の前方に配置されたカメラ3により、運転席900を撮影して撮影画像を取得し、この撮影画像から顔の器官(目、鼻、口など)が検出されたとき、運転席900に運転者800が着座していると判定するものである。具体的には、図2に示すように、この着座判定システムは、着座判定装置1、学習装置2、及びカメラ3を備えている。
<1. Overview of seating determination system>
First, a seating determination system including the seating determination device according to the present embodiment will be described. FIG. 1 is a partial schematic configuration diagram of an automobile to which a seating determination device is attached, and FIG. 2 is a diagram illustrating a schematic configuration of a seating determination system. As shown in FIG. 1, in this system, the driver's seat 900 is photographed by the camera 3 disposed in front of the driver's seat 900 to obtain a photographed image, and facial organs (eyes, nose, mouth) are obtained from the photographed image. ) Is detected, the driver 800 is determined to be seated in the driver's seat 900. Specifically, as shown in FIG. 2, the seating determination system includes a seating determination device 1, a learning device 2, and a camera 3.
 着座判定装置1は、例えば、ネットワーク10を介して、学習装置2により作成された学習済みの学習器を取得することができる。ネットワーク10の種類は、例えば、インターネット、無線通信網、移動通信網、電話網、専用網等から適宜選択されてよい。その他、着座判定装置1と学習装置2とを直接接続して、学習器を送信することもできる。あるいは、着座判定装置1と学習装置2とを接続せず、学習装置2で学習された学習済みの学習器を、CD-ROM等の記憶媒体に記憶し、この記憶媒体に記憶された学習器を着座判定装置1に保存することもできる。以下、各装置について詳細に説明する。 The seating determination device 1 can acquire a learned learning device created by the learning device 2 via the network 10, for example. The type of the network 10 may be appropriately selected from, for example, the Internet, a wireless communication network, a mobile communication network, a telephone network, a dedicated network, and the like. In addition, the learning device can be transmitted by directly connecting the seating determination device 1 and the learning device 2. Alternatively, the learning device learned by the learning device 2 is stored in a storage medium such as a CD-ROM without connecting the seating determination device 1 and the learning device 2, and the learning device stored in the storage medium Can also be stored in the seating determination apparatus 1. Hereinafter, each device will be described in detail.
 <1-1.カメラ>
 カメラ3は、デジタルカメラやビデオカメラなど、公知のものを用いることができ、前方から運転席を撮影することで撮影画像を生成し、これを着座判定装置1に出力する。カメラ3は、運転席900において、少なくとも人の顔801が配置される付近を撮影する。このとき、運転者の身長の高低を考慮し、概ねほとんどの運転者の顔が位置する範囲を含むようにする。なお、撮影画像は、静止画または動画のいずれであってもよく、動画の場合には、フレーム毎に着座判定装置1に撮影画像が送信され、着座の判定が行われる。
<1-1. Camera>
As the camera 3, a known device such as a digital camera or a video camera can be used. A photographed image is generated by photographing the driver's seat from the front, and this is output to the seating determination device 1. The camera 3 captures at least the vicinity of the human face 801 in the driver's seat 900. At this time, in consideration of the height of the driver, a range in which almost the driver's face is located is included. The captured image may be either a still image or a moving image. In the case of a moving image, the captured image is transmitted to the seating determination device 1 for each frame, and seating determination is performed.
 <1-2.着座判定装置> <1-2. Seating determination device>
 図3は、本実施形態に係る着座判定装置を示すブロック図である。図3に示すように、本実施形態に係る着座判定装置1は、制御部11、記憶部12、外部インタフェース13、入力装置14、出力装置15、通信インタフェース13、及びドライブ17が電気的に接続されたコンピュータである。なお、図1では、通信インタフェース及び外部インタフェースをそれぞれ、「通信I/F」及び「外部I/F」と記載している。 FIG. 3 is a block diagram showing a seating determination apparatus according to the present embodiment. As shown in FIG. 3, the seating determination apparatus 1 according to the present embodiment is electrically connected to the control unit 11, the storage unit 12, the external interface 13, the input device 14, the output device 15, the communication interface 13, and the drive 17. Computer. In FIG. 1, the communication interface and the external interface are described as “communication I / F” and “external I / F”, respectively.
 制御部11は、CPU(Central Processing Unit)、RAM(Random Access Memory)、ROM(Read Only Memory)等を含み、情報処理に応じて各構成要素の制御を行う。記憶部12は、例えば、ハードディスクドライブ、ソリッドステートドライブ等の補助記憶装置であり、制御部11で実行される着座判定プログラム121、学習済みの学習器に関する情報を示す学習結果データ122等を記憶する。 The control unit 11 includes a CPU (Central Processing Unit), a RAM (Random Access Memory), a ROM (Read Only Memory), and the like, and controls each component according to information processing. The storage unit 12 is, for example, an auxiliary storage device such as a hard disk drive or a solid state drive, and stores a seating determination program 121 executed by the control unit 11, learning result data 122 indicating information related to a learned learning device, and the like. .
 着座判定プログラム121は、撮影された画像から人の顔の器官を検出できるか否かの処理を、着座判定装置1に実行させるためのプログラムである。また、学習結果データ122は、学習済みの学習器の設定を行うためのデータである。詳細は後述する。 The seating determination program 121 is a program for causing the seating determination apparatus 1 to execute a process as to whether or not a human face organ can be detected from a captured image. The learning result data 122 is data for setting a learned learner. Details will be described later.
 通信インタフェース16は、例えば、有線LAN(Local Area Network)モジュール、無線LANモジュール等であり、ネットワークを介した有線又は無線通信を行うためのインタフェースである。入力装置14は、例えば、マウス、キーボード等の入力を行うための装置である。出力装置15は、例えば、ディスプレイ、スピーカ等の出力を行うための装置である。外部インタフェース13は、USB(Universal Serial Bus)ポート等であり、カメラ3、車内のスピーカ、ディスプレイ、速度を制御する装置等の外部装置と接続するためのインタフェースである。なお、車内のディスプレイとは、例えば、ダッシュボードに設けられた、カーナビゲーション用のディスプレイなど、種々のものを適用することができる。また、外部インタフェース13に接続する外部装置は、上記の各装置に限定されなくてもよく、実施の形態に応じて適宜選択されてよい。そのため、外部インタフェース13は、接続する外部装置毎に設けられてもよく、その数は、実施の形態に応じて適宜選択可能である。 The communication interface 16 is, for example, a wired LAN (Local Area Network) module, a wireless LAN module, or the like, and is an interface for performing wired or wireless communication via a network. The input device 14 is a device for performing input using, for example, a mouse or a keyboard. The output device 15 is a device for outputting, for example, a display or a speaker. The external interface 13 is a USB (Universal Serial Bus) port or the like, and is an interface for connecting to an external device such as the camera 3, a speaker in a vehicle, a display, or a device for controlling the speed. For example, various displays such as a display for car navigation provided on a dashboard can be used as the display in the vehicle. Moreover, the external device connected to the external interface 13 may not be limited to each of the above devices, and may be appropriately selected according to the embodiment. Therefore, the external interface 13 may be provided for each external device to be connected, and the number thereof can be selected as appropriate according to the embodiment.
 ドライブ17は、例えば、CD(Compact Disk)ドライブ、DVD(Digital Versatile Disk)ドライブ等であり、記憶媒体91に記憶されたプログラムを読み込むための装置である。ドライブ17の種類は、記憶媒体91の種類に応じて適宜選択されてよい。上記着座判定プログラム121及び/又は学習結果データ122は、この記憶媒体91に記憶されていてもよい。 The drive 17 is, for example, a CD (Compact Disk) drive, a DVD (Digital Versatile Disk) drive, or the like, and is a device for reading a program stored in the storage medium 91. The type of the drive 17 may be appropriately selected according to the type of the storage medium 91. The seating determination program 121 and / or the learning result data 122 may be stored in the storage medium 91.
 記憶媒体91は、コンピュータその他装置、機械等が記録されたプログラム等の情報を読み取り可能なように、当該プログラム等の情報を、電気的、磁気的、光学的、機械的又は化学的作用によって蓄積する媒体である。着座判定装置1は、この記憶媒体91から、着座判定プログラム121及び/又は学習結果データ122を取得してもよい。 The storage medium 91 stores information such as a program by an electrical, magnetic, optical, mechanical, or chemical action so that the information such as a program recorded by a computer or other device or machine can be read. It is a medium to do. The seating determination apparatus 1 may acquire the seating determination program 121 and / or the learning result data 122 from the storage medium 91.
 ここで、図3では、記憶媒体91の一例として、CD、DVD等のディスク型の記憶媒体を例示している。しかしながら、記憶媒体91の種類は、ディスク型に限定される訳ではなく、ディスク型以外であってもよい。ディスク型以外の記憶媒体として、例えば、フラッシュメモリ等の半導体メモリを挙げることができる。 Here, in FIG. 3, as an example of the storage medium 91, a disk-type storage medium such as a CD or a DVD is illustrated. However, the type of the storage medium 91 is not limited to the disk type and may be other than the disk type. Examples of the storage medium other than the disk type include a semiconductor memory such as a flash memory.
 なお、着座判定装置1の具体的なハードウェア構成に関して、実施形態に応じて、適宜、構成要素の省略、置換及び追加が可能である。例えば、制御部11は、複数のプロセッサを含んでもよい。着座判定装置1は、複数台の情報処理装置で構成されてもよい。 It should be noted that regarding the specific hardware configuration of the seating determination device 1, components can be omitted, replaced, and added as appropriate according to the embodiment. For example, the control unit 11 may include a plurality of processors. The seating determination device 1 may be composed of a plurality of information processing devices.
 <1-3.学習装置>
 図4は、本実施形態に係る学習装置を示すブロック図である。図4に示すように、本実施形態に係る学習装置2は、上記第2検出部102に含まれる学習器を学習するためのものであり、制御部21、記憶部22、通信インタフェース23、入力装置24、出力装置25、外部インタフェース26、及びドライブ27が電気的に接続されたコンピュータである。なお、図4では、図3と同様に、通信インタフェース及び外部インタフェースをそれぞれ、「通信I/F」及び「外部I/F」と記載している。
<1-3. Learning device>
FIG. 4 is a block diagram illustrating the learning device according to the present embodiment. As shown in FIG. 4, the learning device 2 according to the present embodiment is for learning the learning device included in the second detection unit 102, and includes a control unit 21, a storage unit 22, a communication interface 23, and an input. A computer in which the device 24, the output device 25, the external interface 26, and the drive 27 are electrically connected. In FIG. 4, as in FIG. 3, the communication interface and the external interface are described as “communication I / F” and “external I / F”, respectively.
 制御部21~ドライブ27及び記憶媒体92はそれぞれ、上記着座判定装置1の制御部11~ドライブ17及び記憶媒体91と同様である。ただし、学習装置2の記憶部22は、制御部21で実行される学習プログラム221、学習器の学習に利用する学習データ222、学習プログラム221を実行して作成した学習結果データ122等を記憶する。  The control unit 21 to the drive 27 and the storage medium 92 are the same as the control unit 11 to the drive 17 and the storage medium 91 of the seating determination device 1, respectively. However, the storage unit 22 of the learning device 2 stores a learning program 221 executed by the control unit 21, learning data 222 used for learning of the learning device, learning result data 122 created by executing the learning program 221, and the like. . *
 学習プログラム221は、学習装置2に後述するニューラルネットワークの学習処理(図8)を実行させるためのプログラムである。また、学習データ222は、撮影画像から人の顔の器官を検出するために学習器の学習を行うためのデータである。詳細は後述する。 The learning program 221 is a program for causing the learning device 2 to execute a neural network learning process (FIG. 8) described later. The learning data 222 is data for performing learning of a learning device in order to detect a human facial organ from a captured image. Details will be described later.
 なお、上記着座判定装置1と同様に、学習プログラム221及び/又は学習データ222は、記憶媒体92に記憶されていてもよい。これに応じて、学習装置2は、利用する学習プログラム221及び/又は学習データ222を記憶媒体92から取得してもよい。 Note that the learning program 221 and / or the learning data 222 may be stored in the storage medium 92 as in the seating determination apparatus 1. In response to this, the learning device 2 may acquire the learning program 221 and / or the learning data 222 to be used from the storage medium 92.
 また、上記着座判定装置1と同様に、学習装置2の具体的なハードウェア構成に関して、実施形態に応じて、適宜、構成要素の省略、置換及び追加が可能である。更に、学習装置2は、提供されるサービス専用に設計された情報処理装置の他、汎用のサーバ装置、デスクトップPC等が用いられてもよい。 Also, as with the seating determination device 1, regarding the specific hardware configuration of the learning device 2, components can be omitted, replaced, and added as appropriate according to the embodiment. Further, the learning device 2 may be a general-purpose server device, a desktop PC, or the like, in addition to an information processing device designed exclusively for the provided service.
 <2.着座判定装置の機能構成>
 次に、図5を参照しつつ、本実施形態に係る着座判定装置1の機能構成の一例を説明する。図5は、本実施形態に係る着座判定装置1の機能構成の一例を模式的に例示する。
<2. Functional configuration of seating determination device>
Next, an example of a functional configuration of the seating determination device 1 according to the present embodiment will be described with reference to FIG. FIG. 5 schematically illustrates an example of a functional configuration of the seating determination apparatus 1 according to the present embodiment.
 <2-1.概略構成>
 着座判定装置1の制御部11は、記憶部12に記憶された着座判定プログラム121をRAMに展開する。そして、制御部11は、RAMに展開された着座判定プログラム121をCPUにより解釈及び実行して、各構成要素を制御する。これによって、図5に示すように、本実施形態に係る着座判定装置1は、画像取得部111、解析部116、及び警告部117を備えるコンピュータとして機能する。
<2-1. Schematic configuration>
The control unit 11 of the seating determination device 1 expands the seating determination program 121 stored in the storage unit 12 in the RAM. And the control part 11 interprets and performs the seating determination program 121 expand | deployed by RAM by CPU, and controls each component. Accordingly, as illustrated in FIG. 5, the seating determination apparatus 1 according to the present embodiment functions as a computer including an image acquisition unit 111, an analysis unit 116, and a warning unit 117.
 画像取得部111は、カメラ3で生成された撮影画像123を取得する。また、解析部116は、撮影画像123から運転者が運転席に着座しているか否かを判定する。そして、解析部116が、運転者が運転席に着座していないと判定したときには、警告部117が警告を発するように構成されている。以下、これらの機能構成について、詳細に説明する。 The image acquisition unit 111 acquires the captured image 123 generated by the camera 3. Further, the analysis unit 116 determines whether or not the driver is seated in the driver's seat from the captured image 123. When the analysis unit 116 determines that the driver is not seated in the driver's seat, the warning unit 117 is configured to issue a warning. Hereinafter, these functional configurations will be described in detail.
 <2-2.解析部>
 まず、解析部116について説明する。図5に示すように、解析部116では、撮影画像123を、顔の器官を検出するために学習した学習器の入力として用いる。この学習器の演算処理により、当該学習器から出力値を得る。そして、解析部116は、学習器から得られた出力値に基づいて、撮影画像123の中の人の顔の器官が存在するか否かを判定する。なお、顔の器官とは、目、鼻、口などが該当するが、これらの少なくとも1つの特徴点を検出できるようにする。但し、カメラの種類によっては、運転者がサングラスを着用している場合には目の検出ができない場合があるため、例えば、鼻や口の特徴点を検出できるようにする。また、運転者がマスクを着用している場合には、鼻や口が検出できないため、例えば、目の特徴点を検出できるようにする。
<2-2. Analysis Department>
First, the analysis unit 116 will be described. As shown in FIG. 5, the analysis unit 116 uses the photographed image 123 as an input of a learning device learned to detect a facial organ. An output value is obtained from the learning device by the arithmetic processing of the learning device. Then, the analysis unit 116 determines whether or not a human face organ in the captured image 123 exists based on the output value obtained from the learning device. The facial organ includes eyes, nose, mouth, and the like, and at least one of these feature points can be detected. However, depending on the type of camera, the eyes may not be detected when the driver is wearing sunglasses. For example, feature points of the nose and mouth can be detected. In addition, when the driver wears a mask, the nose and mouth cannot be detected, so that, for example, the eye feature point can be detected.
 次に、学習器について説明する。図5に示すように、本実施形態に係る着座判定装置1は、一例として、撮影画像123中の顔の器官の有無について学習した学習器が用いられる。この学習器7はニューラルネットワークで構成されている。具体的には、図4に示すような、いわゆる深層学習に用いられる多層構造のニューラルネットワークであり、入力から順に、入力層71、中間層(隠れ層)72、及び出力層73を備えている。 Next, the learning device will be described. As shown in FIG. 5, the seating determination apparatus 1 according to the present embodiment uses, as an example, a learning device that learns about the presence or absence of a facial organ in the captured image 123. The learning device 7 is composed of a neural network. Specifically, it is a neural network having a multilayer structure used for so-called deep learning as shown in FIG. 4, and includes an input layer 71, an intermediate layer (hidden layer) 72, and an output layer 73 in order from the input. .
 図5の例では、ニューラルネットワーク7は1層の中間層72を備えており、入力層71の出力が中間層72の入力となり、中間層72の出力が出力層73の入力となっている。ただし、中間層72の数は1層に限られなくてもよく、ニューラルネットワーク7は、中間層72を2層以上備えてもよい。 In the example of FIG. 5, the neural network 7 includes one intermediate layer 72, the output of the input layer 71 is the input of the intermediate layer 72, and the output of the intermediate layer 72 is the input of the output layer 73. However, the number of intermediate layers 72 is not limited to one, and the neural network 7 may include two or more intermediate layers 72.
 各層71~73は、1又は複数のニューロンを備えている。例えば、入力層71のニューロンの数は、各撮影画像123の画素数に応じて設定することができる。中間層72のニューロンの数は実施の形態に応じて適宜設定することができる。また、出力層73は、顔の器官の有無の判定に応じて設定することができる。 Each layer 71 to 73 includes one or a plurality of neurons. For example, the number of neurons in the input layer 71 can be set according to the number of pixels in each captured image 123. The number of neurons in the intermediate layer 72 can be set as appropriate according to the embodiment. The output layer 73 can be set according to the determination of the presence or absence of a facial organ.
 隣接する層のニューロン同士は適宜結合され、各結合には重み(結合荷重)が設定されている。図5の例では、各ニューロンは、隣接する層の全てのニューロンと結合されているが、ニューロンの結合は、このような例に限定されなくてもよく、実施の形態に応じて適宜設定されてよい。 Adjacent layers of neurons are appropriately connected to each other, and a weight (connection load) is set for each connection. In the example of FIG. 5, each neuron is connected to all neurons in the adjacent layers, but the neuron connection is not limited to such an example, and is appropriately set according to the embodiment. It's okay.
 各ニューロンには閾値が設定されており、基本的には、各入力と各重みとの積の和が閾値を超えているか否かによって各ニューロンの出力が決定される。着座判定装置1は、このようなニューラルネットワーク7の入力層71に上記各撮影画像を入力することで出力層73から得られる出力値に基づいて、運転者が運転席に着座しているか否かを判定する。 A threshold is set for each neuron, and basically, the output of each neuron is determined by whether or not the sum of products of each input and each weight exceeds the threshold. The seating determination apparatus 1 determines whether or not the driver is seated in the driver's seat based on the output value obtained from the output layer 73 by inputting the respective captured images to the input layer 71 of the neural network 7. Determine.
 なお、このようなニューラルネットワーク7の構成(例えば、ニューラルネットワーク7の層数、各層におけるニューロンの個数、ニューロン同士の結合関係、各ニューロンの伝達関数)、各ニューロン間の結合の重み、及び各ニューロンの閾値を示す情報は、学習結果データ122に含まれている。着座判定装置1は、学習結果データ122を参照して、運転者が運転席に着座しているか否かを判定するための処理に用いる学習済みの学習器7の設定を行う。この点は、後述する第2実施形態においても同様である。 The configuration of the neural network 7 (for example, the number of layers of the neural network 7, the number of neurons in each layer, the connection relationship between neurons, the transfer function of each neuron), the weight of connection between each neuron, and each neuron The information indicating the threshold value is included in the learning result data 122. The seating determination device 1 refers to the learning result data 122 and sets the learned learning device 7 used for processing for determining whether or not the driver is seated in the driver's seat. This also applies to a second embodiment described later.
 <2-3.警告部>
 警告部117は、上記解析部116で、運転席に運転者が着座していないと判定したとき、外部インタフェース16を通じて車内のディスプレイやスピーカ等を駆動し、警告を行う。すなわち、運転者が着座していないことをディスプレイに表示したり、スピーカを通じて、着座していないことを車内に報知する。その他、ブレーキを駆動するなど、走行中の自動車の速度を落としたり、停止するなどで警告を行うこともできる。
<2-3. Warning section>
When the analysis unit 116 determines that the driver is not seated in the driver's seat, the warning unit 117 drives a display, a speaker, and the like in the vehicle through the external interface 16 to give a warning. That is, it is displayed on the display that the driver is not seated, or the vehicle is informed through the speaker that the driver is not seated. In addition, a warning can also be given by reducing the speed of a running car, such as driving a brake, or stopping.
 <3.学習装置の機能構成>
 次に、図6を用いて、本実施形態に係る学習装置2の機能構成の一例を説明する。図6は、本実施形態に係る学習装置2の機能構成の一例を模式的に例示する。
<3. Functional configuration of learning device>
Next, an example of a functional configuration of the learning device 2 according to the present embodiment will be described with reference to FIG. FIG. 6 schematically illustrates an example of a functional configuration of the learning device 2 according to the present embodiment.
 学習装置2の制御部21は、記憶部22に記憶された学習プログラム221をRAMに展開する。そして、制御部21は、RAMに展開された学習プログラム221をCPUにより解釈及び実行して、各構成要素を制御する。これによって、図6に示されるとおり、本実施形態に係る学習装置2は、学習データ取得部211、及び学習処理部212を備えるコンピュータとして機能する。 The control unit 21 of the learning device 2 expands the learning program 221 stored in the storage unit 22 in the RAM. Then, the control unit 21 interprets and executes the learning program 221 expanded in the RAM, and controls each component. Accordingly, as illustrated in FIG. 6, the learning device 2 according to the present embodiment functions as a computer including the learning data acquisition unit 211 and the learning processing unit 212.
 学習データ取得部211は、図6に示すように、学習データ222として、カメラ3で撮影した撮影画像223と、この撮影画像223に顔の器官が示されているか否かを示す着座情報2241の組を取得する。この撮影画像223と着座情報2241が、ニューラルネットワーク8の教師データに相当する。 As shown in FIG. 6, the learning data acquisition unit 211 includes, as learning data 222, a captured image 223 captured by the camera 3 and seating information 2241 indicating whether or not a facial organ is indicated in the captured image 223. Get a pair. The captured image 223 and the seating information 2241 correspond to the teacher data of the neural network 8.
 一方、学習処理部212は、学習データ222を用いて、取得した各撮影画像223を入力すると、上記着座情報2241に対応する出力値を出力するようにニューラルネットワーク8を学習させる。 On the other hand, when the acquired captured image 223 is input using the learning data 222, the learning processing unit 212 causes the neural network 8 to learn so as to output an output value corresponding to the seating information 2241.
 図6に示すように、学習器の一例であるニューラルネットワーク8は、入力層81、中間層(隠れ層)82、及び出力層83を備え、上記ニューラルネットワーク7と同様に構成される。各層81~83は、上記各層71~73と同様である。学習処理部212は、ニューラルネットワークの学習処理により、撮影画像223を入力すると、上記着座情報2241に対応する出力値を出力するニューラルネットワーク8を構築する。そして、学習処理部212は、構築したニューラルネットワーク8の構成、各ニューロン間の結合の重み、及び各ニューロンの閾値を示す情報を学習結果データ122として記憶部22に格納する。そして、この学習結果データ122は、上述した種々の方法で、着座判定装置1に送信される。また、このような学習結果データ122を定期的に更新してもよい。そして、制御部21は、作成した学習結果データ122を学習処理の実行毎に着座判定装置1に転送することで、着座判定装置1の保持する学習結果データ122を定期的に更新してもよい。 As shown in FIG. 6, a neural network 8 as an example of a learning device includes an input layer 81, an intermediate layer (hidden layer) 82, and an output layer 83, and is configured in the same manner as the neural network 7. The layers 81 to 83 are the same as the layers 71 to 73 described above. The learning processing unit 212 constructs the neural network 8 that outputs an output value corresponding to the seating information 2241 when the captured image 223 is input by the learning processing of the neural network. The learning processing unit 212 stores information indicating the configuration of the constructed neural network 8, the weight of the connection between the neurons, and the threshold value of each neuron as the learning result data 122 in the storage unit 22. The learning result data 122 is transmitted to the seating determination apparatus 1 by the various methods described above. Further, such learning result data 122 may be periodically updated. And the control part 21 may update the learning result data 122 which the seating determination apparatus 1 hold | maintains by transferring the created learning result data 122 to the seating determination apparatus 1 for every execution of a learning process. .
 <4.その他>
 着座判定装置1及び学習装置2の各機能に関しては後述する動作例で詳細に説明する。なお、本実施形態では、着座判定装置1及び学習装置2の各機能がいずれも汎用のCPUによって実現される例について説明している。しかしながら、以上の機能の一部又は全部が、1又は複数の専用のプロセッサにより実現されてもよい。また、着座判定装置1及び学習装置2それぞれの機能構成に関して、実施形態に応じて、適宜、機能の省略、置換及び追加が行われてもよい。
<4. Other>
Each function of the seating determination device 1 and the learning device 2 will be described in detail in an operation example described later. In the present embodiment, an example in which each function of the seating determination device 1 and the learning device 2 is realized by a general-purpose CPU is described. However, part or all of the above functions may be realized by one or a plurality of dedicated processors. In addition, regarding the functional configurations of the seating determination device 1 and the learning device 2, functions may be omitted, replaced, and added as appropriate according to the embodiment.
 <5.着座判定装置の動作>
 次に、図7を参照しつつ、着座判定装置1の動作例を説明する。図7は、着座判定装置1の処理手順の一例を例示するフローチャートである。なお、以下で説明する処理手順は一例に過ぎず、各処理は可能な限り変更されてよい。また、以下で説明する処理手順について、実施の形態に応じて、適宜、ステップの省略、置換、及び追加が可能である。
<5. Operation of Seating Determination Device>
Next, an operation example of the seating determination device 1 will be described with reference to FIG. FIG. 7 is a flowchart illustrating an example of a processing procedure of the seating determination apparatus 1. Note that the processing procedure described below is merely an example, and each processing may be changed as much as possible. Further, in the processing procedure described below, steps can be omitted, replaced, and added as appropriate according to the embodiment.
 まず、利用者(運転者)は、着座判定装置1を起動し、起動した着座判定装置1に着座判定プログラム121を実行させる。着座判定装置1の制御部11は、学習結果データ122を参照して、ニューラルネットワーク7の構造、各ニューロン間の結合の重み及び各ニューロンの閾値の設定を行う。そして、制御部11は、以下の処理手順に従って、撮影画像123から、運転者が運転席に着座しているか否かを判定する。 First, the user (driver) activates the seating determination apparatus 1 and causes the activated seating determination apparatus 1 to execute the seating determination program 121. The control unit 11 of the seating determination apparatus 1 refers to the learning result data 122 and sets the structure of the neural network 7, the weight of connection between neurons, and the threshold value of each neuron. And the control part 11 determines whether the driver | operator is seated in the driver's seat from the picked-up image 123 according to the following processing procedures.
 まず、制御部11は、自動車の運転が行われている場合(ステップS101のYES)、画像取得部111として機能し、外部インタフェース16を介して接続されるカメラ3から、運転席を前方から撮影した撮影画像123を取得する(ステップS102)。上述したように、撮影画像123は、静止画でもよいし、動画である場合は、フレームごとに撮影画像が取得される。 First, when the vehicle is being driven (YES in step S101), the control unit 11 functions as the image acquisition unit 111 and photographs the driver's seat from the front from the camera 3 connected via the external interface 16. The acquired captured image 123 is acquired (step S102). As described above, the captured image 123 may be a still image, or in the case of a moving image, a captured image is acquired for each frame.
 次に、制御部11は、解析部116として機能し、ステップS102で取得した各撮影画像123に顔の器官が含まれているか否かを判断する(ステップS103)。そして、撮影画像123の中に顔の器官を検出した場合、運転席に運転者が着座していると判定する(ステップS103のYES)。その後、運転が行われていれば(ステップS101のYES)、引き続き、撮影画像123を取得し(ステップS102)、運転者の着座を判定する(ステップS103)。一方、撮影画像123中に顔の器官の検出ができない場合には、運転席に運転者が着座していないと判定し(ステップS101のNO)、警告を発信する(ステップS104)。すなわち、制御部11は、警告部117として機能し、車内のディスプレイまたはスピーカを用いて、運転席に運転者が着座していないことを車内に報知する。あるいは、自動車を減速したり、停止させることもできる。その後、運転が行われていれば(ステップS101のYES)、引き続き、撮影画像123を取得し(ステップS102)、運転者の着座を判定する(ステップS103)。一方、運転が行われていない場合には、処理を停止する。 Next, the control unit 11 functions as the analysis unit 116 and determines whether or not a facial organ is included in each captured image 123 acquired in step S102 (step S103). If a facial organ is detected in the captured image 123, it is determined that the driver is seated in the driver's seat (YES in step S103). Thereafter, if driving is being performed (YES in step S101), the captured image 123 is continuously acquired (step S102), and the driver's seating is determined (step S103). On the other hand, if the facial organ cannot be detected in the captured image 123, it is determined that the driver is not seated in the driver's seat (NO in step S101), and a warning is transmitted (step S104). That is, the control unit 11 functions as the warning unit 117 and notifies the vehicle that the driver is not seated in the driver's seat using the display or speaker in the vehicle. Alternatively, the automobile can be decelerated or stopped. Thereafter, if driving is being performed (YES in step S101), the captured image 123 is continuously acquired (step S102), and the driver's seating is determined (step S103). On the other hand, when the operation is not performed, the process is stopped.
 なお、運転席に運転者が着座しているか否かの判定に当たって、撮影画像123から顔の器官を検出できない場合、上記のように、即座に警告を発してもよいが、例えば、所定時間(または所定の数のフレーム)、顔の器官が検出できない場合に、警告を発することもできる。 When determining whether or not the driver is seated in the driver's seat, if the facial organ cannot be detected from the captured image 123, a warning may be issued immediately as described above. (Alternatively, a predetermined number of frames), a warning can be issued if a facial organ cannot be detected.
 また、上記の処理は、自動車のイグニッション電源がONになった直後から行ってもよいし、例えば、自動車が手動運転モードと自動運転モードとを切替可能な場合、自動車が自動運転モードに移行した場合にのみ行ってもよい。 The above processing may be performed immediately after the ignition power of the automobile is turned on. For example, when the automobile can be switched between the manual operation mode and the automatic operation mode, the automobile has shifted to the automatic operation mode. You may only do that.
 <6.特徴>
 以上のように、本実施形態によれば、カメラ3によって運転席を撮影した撮影画像123を取得し、この撮影画像123に人の顔の器官が含まれているか否かを解析することで、運転者が運転席に着座しているか否かを判断するようにしている。したがって、運転者が運転席に着座していることを確実に判断することができる。
<6. Features>
As described above, according to the present embodiment, the captured image 123 obtained by capturing the driver's seat with the camera 3 is acquired, and by analyzing whether or not the captured image 123 includes a human facial organ, It is determined whether or not the driver is seated in the driver's seat. Therefore, it can be reliably determined that the driver is seated in the driver's seat.
 特に、着座の判定に当たっては、ニューラルネットワークにより構成された学習器7によって判定を行っている。すなわち、学習器7は、多くの撮影画像123から顔の器官を検出するための学習がなされているため、精度の高い判定を行うことができる。 In particular, in the determination of sitting, the determination is performed by the learning device 7 configured by a neural network. That is, since the learning device 7 has learned to detect a facial organ from many captured images 123, it can make a highly accurate determination.
 <B.第2実施形態>
 次に、本発明に係る運転者の着座判定装置、着座判定方法、及び着座判定プログラムの第2実施形態について、図面を参照しつつ説明する。
<B. Second Embodiment>
Next, a second embodiment of the driver's seating determination apparatus, seating determination method, and seating determination program according to the present invention will be described with reference to the drawings.
 <1.着座判定システムの概要>
 まず、本実施形態に係る着座判定装置が含まれる着座判定システムについて、説明する。但し、本実施形態に係る着座判定システムは、第1実施形態と同様に、着座判定装置1及び学習装置2を備えている。
<1. Overview of seating determination system>
First, a seating determination system including the seating determination device according to the present embodiment will be described. However, the seating determination system according to the present embodiment includes the seating determination device 1 and the learning device 2 as in the first embodiment.
 本実施形態に係る着座判定装置1は、図1及び図2に示す第1実施形態と同様に、車両の運転席に着いた運転者800を撮影するように配置されたカメラ3から撮影画像を取得する。また、この着座判定装置1は、運転者800の顔の挙動に関する顔挙動情報を含む運転者の観測情報を取得する。そして、着座判定装置1は、運転者800が運転席900に着座しているか否かを判定するための学習を行った学習済みの学習器(後述するニューラルネットワーク)に、取得した撮影画像及び観測情報を入力することで、運転者800が運転席900に着座しているか否かを判定する。 As in the first embodiment shown in FIGS. 1 and 2, the seating determination device 1 according to the present embodiment captures a captured image from a camera 3 that is arranged to capture a driver 800 that has arrived at the driver's seat of the vehicle. get. In addition, the seating determination apparatus 1 acquires driver observation information including face behavior information related to the behavior of the driver 800 face. Then, the seating determination device 1 uses the acquired captured image and observation to a learned learning device (a neural network described later) that has performed learning to determine whether or not the driver 800 is seated in the driver's seat 900. By inputting information, it is determined whether or not the driver 800 is seated in the driver's seat 900.
 一方、本実施形態に係る学習装置2は、着座判定装置1で利用する学習器を構築する、すなわち、撮影画像及び観測情報の入力に応じて、運転者800が運転席に着座しているか否かを示す着座情報を出力するように学習器の機械学習を行うコンピュータである。具体的には、学習装置2は、上記の撮影画像、観測情報、及び着座情報の組を学習データとして取得する。そして、学習装置2は、撮影画像及び観測情報を入力すると着座情報に対応する出力値を出力するように学習器(後述するニューラルネットワーク6)を学習させる。これにより、着座判定装置1で利用する学習済みの学習器が作成される。着座判定装置1と学習装置2と接続は、第1実施形態と同様である。 On the other hand, the learning device 2 according to the present embodiment constructs a learning device to be used in the seating determination device 1, that is, whether or not the driver 800 is seated in the driver's seat according to the input of the captured image and the observation information. It is a computer that performs machine learning of a learning device so as to output seating information indicating the above. Specifically, the learning device 2 acquires a set of the above-described captured image, observation information, and seating information as learning data. Then, the learning device 2 causes the learning device (a neural network 6 described later) to learn so as to output an output value corresponding to the seating information when the captured image and the observation information are input. As a result, a learned learning device used in the seating determination apparatus 1 is created. The connection between the seating determination device 1 and the learning device 2 is the same as that in the first embodiment.
 以上のとおり、本実施形態では、運転者800の状態を推定するために、運転者の着座を推定するための学習を行った学習済みの学習器を利用する。そして、この学習器の入力として、運転者の顔の挙動に関する顔挙動情報を含む、運転者を観測することで得られる観測情報の他に、車両の運転席に着いた運転者を撮影するように配置されたカメラ3から得られる撮影画像を用いる。そのため、運転者800の顔の挙動だけではなく、運転者800の身体の状態(例えば、身体の向き、姿勢等)を撮影画像から解析することができる。したがって、本実施形態によれば、運転者800の取り得る多様な状態を反映して、運転者800が運転席900に着座しているか否かを判定することができる。以下、詳細に説明する。 As described above, in this embodiment, in order to estimate the state of the driver 800, a learned learning device that has performed learning for estimating the seating of the driver is used. In addition to the observation information obtained by observing the driver, including the face behavior information related to the behavior of the driver's face as an input to the learning device, the driver who took the driver's seat is photographed. A photographed image obtained from the camera 3 arranged in the above is used. Therefore, not only the behavior of the driver 800's face but also the state of the driver's 800 body (for example, body orientation, posture, etc.) can be analyzed from the captured image. Therefore, according to the present embodiment, it is possible to determine whether or not the driver 800 is seated in the driver's seat 900, reflecting various states that the driver 800 can take. Details will be described below.
 <2.着座判定装置>
 まず、図8を用いて、本実施形態に係る着座判定装置について説明する。図8は、本実施形態に係る着座判定装置のブロック図である。
<2. Seating determination device>
First, the seating determination apparatus according to the present embodiment will be described with reference to FIG. FIG. 8 is a block diagram of the seating determination apparatus according to the present embodiment.
 図8に示すように、本実施形態に係る着座判定装置のハードウェア構成は、第1実施形態と概ね同じであり、外部I/Fに接続される装置が相違している。したがって、以下では、第1実施形態と相違する点についてのみ説明し、同一構成には同一符号を付して説明を省略する。本実施形態では、外部インタフェース13は、例えば、CAN(Controller Area Network)を介して、上述したカメラ3のほか、ナビゲーション装置30、生体センサ32、及びスピーカ33に接続される。 As shown in FIG. 8, the hardware configuration of the seating determination device according to the present embodiment is substantially the same as that of the first embodiment, and the devices connected to the external I / F are different. Therefore, hereinafter, only differences from the first embodiment will be described, and the same components will be denoted by the same reference numerals and description thereof will be omitted. In the present embodiment, the external interface 13 is connected to the navigation device 30, the biosensor 32, and the speaker 33 in addition to the above-described camera 3 via, for example, CAN (Controller (Area Network).
 ナビゲーション装置30は、車両の走行時に経路案内を行うコンピュータである。ナビゲーション装置30には、公知のカーナビゲーション装置が用いられてよい。ナビゲーション装置30は、GPS(Global Positioning System)信号に基づいて自車位置を測定し、地図情報及び周辺の建物等に関する周辺情報を利用して、経路案内を行うように構成される。なお、以下では、GPS信号に基づいて測定される自車位置を示す情報を「GPS情報」と称する。 The navigation device 30 is a computer that provides route guidance when the vehicle is traveling. A known car navigation device may be used as the navigation device 30. The navigation device 30 is configured to measure the position of the vehicle based on a GPS (Global Positioning System) signal, and to perform route guidance using map information and surrounding information on surrounding buildings and the like. Hereinafter, information indicating the vehicle position measured based on the GPS signal is referred to as “GPS information”.
 生体センサ32は、運転者800の生体情報を測定するように構成される。測定対象となる生体情報は、特に限定されなくてもよく、例えば、脳波、心拍数等であってよい。生体センサ32は、測定対象となる生体情報を測定可能であれば特に限定されなくてもよく、例えば、公知の脳波センサ、脈拍センサ等が用いられてよい。生体センサ32は、測定対象となる生体情報に応じた運転者800の身体部位に装着される。 The biological sensor 32 is configured to measure the biological information of the driver 800. The biological information to be measured is not particularly limited, and may be, for example, an electroencephalogram, a heart rate, or the like. The biological sensor 32 is not particularly limited as long as biological information to be measured can be measured. For example, a known brain wave sensor, pulse sensor, or the like may be used. The biosensor 32 is attached to the body part of the driver 800 corresponding to the biometric information to be measured.
 スピーカ33は、音声を出力するように構成される。スピーカ33は、車両の走行中に運転者800が当該車両の運転に適した状態ではない場合に、当該車両の運転に適した状態をとるように当該運転者800に対して警告するのに利用される。詳細は後述する。 The speaker 33 is configured to output sound. The speaker 33 is used to warn the driver 800 to take a state suitable for driving the vehicle when the driver 800 is not in a state suitable for driving the vehicle while the vehicle is running. Is done. Details will be described later.
 <3.学習装置>
 本実施形態に係る学習装置のハードウェア構成は、第1実施形態の学習装置と同じであるため、ここでは説明を省略する。
<3. Learning device>
Since the hardware configuration of the learning device according to the present embodiment is the same as that of the learning device of the first embodiment, the description thereof is omitted here.
 <4.着座判定装置の機能構成>
 次に、図9を用いて、本実施形態に係る着座判定装置1の機能構成の一例を説明する。図9は、本実施形態に係る着座判定装置1の機能構成の一例を模式的に例示する。
<4. Functional configuration of seating determination device>
Next, an example of a functional configuration of the seating determination apparatus 1 according to the present embodiment will be described with reference to FIG. FIG. 9 schematically illustrates an example of a functional configuration of the seating determination apparatus 1 according to the present embodiment.
 図9に示すように、着座判定装置1の制御部11は、記憶部12に記憶されたプログラム121をRAMに展開する。そして、制御部11は、RAMに展開されたプログラム121をCPUにより解釈及び実行して、各構成要素を制御する。これによって、図9に示されるとおり、本実施形態に係る着座判定装置1は、画像取得部111、観測情報取得部112、解像度変換部113、運転状態推定部114、及び警告部115を備えるコンピュータとして機能する。このうち、解像度変換部113及び運転状態推定部114が、本発明の解析部に相当する。 As shown in FIG. 9, the control unit 11 of the seating determination apparatus 1 expands the program 121 stored in the storage unit 12 in the RAM. The control unit 11 interprets and executes the program 121 expanded in the RAM by the CPU and controls each component. Accordingly, as shown in FIG. 9, the seating determination apparatus 1 according to the present embodiment includes a computer including an image acquisition unit 111, an observation information acquisition unit 112, a resolution conversion unit 113, a driving state estimation unit 114, and a warning unit 115. Function as. Among these, the resolution conversion unit 113 and the operation state estimation unit 114 correspond to the analysis unit of the present invention.
 画像取得部111は、車両の運転席に着いた運転者800を撮影するように配置されたカメラ31から撮影画像123を取得する。観測情報取得部112は、運転者800の顔の挙動に関する顔挙動情報1241及び生体センサ32により測定された生体情報1242を含む観測情報124を取得する。本実施形態では、顔挙動情報1241は、撮影画像123を画像解析することで得られる。なお、観測情報124は、このような例に限定されなくてもよく、生体情報1242は、省略されてもよい。この場合、生体センサ32は省略されてもよい。すなわち、撮影画像123から取得した顔挙動情報1241のみを観測情報124として使用することもできる。 The image acquisition unit 111 acquires the captured image 123 from the camera 31 that is arranged so as to capture the driver 800 seated in the driver's seat of the vehicle. The observation information acquisition unit 112 acquires observation information 124 including face behavior information 1241 related to the behavior of the face of the driver 800 and biological information 1242 measured by the biological sensor 32. In the present embodiment, the face behavior information 1241 is obtained by image analysis of the captured image 123. Note that the observation information 124 may not be limited to such an example, and the biological information 1242 may be omitted. In this case, the biosensor 32 may be omitted. That is, only the face behavior information 1241 acquired from the captured image 123 can be used as the observation information 124.
 解像度変換部113は、画像取得部111により取得した撮影画像123の解像度を低下させる。これにより、解像度変換部113は、低解像度撮影画像1231を形成する。 The resolution conversion unit 113 reduces the resolution of the captured image 123 acquired by the image acquisition unit 111. Thereby, the resolution conversion unit 113 forms a low-resolution captured image 1231.
 運転状態推定部114は、運転者が運転席に着座しているか否かの反転を行うための学習を行った学習済みの学習器(ニューラルネットワーク5)に、撮影画像123を低解像度化することで得られた低解像度撮影画像1231及び観測情報124を入力する。これにより、運転状態推定部114は、運転者800の着座に関する着座情報125を当該学習器から取得する。なお、低解像度化の処理は省略されてもよい。この場合、運転状態推定部114は、撮影画像123を学習器に入力してもよい。 The driving state estimation unit 114 reduces the resolution of the captured image 123 to a learned learning device (neural network 5) that has performed learning for reversing whether or not the driver is seated in the driver's seat. The low-resolution captured image 1231 and the observation information 124 obtained in the above are input. Accordingly, the driving state estimation unit 114 acquires the sitting information 125 related to the sitting of the driver 800 from the learning device. Note that the resolution reduction process may be omitted. In this case, the driving state estimation unit 114 may input the captured image 123 to the learning device.
 ところで、運転席900に運転者800が着座しているか否かを判定するに当たっては、誤判定がなされる可能性がある。例えば、助手席の乗車者が運転席900へ顔を含む体を乗り出している場合、後部座席の乗車者が運転席900へ顔を含む体を乗り出している場合などに、運転席900に運転者800が着座していると判断されるおそれがある。また、運転席900に着座しているにもかかわらず、運転者が着座していないと判断される可能性もある。例えば、運転者800がうつむいている場合、あるいは後を向いている場合には、後述するように顔の器官を検出できず、運転者800が着座していないと判断されるおそれがある。したがって、この着座判定装置1では、後述するように、観測情報124に含まれる顔の器官に関する情報を判定の材料とすることに加え、低解像度撮影画像1231から運転席に写る者の身体の状態(例えば、身体の向き、姿勢等)を判定の材料としている。このような身体の状態から、運転席に写る者が、運転席に座る者であるか、助手席や後部座席の者であるか、あるいは子供であるかを判定することができる。 By the way, in determining whether or not the driver 800 is seated in the driver's seat 900, there is a possibility that an erroneous determination is made. For example, when the passenger in the front passenger seat has a body with a face on the driver's seat 900, or when the passenger in the rear seat has a body with a face on the driver's seat 900, the driver can enter the driver's seat 900. There is a risk that 800 is seated. Further, there is a possibility that it is determined that the driver is not seated even though the driver seat 900 is seated. For example, when the driver 800 is depressed or facing backward, the facial organ cannot be detected as will be described later, and it may be determined that the driver 800 is not seated. Therefore, in the seating determination device 1, as described later, in addition to using information on facial organs included in the observation information 124 as a material for determination, the state of the body of the person shown in the driver's seat from the low-resolution captured image 1231 (For example, body orientation, posture, etc.) is used as a material for determination. From such a physical state, it can be determined whether the person in the driver's seat is a person sitting in the driver's seat, a passenger seat or a rear seat, or a child.
 警告部115は、第1実施形態と同じであり、運転席900に運転者800が着座していないと判定したとき、外部インタフェース16を通じて車内のディスプレイやスピーカ等を駆動し、警告を行う。 The warning unit 115 is the same as that of the first embodiment, and when it is determined that the driver 800 is not seated in the driver's seat 900, the display unit and the speaker in the vehicle are driven through the external interface 16 to give a warning.
 (学習器)
 次に、学習器について説明する。図9に示されるとおり、本実施形態に係る着座判定装置1は、運転者800が運転席900に着座しているか否かの判定のための学習を行った学習済みの学習器として、ニューラルネットワーク5を利用する。本実施形態に係るニューラルネットワーク5は、複数種類のニューラルネットワークを組み合わせることで構成されている。
(Learning device)
Next, the learning device will be described. As shown in FIG. 9, the seating determination device 1 according to the present embodiment is a neural network as a learned learner that has performed learning for determining whether or not the driver 800 is seated in the driver seat 900. 5 is used. The neural network 5 according to the present embodiment is configured by combining a plurality of types of neural networks.
 具体的には、ニューラルネットワーク5は、全結合ニューラルネットワーク51、畳み込みニューラルネットワーク52、結合層53、及びLSTMネットワーク54の4つの部分に分かれている。全結合ニューラルネットワーク51及び畳み込みニューラルネットワーク52は入力側に並列に配置されており、全結合ニューラルネットワーク51には観測情報124が入力され、畳み込みニューラルネットワーク52には低解像度撮影画像1231が入力される。結合層53は、全結合ニューラルネットワーク51及び畳み込みニューラルネットワーク52の出力を結合する。LSTMネットワーク54は、結合層53からの出力を受けて、着座情報125を出力する。 Specifically, the neural network 5 is divided into four parts: a fully connected neural network 51, a convolutional neural network 52, a connected layer 53, and an LSTM network 54. The fully connected neural network 51 and the convolutional neural network 52 are arranged in parallel on the input side. Observation information 124 is input to the fully connected neural network 51, and a low-resolution captured image 1231 is input to the convolutional neural network 52. . The connection layer 53 combines the outputs of the fully connected neural network 51 and the convolutional neural network 52. The LSTM network 54 receives the output from the coupling layer 53 and outputs the seating information 125.
 (a)全結合ニューラルネットワーク
 全結合ニューラルネットワーク51は、いわゆる多層構造のニューラルネットワークであり、入力側から順に、入力層511、中間層(隠れ層)512、及び出力層513を備えている。ただし、全結合ニューラルネットワーク51の層の数は、このような例に限定されなくてもよく、実施の形態に応じて適宜選択されてよい。
(A) Fully Connected Neural Network The fully connected neural network 51 is a so-called multilayered neural network, and includes an input layer 511, an intermediate layer (hidden layer) 512, and an output layer 513 in order from the input side. However, the number of layers of the fully connected neural network 51 may not be limited to such an example, and may be appropriately selected according to the embodiment.
 各層511~513は、1又は複数のニューロン(ノード)を備えている。各層511~513に含まれるニューロンの個数は、実施の形態に応じて適宜設定されてよい。各層511~513に含まれる各ニューロンが、隣接する層に含まれる全てのニューロンに結合されていることで、全結合ニューラルネットワーク51は構成される。各結合には、重み(結合荷重)が適宜設定されている。 Each layer 511 to 513 includes one or a plurality of neurons (nodes). The number of neurons included in each of the layers 511 to 513 may be set as appropriate according to the embodiment. The all-connected neural network 51 is configured by connecting each neuron included in each layer 511 to 513 to all the neurons included in the adjacent layers. A weight (coupling load) is appropriately set for each coupling.
 (b)畳み込みニューラルネットワーク
 畳み込みニューラルネットワーク52は、畳み込み層521及びプーリング層522を交互に接続した構造を有する順伝播型ニューラルネットワークである。本実施形態に係る畳み込みニューラルネットワーク52では、複数の畳み込み層521及びプーリング層522が入力側に交互に配置されている。そして、最も出力側に配置されたプーリング層522の出力が全結合層523に入力され、全結合層523の出力が出力層524に入力される。
(B) Convolutional Neural Network The convolutional neural network 52 is a forward propagation neural network having a structure in which convolutional layers 521 and pooling layers 522 are alternately connected. In the convolutional neural network 52 according to the present embodiment, a plurality of convolutional layers 521 and pooling layers 522 are alternately arranged on the input side. Then, the output of the pooling layer 522 arranged on the most output side is input to the total coupling layer 523, and the output of the total coupling layer 523 is input to the output layer 524.
 畳み込み層521は、画像の畳み込みの演算を行う層である。画像の畳み込みとは、画像と所定のフィルタとの相関を算出する処理に相当する。そのため、画像の畳み込みを行うことで、例えば、フィルタの濃淡パターンと類似する濃淡パターンを入力される画像から検出することができる。 The convolution layer 521 is a layer that performs an image convolution operation. Image convolution corresponds to processing for calculating the correlation between an image and a predetermined filter. Therefore, by performing image convolution, for example, a shading pattern similar to the shading pattern of the filter can be detected from the input image.
 プーリング層522は、プーリング処理を行う層である。プーリング処理は、画像のフィルタに対する応答の強かった位置の情報を一部捨て、画像内に現れる特徴の微小な位置変化に対する応答の不変性を実現する。 The pooling layer 522 is a layer that performs a pooling process. The pooling process discards a part of the information of the position where the response to the image filter is strong, and realizes the invariance of the response to the minute position change of the feature appearing in the image.
 全結合層523は、隣接する層の間のニューロン全てを結合した層である。すなわち、全結合層523に含まれる各ニューロンは、隣接する層に含まれる全てのニューロンに結合される。畳み込みニューラルネットワーク52は、2層以上の全結合層523を備えてもよい。また、全結合層423に含まれるニューロンの個数は、実施の形態に応じて適宜設定されてよい。 The total connection layer 523 is a layer in which all neurons between adjacent layers are connected. That is, each neuron included in all connection layers 523 is connected to all neurons included in adjacent layers. The convolutional neural network 52 may include two or more fully connected layers 523. Further, the number of neurons included in all connection layers 423 may be set as appropriate according to the embodiment.
 出力層524は、畳み込みニューラルネットワーク52の最も出力側に配置される層である。出力層524に含まれるニューロンの個数は、実施の形態に応じて適宜設定されてよい。なお、畳み込みニューラルネットワーク52の構成は、このような例に限定されなくてもよく、実施の形態に応じて適宜設定されてよい。 The output layer 524 is a layer arranged on the most output side of the convolutional neural network 52. The number of neurons included in the output layer 524 may be appropriately set according to the embodiment. Note that the configuration of the convolutional neural network 52 is not limited to such an example, and may be appropriately set according to the embodiment.
 (c)結合層
 結合層53は、全結合ニューラルネットワーク51及び畳み込みニューラルネットワーク52とLSTMネットワーク54との間に配置される。結合層53は、全結合ニューラルネットワーク51の出力層513からの出力及び畳み込みニューラルネットワーク52の出力層524からの出力を結合する。結合層53に含まれるニューロンの個数は、全結合ニューラルネットワーク51及び畳み込みニューラルネットワーク52の出力の数に応じて適宜設定されてよい。
(C) Connection Layer The connection layer 53 is disposed between the fully connected neural network 51 and the convolutional neural network 52 and the LSTM network 54. The connection layer 53 combines the output from the output layer 513 of the fully connected neural network 51 and the output from the output layer 524 of the convolutional neural network 52. The number of neurons included in the connection layer 53 may be appropriately set according to the number of outputs of the fully connected neural network 51 and the convolutional neural network 52.
 (d)LSTMネットワーク
 LSTMネットワーク54は、LSTMブロック542を備える再起型ニューラルネットワークである。再帰型ニューラルネットワークは、例えば、中間層から入力層への経路のように、内部にループを有するニューラルネットワークのことである。LSTMネットワーク54は、一般的な再起型ニューラルネットワークの中間層をLSTMブロック542に置き換えた構造を有する。
(D) LSTM Network The LSTM network 54 is a recurrent neural network that includes an LSTM block 542. A recursive neural network is a neural network having a loop inside, such as a path from an intermediate layer to an input layer. The LSTM network 54 has a structure in which an intermediate layer of a general recurrent neural network is replaced with an LSTM block 542.
 本実施形態では、LSTMネットワーク54は、入力側から順に、入力層541、LSTMブロック542、及び出力層543を備えており、順伝播の経路の他、LSTMブロック542から入力層541に戻る経路を有している。入力層541及び出力層543に含まれるニューロンの個数は、実施の形態に応じて適宜設定されてよい。 In the present embodiment, the LSTM network 54 includes an input layer 541, an LSTM block 542, and an output layer 543 in order from the input side. In addition to a forward propagation path, a path returning from the LSTM block 542 to the input layer 541 is provided. Have. The number of neurons included in the input layer 541 and the output layer 543 may be set as appropriate according to the embodiment.
 LSTMブロック542は、入力ゲート及び出力ゲートを備え、情報の記憶及び出力のタイミングを学習可能に構成されたブロックである(S.Hochreiter and J.Schmidhuber, "Long short-term memory" Neural Computation, 9(8):1735-1780, November 15, 1997)。また、LSTMブロック542は、情報の忘却のタイミングを調節する忘却ゲートを備えてもよい(Felix A. Gers, Jurgen Schmidhuber and Fred Cummins, "Learning to Forget: Continual Prediction with LSTM" Neural Computation, pages 2451-2471, October 2000)。LSTMネットワーク54の構成は、実施の形態に応じて適宜設定可能である。 The LSTM block 542 includes an input gate and an output gate, and is configured to be able to learn information storage and output timing (S. Hochreiter and J.Schmidhuber, "Long short-term memory" Neural Computation, 9). (8): 1735-1780, November 15, 1997). The LSTM block 542 may also include a forgetting gate that adjusts the timing of forgetting information (FelixFA. Gers, Jurgen Schmidhuber and Fred Cummins, "Learning to Forget: Continual Prediction with LSTM" Neural Computation, pages 2451- 2471, “October” 2000). The configuration of the LSTM network 54 can be set as appropriate according to the embodiment.
 (e)小括
 各ニューロンには閾値が設定されており、基本的には、各入力と各重みとの積の和が閾値を超えているか否かによって各ニューロンの出力が決定される。着座判定装置1は、全結合ニューラルネットワーク51に観測情報124を入力し、畳み込みニューラルネットワーク52に低解像度撮影画像1231を入力する。そして、着座判定装置1は、入力側から順に、各層に含まれる各ニューロンの発火判定を行う。これにより、着座判定装置1は、着座情報125に対応する出力値をニューラルネットワーク5の出力層543から取得する。
(E) Summary A threshold is set for each neuron, and basically, the output of each neuron is determined by whether or not the sum of products of each input and each weight exceeds the threshold. The seating determination apparatus 1 inputs observation information 124 to the fully connected neural network 51 and inputs a low-resolution captured image 1231 to the convolutional neural network 52. And the seating determination apparatus 1 performs the firing determination of each neuron included in each layer in order from the input side. Thereby, the seating determination apparatus 1 acquires an output value corresponding to the seating information 125 from the output layer 543 of the neural network 5.
 <5.学習装置>
 次に、図10を用いて、本実施形態に係る学習装置2の機能構成の一例を説明する。図10は、本実施形態に係る学習装置2の機能構成の一例を模式的に例示する。
<5. Learning device>
Next, an example of a functional configuration of the learning device 2 according to the present embodiment will be described with reference to FIG. FIG. 10 schematically illustrates an example of a functional configuration of the learning device 2 according to the present embodiment.
 図10に示すように、学習装置2の制御部21は、記憶部22に記憶された学習プログラム221をRAMに展開する。そして、制御部21は、RAMに展開された学習プログラム221をCPUにより解釈及び実行して、各構成要素を制御する。これによって、図10に示されるとおり、本実施形態に係る学習装置2は、学習データ取得部211、及び学習処理部212を備えるコンピュータとして機能する。 As shown in FIG. 10, the control unit 21 of the learning device 2 develops the learning program 221 stored in the storage unit 22 in the RAM. Then, the control unit 21 interprets and executes the learning program 221 expanded in the RAM, and controls each component. Accordingly, as illustrated in FIG. 10, the learning device 2 according to the present embodiment functions as a computer including the learning data acquisition unit 211 and the learning processing unit 212.
 学習データ取得部211は、車両の運転席に着いた運転者を撮影するように配置された撮影装置から取得される撮影画像、当該運転者の顔の挙動に関する顔挙動情報を含む当該運転者の観測情報、及び当該運転者の着座に関する着座情報の組を学習データとして取得する。本実施形態では、学習データ取得部211は、低解像度撮影画像223、観測情報224、着座情報225の組を学習データ222として取得する。低解像度撮影画像223及び観測情報224はそれぞれ、上記低解像度撮影画像1231及び観測情報124に対応する。着座情報225は、上記着座情報125に対応する。学習処理部212は、低解像度撮影画像223及び観測情報224を入力すると着座情報225に対応する出力値を出力するように学習器を学習させる。これにより、この学習装置2では、上述した誤判定を避けるための、学習が行われる。 The learning data acquisition unit 211 includes a captured image acquired from an imaging device arranged to capture a driver who has arrived at the driver's seat of the vehicle, and facial behavior information regarding the behavior of the driver's face. A set of observation information and seating information related to the driver's seating is acquired as learning data. In the present embodiment, the learning data acquisition unit 211 acquires a set of the low-resolution captured image 223, the observation information 224, and the seating information 225 as learning data 222. The low-resolution captured image 223 and the observation information 224 correspond to the low-resolution captured image 1231 and the observation information 124, respectively. The seating information 225 corresponds to the seating information 125. When the learning processing unit 212 inputs the low-resolution captured image 223 and the observation information 224, the learning processing unit 212 learns the learning device so as to output an output value corresponding to the seating information 225. Thereby, in this learning apparatus 2, learning for avoiding the erroneous determination described above is performed.
 図10に示されるとおり、本実施形態において、学習対象となる学習器は、ニューラルネットワーク6である。当該ニューラルネットワーク6は、全結合ニューラルネットワーク61、畳み込みニューラルネットワーク62、結合層63、及びLSTMネットワーク64を備え、上記ニューラルネットワーク5と同様に構成される。全結合ニューラルネットワーク61、畳み込みニューラルネットワーク62、結合層63、及びLSTMネットワーク64はそれぞれ、上記全結合ニューラルネットワーク51、畳み込みニューラルネットワーク52、結合層53、及びLSTMネットワーク54と同様である。学習処理部212は、ニューラルネットワークの学習処理により、全結合ニューラルネットワーク61に観測情報224を入力し、畳み込みニューラルネットワーク62に低解像度撮影画像223を入力すると、着座情報225に対応する出力値をLSTMネットワーク64から出力するニューラルネットワーク6を構築する。そして、学習処理部212は、構築したニューラルネットワーク6の構成、各ニューロン間の結合の重み、及び各ニューロンの閾値を示す情報を学習結果データ122として記憶部22に格納する。 As shown in FIG. 10, in this embodiment, the learning device to be learned is a neural network 6. The neural network 6 includes a fully connected neural network 61, a convolutional neural network 62, a connected layer 63, and an LSTM network 64, and is configured in the same manner as the neural network 5. The fully connected neural network 61, the convolutional neural network 62, the connection layer 63, and the LSTM network 64 are the same as the above-described all connection neural network 51, the convolutional neural network 52, the connection layer 53, and the LSTM network 54, respectively. The learning processing unit 212 inputs the observation information 224 to the fully connected neural network 61 and inputs the low-resolution captured image 223 to the convolutional neural network 62 by the learning processing of the neural network, and outputs an output value corresponding to the seating information 225 to the LSTM. The neural network 6 that outputs from the network 64 is constructed. The learning processing unit 212 stores information indicating the configuration of the constructed neural network 6, the weight of the connection between the neurons, and the threshold value of each neuron as the learning result data 122 in the storage unit 22.
 <6.着座判定装置の動作>
 次に、図11を用いて、着座判定装置1の動作例を説明する。図11は、着座判定装置1の処理手順の一例を例示するフローチャートである。ただし、以下で説明する処理手順は一例に過ぎず、各処理は可能な限り変更されてよい。また、以下で説明する処理手順について、実施の形態に応じて、適宜、ステップの省略、置換、及び追加が可能である。
<6. Operation of Seating Determination Device>
Next, an operation example of the seating determination apparatus 1 will be described with reference to FIG. FIG. 11 is a flowchart illustrating an example of a processing procedure of the seating determination apparatus 1. However, the processing procedure described below is merely an example, and each processing may be changed as much as possible. Further, in the processing procedure described below, steps can be omitted, replaced, and added as appropriate according to the embodiment.
 まず、運転者800は、車両のイグニッション電源をオンにすることで、着座判定装置1を起動し、起動した着座判定装置1にプログラム121を実行させる。着座判定装置1が起動するタイミングは、これに限られない。例えば車両が手動運転モードと自動運転モードとを有する場合、着座判定装置1が起動するタイミングは、自動運転モードが起動されたタイミングでもよい。着座判定装置1の制御部11は、ナビゲーション装置30から地図情報、周辺情報、及びGPS情報を取得して、取得した地図情報、周辺情報、及びGPS情報に基づいて車両の自動運転を開始する。自動運転の制御方法には、公知の制御方法が利用可能である。そして、車両の自動運転を開始した後(ステップS201のYES)、制御部11は、以下の処理手順に従って、運転者800の状態を監視する。 First, the driver 800 activates the seating determination device 1 by turning on the ignition power of the vehicle, and causes the activated seating determination device 1 to execute the program 121. The timing which the seating determination apparatus 1 starts is not restricted to this. For example, when the vehicle has a manual operation mode and an automatic operation mode, the timing at which the seating determination device 1 is activated may be the timing at which the automatic operation mode is activated. The control unit 11 of the seating determination device 1 acquires map information, peripheral information, and GPS information from the navigation device 30 and starts automatic driving of the vehicle based on the acquired map information, peripheral information, and GPS information. As a control method for automatic operation, a known control method can be used. Then, after starting the automatic driving of the vehicle (YES in step S201), the control unit 11 monitors the state of the driver 800 according to the following processing procedure.
 (ステップS202)
 ステップS101では、制御部11は、画像取得部111として機能し、車両の運転席についた運転者800を撮影するように配置されたカメラ31から撮影画像123を取得する。取得する撮影画像123は、動画像であってもよいし、静止画像であってもよい。撮影画像123を取得すると、制御部11は、次のステップS203に処理を進める。
(Step S202)
In step S <b> 101, the control unit 11 functions as the image acquisition unit 111, and acquires the captured image 123 from the camera 31 arranged to capture the driver 800 attached to the driver's seat of the vehicle. The captured image 123 to be acquired may be a moving image or a still image. When the captured image 123 is acquired, the control unit 11 advances the processing to the next step S203.
 (ステップS203)
 ステップS203では、制御部11は、観測情報取得部112として機能し、運転者800の顔に挙動する顔挙動情報1241及び生体情報1242を含む観測情報124を取得する。観測情報124を取得すると、制御部11は、次のステップS204に処理を進める。
(Step S203)
In step S <b> 203, the control unit 11 functions as the observation information acquisition unit 112, and acquires observation information 124 including face behavior information 1241 and biological information 1242 that behave on the face of the driver 800. When the observation information 124 is acquired, the control unit 11 advances the processing to the next step S204.
 顔挙動情報1241は適宜取得されてよい。例えば、制御部11は、ステップS202で取得した撮影画像123に対して所定の画像解析を行うことで、運転者800の顔の検出可否、顔の位置、顔の向き、顔の動き、視線の方向、顔の器官の位置、及び目の開閉の少なくともいずれか1つに関する情報を顔挙動情報1241として取得することができる。 The face behavior information 1241 may be acquired as appropriate. For example, the control unit 11 performs predetermined image analysis on the captured image 123 acquired in step S202, thereby determining whether the driver 800 can detect the face, the position of the face, the direction of the face, the movement of the face, and the line of sight. Information regarding at least one of the direction, the position of the facial organ, and the opening and closing of the eyes can be acquired as the face behavior information 1241.
 顔挙動情報1241の取得方法の一例として、まず、制御部11は、撮影画像123から運転者800の顔を検出し、検出した顔の位置を特定する。これにより、制御部11は、顔の検出可否及び位置に関する情報を取得することができる。また、継続的に顔の検出を行うことで、制御部11は、顔の動きに関する情報を取得することができる。次に、制御部11は、検出した顔の画像内において、運転者800の顔に含まれる各器官(眼、口、鼻、耳等)を検出する。これにより、制御部11は、顔の器官の位置に関する情報を取得することができる。そして、制御部11は、検出した各器官(眼、口、鼻、耳等)の状態を解析することで、顔の向き、視線の方向、及び目の開閉に関する情報を取得することができる。顔の検出、器官の検出、及び器官の状態の解析には、公知の画像解析方法が用いられてよい。 As an example of a method for acquiring the face behavior information 1241, first, the control unit 11 detects the face of the driver 800 from the photographed image 123 and specifies the position of the detected face. Thereby, the control part 11 can acquire the information regarding the detectability and position of a face. Moreover, the control part 11 can acquire the information regarding a motion of a face by detecting a face continuously. Next, the control unit 11 detects each organ (eye, mouth, nose, ear, etc.) included in the face of the driver 800 in the detected face image. Thereby, the control part 11 can acquire the information regarding the position of the facial organ. And the control part 11 can acquire the information regarding the direction of a face, the direction of eyes | visual_axis, and opening / closing of an eye by analyzing the state of each detected organ (eye, mouth, nose, ear, etc.). A known image analysis method may be used for face detection, organ detection, and organ state analysis.
 取得される撮影画像123が、動画像又は時系列に並んだ複数の静止画像である場合、制御部11は、これらの画像解析を撮影画像123の各フレームに実行することで、時系列に沿って各種情報を取得することができる。これにより、制御部11は、ヒストグラム又は統計量(平均値、分散値等)で表された各種情報を時系列データで取得することができる。 When the captured image 123 to be acquired is a moving image or a plurality of still images arranged in time series, the control unit 11 performs these image analyzes on each frame of the captured image 123 so that the acquired images are arranged in time series. Various information can be acquired. Thereby, the control part 11 can acquire the various information represented by the histogram or the statistic (an average value, a variance value, etc.) with time series data.
 また、制御部11は、生体センサ32から生体情報(例えば、脳波、心拍数等)1242を取得する。例えば、生体情報1242は、ヒストグラム又は統計量(平均値、分散値等)で表されてよい。顔挙動情報1241と同様に、制御部11は、生体センサ32に継続的にアクセスすることで、生体情報1242を時系列データで取得することができる。なお、上述したように、生体情報1242は、必ずしも必要ではなく、制御部11は、顔挙動情報1241のみを用いて観測情報124を生成することができる。 Further, the control unit 11 acquires biological information (for example, brain waves, heart rate, etc.) 1242 from the biological sensor 32. For example, the biological information 1242 may be represented by a histogram or a statistic (average value, variance value, etc.). Similar to the face behavior information 1241, the control unit 11 can obtain the biological information 1242 as time-series data by continuously accessing the biological sensor 32. As described above, the biological information 1242 is not necessarily required, and the control unit 11 can generate the observation information 124 using only the face behavior information 1241.
 (ステップS204)
 ステップS204では、制御部11は、解像度変換部113として機能し、ステップS202で取得した撮影画像123の解像度を低下させる。これにより、制御部11は、低解像度撮影画像1231を形成する。低解像度化の処理方法は、特に限定されなくてもよく、実施の形態に応じて適宜選択されてよい。例えば、制御部11は、ニアレストネイバー法、バイリニア補間法、バイキュービック法等により、低解像度撮影画像1231を形成することができる。低解像度撮影画像1231を形成すると、制御部11は、次のステップS104に処理を進める。なお、本ステップS103は省略されてもよい。すなわち、制御部11は、撮影画像123の低解像度化を行うことなく、撮影画像123を学習器5への入力とすることができる。
(Step S204)
In step S204, the control unit 11 functions as the resolution conversion unit 113, and reduces the resolution of the captured image 123 acquired in step S202. Thereby, the control unit 11 forms a low-resolution captured image 1231. The processing method for reducing the resolution is not particularly limited, and may be appropriately selected according to the embodiment. For example, the control unit 11 can form the low-resolution captured image 1231 by the nearest neighbor method, the bilinear interpolation method, the bicubic method, or the like. When the low-resolution captured image 1231 is formed, the control unit 11 advances the processing to the next step S104. Note that this step S103 may be omitted. That is, the control unit 11 can input the captured image 123 to the learning device 5 without reducing the resolution of the captured image 123.
 (ステップS205及びS206)
 ステップS205では、制御部11は、運転状態推定部114として機能し、取得した観測情報124及び低解像度撮影画像1231をニューラルネットワーク5の入力として用いて、当該ニューラルネットワーク5の演算処理を実行する。これにより、ステップS105では、制御部11は、着座情報125それぞれに対応する出力値を当該ニューラルネットワーク5から得る。
(Steps S205 and S206)
In step S <b> 205, the control unit 11 functions as the driving state estimation unit 114, and executes arithmetic processing of the neural network 5 using the acquired observation information 124 and the low-resolution captured image 1231 as inputs of the neural network 5. Thereby, in step S <b> 105, the control unit 11 obtains an output value corresponding to each of the seating information 125 from the neural network 5.
 具体的には、制御部11は、ステップS203で取得した観測情報124を全結合ニューラルネットワーク51の入力層511に入力し、ステップS204で取得した低解像度撮影画像1231を、畳み込みニューラルネットワーク52の最も入力側に配置された畳み込み層521に入力する。そして、制御部11は、入力側から順に、各層に含まれる各ニューロンの発火判定を行う。これにより、制御部11は、着座情報125に対応する出力値をLSTMネットワーク54の出力層543から取得する。 Specifically, the control unit 11 inputs the observation information 124 acquired in step S203 to the input layer 511 of the fully connected neural network 51, and the low-resolution captured image 1231 acquired in step S204 is the most in the convolutional neural network 52. It inputs into the convolution layer 521 arrange | positioned at the input side. And the control part 11 performs the firing determination of each neuron contained in each layer in order from the input side. Thereby, the control unit 11 acquires an output value corresponding to the seating information 125 from the output layer 543 of the LSTM network 54.
 (ステップS207及びS208)
 ステップS207では、制御部11は、警告部115として機能し、ステップS206で取得した着座情報125に基づいて、運転者800が運転席に着座しているか否かを判定する。そして、運転者800が運転席に着座していると判定した場合には、制御部11は、次のステップS208を省略して、本動作例に係る処理を終了する。一方、運転者800が運転席に着座していないと判定した場合には、制御部11は、次のステップS208の処理を実行する。すなわち、制御部11は、スピーカ33やディスプレイを介して、運転者が運転席に着座していないことを報知する。あるいは、自動車を減速したり、停止させることもできる。その後、運転が行われていれば(ステップS201のYES)、引き続き、ステップS202~S207の処理を行う。一方、運転が行われていない場合には(ステップS201のNO)、処理を停止する。
(Steps S207 and S208)
In step S207, the control unit 11 functions as the warning unit 115, and determines whether or not the driver 800 is seated in the driver's seat based on the seating information 125 acquired in step S206. And when it determines with the driver | operator 800 sitting on the driver's seat, the control part 11 abbreviate | omits next step S208 and complete | finishes the process which concerns on this operation example. On the other hand, if it is determined that the driver 800 is not seated in the driver's seat, the control unit 11 performs the process of the next step S208. That is, the control unit 11 notifies that the driver is not seated in the driver's seat via the speaker 33 or the display. Alternatively, the automobile can be decelerated or stopped. Thereafter, if the operation is being performed (YES in step S201), the processes in steps S202 to S207 are subsequently performed. On the other hand, when the operation is not performed (NO in step S201), the process is stopped.
 <7.特徴>
 以上のように、本実施形態に係る着座判定装置1は、上記ステップS202~ステップS204までの処理により、運転者800の顔挙動情報1241を含む観測情報124と車両の運転席に着いた運転者を撮影するように配置されたカメラ3から得られる撮影画像(低解像度撮影画像1231)とを取得する。そして、着座判定装置1は、上記ステップS205及びS206により、取得した観測情報124及び低解像度撮影画像1231を学習済みのニューラルネットワーク(ニューラルネットワーク5)の入力として用いることで、運転者800が運転席900に着座しているか否か判定する。この学習済みのニューラルネットワークは、上記学習装置2により、低解像度撮影画像223、観測情報224、着座情報225を含む学習データを用いて作成される。したがって、本実施形態では、運転者の着座を判定する過程に、運転者800の顔の挙動だけではなく、低解像度撮影画像から判別され得る運転者800の身体の状態(例えば、身体の向き、姿勢等)を反映することができる。
<7. Features>
As described above, the seating determination apparatus 1 according to the present embodiment performs the processing from step S202 to step S204, the observation information 124 including the face behavior information 1241 of the driver 800 and the driver who has arrived at the driver's seat of the vehicle. And a captured image (low-resolution captured image 1231) obtained from the camera 3 arranged so as to capture the image. The seating determination device 1 uses the acquired observation information 124 and the low-resolution captured image 1231 as inputs of the learned neural network (the neural network 5) in steps S205 and S206, so that the driver 800 It is determined whether the user is seated at 900. The learned neural network is created by the learning device 2 using learning data including the low-resolution captured image 223, the observation information 224, and the seating information 225. Therefore, in the present embodiment, in the process of determining the driver's seating, not only the behavior of the driver 800's face but also the state of the driver's 800 body (for example, the body orientation, Attitude).
 よって、撮影画像123において、運転席900に写る者の身体の状態から、これが、運転者800であるか、助手席や後部座席の者が運転席900に身体を乗り出しているのであるか、を判定することができる。また、運転者800が着座している場合でも、運転者800がうつむいたり、後を向いたりする場合には、顔の器官を正確に検出できないが、このような場合でも運転者800の身体の状態を検出することで、運転者800が運転席900に着座していると判定することができる。したがって、運転者800が運転席900に着座しているか否かを精度よく判定することができる。 Therefore, in the captured image 123, it is determined from the state of the body of the person shown in the driver's seat 900 whether this is the driver 800 or whether the person in the front passenger seat or the rear seat is riding on the driver's seat 900. Can be determined. Even when the driver 800 is seated, the facial organ cannot be accurately detected when the driver 800 is depressed or turned backward. Even in such a case, the body of the driver 800 is not detected. By detecting the state, it can be determined that the driver 800 is seated in the driver's seat 900. Therefore, it can be accurately determined whether or not the driver 800 is seated on the driver's seat 900.
 また、本実施形態では、ニューラルネットワーク(5、6)の入力として、運転者の顔挙動情報を含む観測情報(124、224)を利用している。そのため、ニューラルネットワーク(5、6)に入力するための撮影画像は、運転者の顔の挙動を判別できるほど高解像度のものでなくてもよい。そこで、本実施形態では、ニューラルネットワーク(5、6)の入力として、カメラ31により得られる撮影画像を低解像度化した低解像度撮影画像(1231、223)を用いてもよい。これにより、ニューラルネットワーク(5、6)の演算処理の計算量を低減することができ、プロセッサの負荷を低減することができる。なお、低解像度撮影画像(1231、223)の解像度は、運転者の顔の挙動は判別できないが、運転者の姿勢に関する特徴を抽出可能な程度であるのが好ましい。なお、上述したように、撮影画像123の低解像度化を必ずしも行う必要はなく、例えば、演算処理の負荷を考慮しなければ、撮影画像123を学習器5の入力とすることができる。 In this embodiment, observation information (124, 224) including driver's face behavior information is used as an input to the neural network (5, 6). Therefore, the captured image to be input to the neural network (5, 6) does not have to have a high resolution so that the behavior of the driver's face can be determined. Therefore, in this embodiment, low-resolution captured images (1231, 223) obtained by reducing the resolution of the captured image obtained by the camera 31 may be used as the input of the neural network (5, 6). Thereby, the calculation amount of the arithmetic processing of the neural network (5, 6) can be reduced, and the load on the processor can be reduced. It should be noted that the resolution of the low-resolution captured image (1231, 223) is preferably such that the behavior of the driver's face cannot be discriminated, but the feature relating to the driver's posture can be extracted. Note that, as described above, it is not always necessary to reduce the resolution of the captured image 123. For example, the captured image 123 can be used as the input of the learning device 5 without considering the processing load.
 また、本実施形態に係るニューラルネットワーク5は、全結合ニューラルネットワーク51及び畳み込みニューラルネットワーク52を入力側に備えている。そして、上記ステップS104では、全結合ニューラルネットワーク51に観測情報124を入力し、畳み込みニューラルネットワーク52に低解像度撮影画像1231を入力している。これにより、各入力に適した解析を行うことができる。また、本実施形態に係るニューラルネットワーク5は、LSTMネットワーク54を備えている。これにより、観測情報124及び低解像度撮影画像1231に時系列データを利用し、短期的な依存関係だけでなく、長期的な依存関係を考慮して、運転者800の着座を判定することができる。したがって、本実施形態によれば、運転者800の着座の判定精度を高めることができる。 Further, the neural network 5 according to the present embodiment includes a fully connected neural network 51 and a convolutional neural network 52 on the input side. In step S <b> 104, the observation information 124 is input to the fully connected neural network 51, and the low-resolution captured image 1231 is input to the convolutional neural network 52. Thereby, analysis suitable for each input can be performed. The neural network 5 according to this embodiment includes an LSTM network 54. Thereby, the time series data is used for the observation information 124 and the low-resolution captured image 1231, and the seating of the driver 800 can be determined in consideration of not only short-term dependency but also long-term dependency. . Therefore, according to the present embodiment, the seating determination accuracy of the driver 800 can be increased.
 <C.変形例>
 以上、本発明の実施の形態を詳細に説明してきたが、前述までの説明はあらゆる点において本発明の例示に過ぎない。本発明の範囲を逸脱することなく種々の改良や変形を行うことができることは言うまでもない。例えば、以下のような変更が可能である。なお、以下では、上記実施形態と同様の構成要素に関しては同様の符号を用い、上記実施形態と同様の点については、適宜説明を省略した。以下の変形例は適宜組み合わせ可能である。
<C. Modification>
As mentioned above, although embodiment of this invention has been described in detail, the above description is only illustration of this invention in all the points. It goes without saying that various improvements and modifications can be made without departing from the scope of the present invention. For example, the following changes are possible. In the following, the same reference numerals are used for the same components as in the above embodiment, and the description of the same points as in the above embodiment is omitted as appropriate. The following modifications can be combined as appropriate.
 <1>
 例えば、上記実施形態では、図5及び図6に示されるとおり、各ニューラルネットワーク(7、8)として、多層構造を有する一般的な順伝播型ニューラルネットワークを用いている。しかしながら、各ニューラルネットワーク(7、8)の種類は、このような例に限定されなくてもよく、実施の形態に応じて適宜選択されてよい。例えば、各ニューラルネットワーク(7、8)は、入力層71及び中間層72を畳み込み層及びプーリング層として利用する畳み込みニューラルネットワークであってもよい。また、例えば、各ニューラルネットワーク(7、8)は、中間層72から入力層71等のように出力側から入力側に再帰する結合を有する再帰型ニューラルネットワークであってもよい。なお、各ニューラルネットワーク(7、8)の層数、各層におけるニューロンの個数、ニューロン同士の結合関係、及び各ニューロンの伝達関数は、実施の形態に応じて適宜決定されてよい。
<1>
For example, in the above embodiment, as shown in FIGS. 5 and 6, a general forward propagation type neural network having a multilayer structure is used as each neural network (7, 8). However, the type of each neural network (7, 8) may not be limited to such an example, and may be appropriately selected according to the embodiment. For example, each neural network (7, 8) may be a convolutional neural network that uses the input layer 71 and the intermediate layer 72 as a convolution layer and a pooling layer. Further, for example, each neural network (7, 8) may be a recursive neural network having a connection that recurs from the output side to the input side, such as the intermediate layer 72 to the input layer 71. The number of layers in each neural network (7, 8), the number of neurons in each layer, the connection relationship between neurons, and the transfer function of each neuron may be determined as appropriate according to the embodiment.
 <2>
 上記実施形態では、着座判定装置1と学習器(ニューラルネットワーク)7の学習を行う学習装置2とは別々のコンピュータで構成されている。しかしながら、着座判定装置1及び学習装置2の構成はこのような例に限定されなくてもよく、着座判定装置1及び学習装置2の両方の機能を有するシステムを1台又は複数台のコンピュータで実現してもよい。また、学習装置2を着座判定装置1に組み込んで使用することもできる。
<2>
In the above embodiment, the seating determination device 1 and the learning device 2 that learns the learning device (neural network) 7 are configured by separate computers. However, the configuration of the seating determination device 1 and the learning device 2 may not be limited to such an example, and a system having both functions of the seating determination device 1 and the learning device 2 is realized by one or a plurality of computers. May be. The learning device 2 can also be used by being incorporated in the seating determination device 1.
 <3>
 上記実施形態では、学習器は、ニューラルネットワークにより構成されている。しかしながら、学習器の種類は、カメラ3で撮影された撮影画像123を入力として利用可能であれば、ニューラルネットワークに限られなくてもよく、実施の形態に応じて適宜選択されてよい。複数の撮影画像123を入力可能な学習器として、例えば、上記ニューラルネットワークの他、サポートベクターマシン、自己組織化マップ、又は強化学習により学習を行う学習器によって構成された学習器を挙げることができる。
<3>
In the above embodiment, the learning device is configured by a neural network. However, the type of learning device is not limited to the neural network as long as the captured image 123 captured by the camera 3 can be used as an input, and may be appropriately selected according to the embodiment. As a learning device capable of inputting a plurality of captured images 123, for example, a learning device configured by a learning device that learns by support vector machine, self-organizing map, or reinforcement learning in addition to the neural network can be cited. .
 <4>
 上記実施形態では、着座判定装置1を単独の装置として、自動車に実装しているが、例えば、自動車のコンピュータに、上記着座判定プログラムをインストールして、着座判定を行うこともできる。
<4>
In the above-described embodiment, the seating determination device 1 is mounted on a vehicle as a single device. However, for example, the seating determination program can be installed in a computer of the vehicle to perform seating determination.
 <5>
 顔の器官を検出する方法は、上記のような学習器以外でも可能である。また、顔の器官を検出する以外に、顔自体の検出や、人の体の検出を行うこともできる。そのような方法としては、種々の方法があり、例えば、公知のパターンマッチングを用いることができる。あるいは三次元モデルを用いて特徴点の抽出を行う手法があり、具体的には、例えば、国際公開2006/051607号公報、特開2007-249280号公報などに記載されている手法を採用することができる。
<5>
The method for detecting the facial organ can be used other than the learning device as described above. In addition to detecting a facial organ, it is also possible to detect the face itself or a human body. As such a method, there are various methods, for example, known pattern matching can be used. Alternatively, there is a method of extracting feature points using a three-dimensional model. Specifically, for example, a method described in International Publication No. 2006/051607, Japanese Patent Application Laid-Open No. 2007-249280, or the like is adopted. Can do.
 <6>
 また、視体積交差法により運転席に位置する物体の三次元形状を取得し、この三次元形状が人であるか否かを判断することで、着座を判定することができる。この場合、車内に複数のカメラを設け、これら複数のカメラで複数の角度から運転席を撮影し、複数の撮影画像を取得する。そして、着座判定装置1では、解析部において、複数の撮影画像から、視体積交差法により運転席に着座する物体の三次元形状を取得する。そして、この三次元形状が人であるか否かを判断し、人である場合には、運転席に運転者が着座していると判定することができる。一方、三次元形状を取得できない場合、あるいは三次元形状が人ではないと判定されたた場合には、上記のように警告を行うことができる。
<6>
In addition, seating can be determined by acquiring a three-dimensional shape of an object located in the driver's seat by the visual volume intersection method and determining whether the three-dimensional shape is a person. In this case, a plurality of cameras are provided in the vehicle, and the driver's seat is photographed from a plurality of angles with the plurality of cameras, and a plurality of photographed images are acquired. And in the seating determination apparatus 1, in an analysis part, the three-dimensional shape of the object seated in a driver's seat is acquired from a some picked-up image by a visual volume intersection method. Then, it is determined whether or not the three-dimensional shape is a person. If the person is a person, it can be determined that the driver is seated in the driver's seat. On the other hand, when the three-dimensional shape cannot be obtained, or when it is determined that the three-dimensional shape is not a person, a warning can be given as described above.
 また、上記実施形態では、観測情報124は、顔挙動情報1241の他に、生体情報1242を含んでいる。しかしながら、観測情報124の構成は、このような例に限定されなくてもよく、実施の形態に応じて適宜選択されてよい。例えば、生体情報1242は省略されてもよい。また、例えば、観測情報124は、生体情報1242以外の情報を含んでいてもよい。 In the above embodiment, the observation information 124 includes biological information 1242 in addition to the face behavior information 1241. However, the configuration of the observation information 124 is not limited to such an example, and may be appropriately selected according to the embodiment. For example, the biological information 1242 may be omitted. For example, the observation information 124 may include information other than the biological information 1242.
 (付記1)
 自動車の運転席を撮影する少なくとも1つのカメラと接続される運転者の着座判定装置であって、
 少なくとも1つのハードウェアプロセッサを備え、
 前記ハードウェアプロセッサは、
 前記カメラによって撮影された撮影画像を取得し、
 前記撮影画像から前記運転席に運転者が着座しているか否かを判断する、運転者の着座判定装置。
(Appendix 1)
A driver's seating determination device connected to at least one camera for photographing a driver's seat of a car,
Comprising at least one hardware processor;
The hardware processor is
Obtaining a captured image captured by the camera,
A driver seating determination device that determines whether a driver is seated in the driver seat from the captured image.
 (付記2)
 自動車の運転席を、少なくとも1つのカメラで撮影するステップと、
 少なくとも1つのハードウェアプロセッサにより、前記カメラによって撮影された撮影画像から前記運転席に運転者が着座しているか否かを判断するステップと、
を備えている、運転者の着座判定方法。
(Appendix 2)
Photographing the driver's seat of the car with at least one camera;
Determining, by at least one hardware processor, whether or not a driver is seated in the driver's seat from a captured image captured by the camera;
A driver's seating determination method.
 1…着座判定装置、100…着座判定システム、
 11…制御部、12…記憶部、13…通信インタフェース、
 14…入力装置、15…出力装置、16…外部インタフェース、
 17…ドライブ、
 111…画像取得部、112…解析部、113…警告部、
 121…着座判定プログラム、122…学習結果データ、
 123…撮影画像、
 2…学習装置、
 21…制御部、22…記憶部、23…通信インタフェース、
 24…入力装置、25…出力装置、26…外部インタフェース、
 27…ドライブ、
 211…学習画像取得部、212…学習処理部、
 221…学習プログラム、222…学習データ、
 3…カメラ、
 7…ニューラルネットワーク、
 71…入力層、72…中間層(隠れ層)、73…出力層、
 8…ニューラルネットワーク、
 81…入力層、82…中間層(隠れ層)、83…出力層、
 91・92…記憶媒体
DESCRIPTION OF SYMBOLS 1 ... Seating determination apparatus, 100 ... Seating determination system,
11 ... Control unit, 12 ... Storage unit, 13 ... Communication interface,
14 ... input device, 15 ... output device, 16 ... external interface,
17 ... Drive,
111 ... Image acquisition unit, 112 ... Analysis unit, 113 ... Warning unit,
121 ... Seating determination program, 122 ... Learning result data,
123 ... Photos taken,
2 ... Learning device,
21 ... Control unit, 22 ... Storage unit, 23 ... Communication interface,
24 ... Input device, 25 ... Output device, 26 ... External interface,
27 ... Drive,
211 ... a learning image acquisition unit, 212 ... a learning processing unit,
221 ... Learning program, 222 ... Learning data,
3 ... Camera,
7 ... Neural network,
71 ... input layer, 72 ... intermediate layer (hidden layer), 73 ... output layer,
8 ... Neural network,
81 ... input layer, 82 ... intermediate layer (hidden layer), 83 ... output layer,
91.92 ... Storage medium

Claims (15)

  1.  自動車の運転席を撮影する少なくとも1つのカメラと接続される運転者の着座判定装置であって、
     前記カメラによって撮影された撮影画像を取得する画像取得部と、
     前記撮影画像から前記運転席に運転者が着座しているか否かを判断する解析部と、
    を備えている、運転者の着座判定装置。
    A driver's seating determination device connected to at least one camera for photographing a driver's seat of a car,
    An image acquisition unit for acquiring a captured image captured by the camera;
    An analysis unit for determining whether a driver is seated in the driver's seat from the captured image;
    A driver's seating determination device.
  2.  前記運転者の顔の挙動に関する顔挙動情報を含む当該運転者の観測情報を取得する観測情報取得部をさらに備え、
     前記解析部は、
     前記運転者の前記運転席への着座を判定するための学習を行った学習済みの学習器に、
    前記撮影画像及び前記観測情報を入力することで、前記運転者が着座しているか否かの着座情報を当該学習器から取得する運転者状態推定部と、
    を備える、請求項1に記載の運転者の着座判定装置。
    An observation information acquisition unit that acquires observation information of the driver including facial behavior information relating to the behavior of the driver's face;
    The analysis unit
    In a learned device that has been learned to determine whether the driver is seated in the driver's seat,
    A driver state estimation unit that acquires seating information as to whether or not the driver is seated by inputting the captured image and the observation information; and
    The driver's seating determination device according to claim 1, comprising:
  3.  前記観測情報取得部は、取得した前記撮影画像に対して所定の画像解析を行うことで、前記運転者の顔の検出可否、顔の位置、顔の向き、顔の動き、視線の方向、顔の器官の位置、及び目の開閉の少なくともいずれか1つに関する情報を前記顔挙動情報として取得する、請求項2に記載の運転者の着座判定装置。 The observation information acquisition unit performs predetermined image analysis on the acquired captured image, thereby detecting whether or not the driver's face can be detected, face position, face direction, face motion, line-of-sight direction, face The driver's seating determination apparatus according to claim 2, wherein information relating to at least one of a position of an organ of an eye and opening / closing of an eye is acquired as the face behavior information.
  4.  前記解析部は、取得した前記撮影画像の解像度を低下させる解像度変換部を更に備え、
     前記運転者状態推定部は、解像度を低下させた前記撮影画像を前記学習器に入力する、
    請求項2または3に記載の運転者の着座判定装置。
    The analysis unit further includes a resolution conversion unit that reduces the resolution of the acquired captured image,
    The driver state estimation unit inputs the captured image with reduced resolution to the learning device.
    The driver's seating determination device according to claim 2 or 3.
  5.  前記解析部は、前記撮影画像から前記運転者の顔を検出することで、前記運転席に運転者が着座していると判断する、請求項1に記載の運転者の着座判定装置。 The driver seating determination device according to claim 1, wherein the analysis unit determines that the driver is seated in the driver seat by detecting the driver's face from the captured image.
  6.  前記解析部は、前記撮影画像に含まれる像から人の顔の器官を検出することで、前記運転席に運転者が着座していると判断する、請求項5に記載の運転者の着座判定装置。 6. The driver seating determination according to claim 5, wherein the analysis unit determines that a driver is seated in the driver seat by detecting a human facial organ from an image included in the captured image. apparatus.
  7.  前記解析部は、
     人の顔を検出するための学習を行った学習済みの学習器であって、運転席を含む撮影画像を入力とし、当該撮影画像に人の顔が含まれるか否かを出力とする学習器を備えている、請求項5または6に記載の運転者の着座判定装置。
    The analysis unit
    A learning device that has been learned to detect a human face, and that takes a captured image including a driver's seat as an input and outputs whether or not the captured image includes a human face The driver's seating determination device according to claim 5 or 6, further comprising:
  8.  前記運転席に運転者が着座していないと判断した場合に、警告を発する警告部をさらに備えている、請求項1から7のいずれかに記載に運転者の着座判定装置。 The driver seating determination device according to any one of claims 1 to 7, further comprising a warning unit that issues a warning when it is determined that the driver is not seated in the driver seat.
  9.  前記自動車は、自動運転機能を有し、
     前記解析部は、前記自動運転機能の作動中に、運転者の着座を判断する、請求項1から8のいずれかに記載に運転者の着座判定装置。
    The automobile has an automatic driving function,
    The driver's seating determination device according to any one of claims 1 to 8, wherein the analysis unit determines the driver's seating during the operation of the automatic driving function.
  10.  自動車の運転席を、少なくとも1つのカメラで撮影するステップと、
     前記カメラによって撮影された撮影画像から前記運転席に運転者が着座しているか否かを判断するステップと、
    を備えている、運転者の着座判定方法。
    Photographing the driver's seat of the car with at least one camera;
    Determining whether a driver is seated in the driver's seat from a photographed image photographed by the camera;
    A driver's seating determination method.
  11.  前記撮影画像から前記運転者の顔を検出することで、前記運転席に運転者が着座していると判断する、請求項10に記載の運転者の着座判定方法。 The driver's seating determination method according to claim 10, wherein it is determined that the driver is seated in the driver's seat by detecting the driver's face from the photographed image.
  12.  前記撮影画像に含まれる像から人の顔の器官を検出することで、前記運転席に運転者が着座していると判断する、請求項10に記載の運転者の着座判定方法。 11. The driver seating determination method according to claim 10, wherein it is determined that a driver is seated in the driver seat by detecting a human facial organ from an image included in the photographed image.
  13.  自動車のコンピュータに、
     自動車の運転席を、少なくとも1つのカメラで撮影するステップと、
     前記カメラによって撮影された撮影画像から前記運転席に運転者が着座しているか否かを判断するステップと、
    を実行させる、運転者の着座判定プログラム。
    To the car computer,
    Photographing the driver's seat of the car with at least one camera;
    Determining whether a driver is seated in the driver's seat from a photographed image photographed by the camera;
    A driver's seating determination program that executes
  14.  前記撮影画像から前記運転者の顔を検出することで、前記運転席に運転者が着座していると判断する、請求項13に記載の運転者の着座判定プログラム。 14. The driver's seating determination program according to claim 13, wherein it is determined that the driver is seated in the driver's seat by detecting the driver's face from the photographed image.
  15.  前記撮影画像に含まれる像から人の顔の器官を検出することで、前記運転席に運転者が着座していると判断する、請求項14に記載の運転者の着座判定プログラム。 The driver's seating determination program according to claim 14, wherein it is determined that a driver is seated in the driver's seat by detecting an organ of a human face from an image included in the captured image.
PCT/JP2017/036276 2017-03-14 2017-10-05 Driver seating determination device WO2018168038A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2017049009A JP2018149990A (en) 2017-03-14 2017-03-14 Device for determining seating of driver
JP2017-049009 2017-03-14

Publications (1)

Publication Number Publication Date
WO2018168038A1 true WO2018168038A1 (en) 2018-09-20

Family

ID=63523251

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/036276 WO2018168038A1 (en) 2017-03-14 2017-10-05 Driver seating determination device

Country Status (2)

Country Link
JP (1) JP2018149990A (en)
WO (1) WO2018168038A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10635917B1 (en) * 2019-01-30 2020-04-28 StradVision, Inc. Method and device for detecting vehicle occupancy using passenger's keypoint detected through image analysis for humans' status recognition
WO2021001942A1 (en) * 2019-07-02 2021-01-07 三菱電機株式会社 Driver monitoring device and driver monitoring method
JP7145830B2 (en) * 2019-09-12 2022-10-03 Kddi株式会社 Object identification method, device and program using encoded parameter feature quantity

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002264826A (en) * 2001-03-09 2002-09-18 Toyota Motor Corp Automatic driving device for moving body
JP2013252863A (en) * 2013-09-27 2013-12-19 Takata Corp Occupant restriction control device and occupant restriction control method
WO2016092773A1 (en) * 2014-12-09 2016-06-16 株式会社デンソー Autonomous driving control device, driving information output device, footrest, autonomous driving control method, and driving information output method
WO2016190253A1 (en) * 2015-05-22 2016-12-01 株式会社デンソー Steering control device and steering control method

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7508979B2 (en) * 2003-11-21 2009-03-24 Siemens Corporate Research, Inc. System and method for detecting an occupant and head pose using stereo detectors
JP4438753B2 (en) * 2006-01-27 2010-03-24 株式会社日立製作所 In-vehicle state detection system, in-vehicle state detection device and method
JP6381353B2 (en) * 2014-08-08 2018-08-29 キヤノン株式会社 Image processing apparatus, imaging apparatus, image processing method, and program
JP6573193B2 (en) * 2015-07-03 2019-09-11 パナソニックIpマネジメント株式会社 Determination device, determination method, and determination program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002264826A (en) * 2001-03-09 2002-09-18 Toyota Motor Corp Automatic driving device for moving body
JP2013252863A (en) * 2013-09-27 2013-12-19 Takata Corp Occupant restriction control device and occupant restriction control method
WO2016092773A1 (en) * 2014-12-09 2016-06-16 株式会社デンソー Autonomous driving control device, driving information output device, footrest, autonomous driving control method, and driving information output method
WO2016190253A1 (en) * 2015-05-22 2016-12-01 株式会社デンソー Steering control device and steering control method

Also Published As

Publication number Publication date
JP2018149990A (en) 2018-09-27

Similar Documents

Publication Publication Date Title
JP6264492B1 (en) Driver monitoring device, driver monitoring method, learning device, and learning method
CN111602137B (en) Evaluation device, action control device, evaluation method, and storage medium
JP6245398B2 (en) State estimation device, state estimation method, and state estimation program
JP7032387B2 (en) Vehicle behavior estimation system and method based on monocular video data
JP7011578B2 (en) Methods and systems for monitoring driving behavior
WO2019149061A1 (en) Gesture-and gaze-based visual data acquisition system
CN112673378B (en) Device for generating estimator, monitoring device, method for generating estimator, and program for generating estimator
EP3588372B1 (en) Controlling an autonomous vehicle based on passenger behavior
WO2018168038A1 (en) Driver seating determination device
US20240096116A1 (en) Devices and methods for detecting drowsiness of drivers of vehicles
KR20230137221A (en) Method for verification of image, diagnostic system performing the same and computer-readable recording medium on which the method of performing the same
CN115719486A (en) Context-based state estimation
WO2017209225A1 (en) State estimation apparatus, state estimation method, and state estimation program
JP6996455B2 (en) Detector generator, monitoring device, detector generator and detector generator
JP6737212B2 (en) Driver state estimating device and driver state estimating method
JP7219787B2 (en) Information processing device, information processing method, learning method, and program
JP2018173757A (en) Detector, learning device, detection method, learning method, and program
JP2022135676A (en) Moving body control system and moving body control method
EP4332886A1 (en) Electronic device, method for controlling electronic device, and program
EP4332885A1 (en) Electronic device, control method for electronic device, and program
JP7466433B2 (en) Driving data recording device, driving assistance system, driving data recording method, and program
JP7035912B2 (en) Detector generator, monitoring device, detector generator method and detector generator
JP2024514994A (en) Image verification method, diagnostic system for executing the method, and computer-readable recording medium having the method recorded thereon
JP2023084580A (en) Hand area detection device, hand area detection method and computer program for hand area detection
KR20240084221A (en) Apparatus and method for recognizing storke symptoms of driver

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17901052

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17901052

Country of ref document: EP

Kind code of ref document: A1