US20210009150A1 - Method for recognizing dangerous action of personnel in vehicle, electronic device and storage medium - Google Patents

Method for recognizing dangerous action of personnel in vehicle, electronic device and storage medium Download PDF

Info

Publication number
US20210009150A1
US20210009150A1 US17/034,290 US202017034290A US2021009150A1 US 20210009150 A1 US20210009150 A1 US 20210009150A1 US 202017034290 A US202017034290 A US 202017034290A US 2021009150 A1 US2021009150 A1 US 2021009150A1
Authority
US
United States
Prior art keywords
action
vehicle
personnel
video stream
danger level
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US17/034,290
Inventor
Yanjie Chen
Fei Wang
Chen Qian
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Sensetime Technology Development Co Ltd
Original Assignee
Beijing Sensetime Technology Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Sensetime Technology Development Co Ltd filed Critical Beijing Sensetime Technology Development Co Ltd
Assigned to BEIJING SENSETIME TECHNOLOGY DEVELOPMENT CO., LTD. reassignment BEIJING SENSETIME TECHNOLOGY DEVELOPMENT CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHEN, YANJIE, QIAN, Chen, WANG, FEI
Publication of US20210009150A1 publication Critical patent/US20210009150A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/087Interaction between the driver and the control system where the control system corrects or modifies a request from the driver
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • G06K9/00744
    • G06K9/00845
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W2050/143Alarm means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2540/00Input parameters relating to occupants
    • B60W2540/225Direction of gaze
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2540/00Input parameters relating to occupants
    • B60W2540/229Attention level, e.g. attentive to driving, reading or sleeping

Definitions

  • the present disclosure relates to computer vision technologies, and in particular, to a method and device for recognizing a dangerous action of personnel in a vehicle, an electronic device, and a storage medium.
  • driver monitoring may be generally summarized as modules such as in-vehicle personnel face recognition module and fatigue detection module.
  • modules such as in-vehicle personnel face recognition module and fatigue detection module.
  • a method for recognizing a dangerous action including:
  • each video stream including information about at least one of the personnel in the vehicle;
  • a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
  • a device for recognizing a dangerous action of personnel in a vehicle including:
  • a video collection unit used for obtaining at least one video stream of the personnel in the vehicle through an image capturing device, each video stream including information about at least one of the personnel in the vehicle;
  • an action recognition unit used for performing action recognition on the personnel in the vehicle based on the video stream
  • a danger processing unit used for, responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
  • an electronic device in a third aspect, includes a processor, where the processor includes the device for recognizing the dangerous action of the personnel in the vehicle according to the first aspect.
  • an electronic device in a fourth aspect, includes: a processor; and a memory configured to store instructions that, when executed by the processor, cause the processor to perform the following operations including:
  • each video stream including information about at least one of the personnel in the vehicle;
  • a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior
  • a non-transitory computer readable storage medium is provided and is used for storing computer readable instructions that, when executed by a processor of an electronic device, cause the processor to perform a method for recognizing a dangerous action is provided, including:
  • each video stream including information about at least one of the personnel in the vehicle;
  • a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
  • a computer program product includes computer readable codes that, when being run on a device, cause a processor in the device to execute instructions for implementing the method for recognizing the dangerous action of the personnel in the vehicle according to the first aspect.
  • FIG. 1 is a schematic flowchart of a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 2 is a part of a schematic flowchart in an optional example of a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 3 a is a part of a schematic flowchart in another optional example of a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 3 b is a schematic diagram of an extracted target area in a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 4 is a schematic structural diagram of a device for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 5 is a schematic structural diagram of an electronic device, which may be a terminal device or a server, suitable for implementing the embodiments of the present disclosure.
  • the embodiments of the present disclosure may be applied to a computer system/server, which may operate with numerous other general-purpose or special-purpose computing system environments or configurations.
  • Examples of well-known computing systems, environments, and/or configurations suitable for use together with the computer system/server include, but are not limited to, personal computer systems, server computer systems, thin clients, thick clients, handheld or laptop devices, microprocessor-based systems, set top boxes, programmable consumer electronics, network personal computers, small computer systems, large computer systems, distributed cloud computing environments that include any one of the foregoing systems, and the like.
  • the computer system/server may be described in the general context of computer system executable instructions (for example, program modules) executed by the computer system.
  • the program modules may include routines, programs, target programs, components, logics, data structures, and the like for performing specific tasks or implementing specific abstract data types.
  • the computer system/server may be practiced in the distributed cloud computing environments in which tasks are performed by remote processing devices that are linked through a communications network.
  • the program modules may be located in local or remote computing system storage media including storage devices.
  • a dangerous action recognition system may give a prompt when a driver makes a dangerous action, thereby performing early-warning and avoiding possible accidents; furthermore, the system may monitor some behaviors that are substandard or may cause discomfort of an in-vehicle passenger to give a prompt and restrain the behaviors; meanwhile, the monitoring of the dangerous action reflects some habits and hobbies of a driver, which facilitates the system to establish a user portrait and perform big data analysis; and an emotional state, a fatigue state, and a behavior habit of the driver is monitored by means of dangerous action recognition.
  • FIG. 1 is a schematic flowchart of a method for recognizing a dangerous action of personnel in a vehicle (i.e., in-vehicle personnel) provided in the embodiments of the present disclosure.
  • the method is performed by any electronic device, such as a terminal device, a server, a mobile device, and a vehicle-mounted device.
  • the method of this embodiment includes the following steps.
  • At step 110 at least one video stream of in-vehicle personnel is obtained through an image capturing device (for example, a camera or the like).
  • an image capturing device for example, a camera or the like.
  • Each video stream includes information about at least one of the personnel in the vehicle; in the embodiments of the present disclosure, an image of the in-vehicle personnel is collected by means of a photographing apparatus (also called the image capturing device, for example, one or more cameras provided in the vehicle and used for performing photographing on a vehicle seat position), so as to obtain the video stream; optionally, the video streams of multiple in-vehicle personnel (for example, all of in-vehicle personnel) in the whole vehicle are collected on the basis of one photographing apparatus, or one photographing apparatus facing one or more back row areas for image collection is provided in the vehicle, or one photographing apparatus is respectively provided in the front of each seat, so as to respectively collect the video stream of at least one in-vehicle personnel (for example, each in-vehicle personnel); and action recognition is respectively performed on the in-vehicle personnel by performing processing on the collected video stream.
  • a photographing apparatus also called the image capturing device, for example, one or more cameras provided in the vehicle and used for performing photograph
  • a situation that video stream collection is performed on the in-vehicle personnel on the basis of an out-vehicle photographing apparatus for example, a camera provided on a road
  • an out-vehicle photographing apparatus for example, a camera provided on a road
  • the photographing apparatus includes, but is not limited to, at least one of the followings: a visible light camera, an infrared camera, or a near-infrared camera.
  • the visible light camera is used for collecting a visible light image
  • the infrared camera is used for collecting an infrared image
  • the near-infrared camera is used for collecting a near-infrared image.
  • step 110 is executed by a processor by invoking a corresponding instruction stored in a memory, or is executed by a video collection unit 41 run by the processor.
  • step 120 action recognition is performed on the in-vehicle personnel on the basis of the video stream.
  • an action category is divided into a dangerous action and a normal action, and the dangerous action needs to be processed to exclude a possible danger
  • the dangerous action includes, but is not limited to, at least one of the followings: a distraction action, a discomfort state, or a non-standard behavior or the like.
  • the dangerous action may have same or different requirements for an ordinary non-driver and driver, for example, the requirements for the driver are relatively stricter, and meanwhile, the independence and safety of the driver need to be protected, for example, the predetermined dangerous action is divided into the dangerous action of the driver and the dangerous action of the non-driver.
  • the embodiments of the present disclosure do not limit a specific mode for recognizing the action category.
  • step 120 is performed by the processor by invoking the corresponding instruction stored in the memory, or is performed by an action recognition unit 42 run by the processor.
  • prompt information is sent and/or an operation is executed to control the vehicle in response to that an action recognition result (i.e., a result of the action recognition) belongs to a predetermined dangerous action.
  • an action recognition result i.e., a result of the action recognition
  • the dangerous action in the embodiments of the present disclosure is a behavior that causes potential safety hazards to the in-vehicle personnel or others.
  • the predetermined dangerous action in the embodiments of the present disclosure includes, but is not limited to, at least one of the following action representations of the in-vehicle personnel: a distraction action, a discomfort state, or a non-standard behavior and the like.
  • the distraction action mainly aims at the driver; in a process of driving a vehicle, the driver needs to concentrate; when the distraction action (for example, actions such as eating food and smoking) occurs, the attention of the driver is influenced, and the vehicle is prone to be in danger; the discomfort state may aim at all of the in-vehicle personnel; when the discomfort state occurs in the in-vehicle personnel, on the basis of human safety considerations, some dangerous situations need to be processed in time, for example, situations such as frequent yawning of the driver or sweat wiping of a passenger; and the non-standard behavior is a behavior that do not comply with safe driving regulations, and may further be a behavior that is dangerous to the driver or other people in the vehicle, etc.
  • the distraction action for example, actions such as eating food and smoking
  • the probability of danger is reduced by sending the prompt information or executing the operation to control the vehicle, and the safety and/or comfort level of the in-vehicle personnel are improved.
  • a representation form of the prompt information may include, but is not limited to, at least one of the followings: voice prompt information, vibration prompt information, light prompt information, or smell prompt information and the like; for example, when the in-vehicle personnel smokes, the voice prompt information is sent to prompt that smoking in the vehicle is not allowed, so as to reduce the danger of smoking to other in-vehicle personnel; for another example: when the in-vehicle personnel wipes sweat, it means that a temperature in the vehicle is too high, and an in-vehicle air conditioning temperature is reduced by means of intelligent control, so as to solve a problem of the discomfort of the in-vehicle personnel.
  • step 130 is performed by the processor by invoking the corresponding instruction stored in the memory, or is performed by a danger processing unit 43 run by the processor.
  • At least one video stream of the in-vehicle personnel is obtained by using the photographing apparatus, each video stream including information about at least one in-vehicle personnel; action recognition is performed on the in-vehicle personnel on the basis of the video stream; and the prompt information is sent and/or the operation is executed to control the vehicle in response to that the action recognition result belongs to the predetermined dangerous action, where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: the distraction action, the discomfort state, or the non-standard behavior.
  • Whether the in-vehicle personnel make the predetermined dangerous action is determined by means of action recognition, and the corresponding prompt and/or operation are made to the predetermined dangerous action so as to control the vehicle, thereby implementing early detection of the vehicle safety condition to reduce the probability of dangerous situations.
  • the in-vehicle personnel may include a driver and/or a non-driver; the number of in-vehicle personnel generally include at least one (for example: merely including the driver); in order to respectively perform recognition on the action of each in-vehicle personnel, after an image or a video stream is acquired, optionally, the image or the video stream is segmented in terms of different in-vehicle personnel according to different positions (for example, a vehicle seat position), so as to implement performing analysis on the image or the video stream corresponding to each in-vehicle personnel. Because evaluations of dangerous actions of the driver and the non-driver are different in the driving process of the vehicle, when recognizing whether the action is the predetermined dangerous action, optionally, whether the in-vehicle personnel is the driver or the non-driver is first determined.
  • FIG. 2 is a part of a schematic flowchart in an optional example of the method for recognizing the dangerous action of personnel in the vehicle provided in the embodiments of the present disclosure. As shown in FIG. 2 , step 120 includes the followings.
  • At step 202 at least one target area included by the in-vehicle personnel is detected in at least one frame of video image of the video stream.
  • the target area may include, but is not limited to, at least one of the followings: a face local area, an action interactive object, or a limb area and the like.
  • a face local area serves as the target area
  • a face action is usually related to five sense organs of the face, for example, a smoking or food-eating action is related to a mouth, and a phone call action is related to an ear
  • the target area includes, but is not limited to, one of the following parts or any combination thereof: mouth, ear, nose, eye, and eyebrow.
  • a target part on the face is determined according to requirements, where the target part may include one or more parts, and the target part on the face is detected by using a face detection technology.
  • a target image corresponding to the target area is captured from the at least one frame of video image of the video stream according to the target area obtained through detection.
  • the target area is a certain area centering on the target part, for example, at least one part of the face is used as a center on the basis of the face action.
  • An area outside the face in the video stream may include an object related to an action; for example, a smoking action is centered on the mouth, smoke may appear in areas other than the face in a detection image.
  • a position of the target area is determined in at least one frame of video image according to a detection result of the target area, and a capturing size and/or a capturing position of a target image are determined according to the position of the target area in the at least one frame of video image.
  • a target image corresponding to the target area is captured according to a set condition, so that the captured target image better meets the requirements of action recognition.
  • the size of the captured target image is determined according to a distance between the target area and a set position of the face; for example, a target image of the mouth of person A is determined by using a distance between the mouth of person A and a face center point of A, and similarly, a target image of the mouth of person B is determined by using a distance between the mouth of person B and a face center point of B. Because the distance between the mouth and the face center point is related to a feature of the face, the captured target image may better meet the feature of the face. According to the target image captured according to the position of the target area in a video image, noise is reduced, and a more complete image area in which the object related to the action is located may further be included.
  • step 206 action recognition is performed on the in-vehicle personnel on according to the target image.
  • the feature of the target image is extracted, and whether the in-vehicle personnel executes the predetermined dangerous action is determined according to the extracted feature.
  • the predetermined dangerous action includes, but is not limited to, at least one of the following action representations of the in-vehicle personnel: the distraction action, the discomfort state, or the non-standard behavior and the like.
  • potential safety hazards may be generated.
  • Applications such as safety analysis are performed on the in-vehicle personnel by using the action recognition result. For example, when the driver makes the smoking action in the video stream, whether the driver smokes is determined by extracting the feature of the target image of the mouth and determining whether a feature of a cigarette exists in the video stream according to the feature; and if the driver has the smoking action, it is considered that the potential safety hazards exist.
  • the target area is recognized in the video stream, the target image corresponding to the target area is captured in the video image according to the detection result of the target area, and whether the in-vehicle personnel executes the predetermined dangerous action is recognized according to the target image.
  • the target image captured according to the detection result of the target area may be applicable to human bodies having different areas in different video images.
  • the embodiments of the present disclosure have a wide application range. On the basis of using the target image as a basis for action recognition, the embodiments of the present disclosure are conducive to obtaining feature extraction corresponding to the dangerous action more accurately, may reduce a detection interference brought by an unrelated area, and improve the accuracy of action recognition.
  • the smoking action is greatly related to a mouth area, and the action of the driver is recognized by using the mouth and the vicinity of the mouth as the mouth area, so as to determine whether the driver smokes, thereby improving the accuracy of smoking action recognition.
  • FIG. 3 a is a part of a schematic flowchart in another optional example of the method for recognizing the dangerous action of personnel in the vehicle provided in the embodiments of the present disclosure.
  • step 202 includes the following steps.
  • a feature of the in-vehicle personnel included in the at least one frame of video image of the video stream is extracted.
  • the embodiments of the present disclosure mainly aim at performing recognition on some dangerous actions made by the in-vehicle personnel when said personnel is inside the vehicle.
  • the dangerous actions usually are actions related to the limb and face, and recognition on the actions cannot be implemented by means of detection of a human body key point and estimation of a human body posture.
  • the feature is extracted by performing convolution operation on the video image, and action recognition in the video image is implemented according to the extracted feature.
  • the feature of the aforementioned dangerous action is: the limb and/or face local areas, and the action interactive object. Therefore, real-time photographing needs to be performed on the in-vehicle personnel by means of the photographing apparatus, and a video image including the face is obtained. Then, the convolution operation is performed on the video image, and the action feature is extracted.
  • the target area is extracted from the at least one frame of video image based on the feature.
  • the target area in the embodiments is a target area that may include the action.
  • the feature of the dangerous action is first defined, and then a neural network implements determination on whether the dangerous action exists in the video image according to the defined feature and the extracted video image.
  • the neural network in the embodiments is trained, i. e., the neural network may extract the feature of a predetermined action in the video image.
  • the neural network divides a feature area that simultaneously includes the limb and face local areas and the action interactive object, so as to obtain the target area, where the target area may include, but is not limited to, at least one of the followings: the face local area, the action interactive object, or the limb area and the like.
  • the face local area includes, but is not limited to, at least one of the followings: the mouth area, the ear area, or the eye area and the like.
  • the action interactive object includes, but is not limited to, at least one of the followings: a container, a cigarette, a mobile phone, food, a tool, a beverage bottle, glasses, or a mask and the like.
  • the limb area includes, but is not limited to, at least one of the followings: a hand area or a foot area and the like.
  • the dangerous action includes, but is not limited to: drinking water/beverage, smoking, calling, wearing glasses, wearing a mask, makeup, using a tool, eating food, and putting both feet on a steering wheel and the like.
  • action features of drinking water may include: the hand area, the face local area, and a cup
  • action features of smoking may include: the hand area, the face local area, and the cigarette
  • action features of calling may include: the hand area, the face local area, and a mobile phone
  • action features of wearing the glasses may include: the hand area, the face local area, and the glasses
  • action features of wearing the mask may include: the hand area, the face local area, and the mask
  • action features of putting both feet on the steering wheel may include: the foot area and the steering wheel.
  • the recognized actions in the embodiments of the present disclosure may further include fine actions related to the face or the limb, and such fine actions at least include two features, i. e., the face local area and the action interactive object, for example, including two features, i. e., the face local area and the action interactive object, or including two of three features, i. e., the face local area, the action interactive object and the limb, and the like. Therefore, the fine actions indicate multiple actions having high similarity, for example, smoking and yawning are both recognized mainly on the basis of the mouth area, and both include actions of opening and closing the mouth, and a difference between smoking and yawning actions is merely in whether the cigarette (the action interactive object) is further included.
  • the fine actions indicate multiple actions having high similarity, for example, smoking and yawning are both recognized mainly on the basis of the mouth area, and both include actions of opening and closing the mouth, and a difference between smoking and yawning actions is merely in whether the cigarette (the action interactive object) is further
  • the embodiments of the present disclosure implement recognition on the fine actions by extracting the target area to implement action recognition.
  • the target area includes: the face local area, a mobile phone (i. e., the action interactive object), and a hand (i. e., the limb area).
  • a target action frame may also include: the mouth area and the cigarette (i. e., the action interactive object).
  • FIG. 3 b is a schematic diagram of an extracted target area in the method for recognizing the dangerous action of personnel in the vehicle in the embodiments of the present disclosure.
  • the method for recognizing the dangerous action of personnel in the vehicle in the embodiments of the present disclosure is used for performing target area extraction on the video image in the video stream, so as to obtain the target area for performing recognition on the actions.
  • the action of the in-vehicle personnel in the embodiments of the present disclosure is the smoking action. Therefore, the obtained target area is based on the mouth area (the face local area) and the cigarette (the action interactive object); the target area obtained on the basis of the embodiments of the present disclosure may confirm that the in-vehicle personnel in FIG. 3 b smokes.
  • action recognition is performed on the basis of the target area, noise interference in areas of the entire image that are not related to the action of the in-vehicle personnel (for example, the smoking action) is removed, and the accuracy of action recognition of the in-vehicle personnel is improved, for example, the accuracy of smoking action recognition in the embodiments.
  • pre-processing is further performed on the target image.
  • pre-processing is performed on the target image by means of a method such as normalization and equalization; and a recognition result obtained by the pre-processed target image is more accurate.
  • the dangerous action includes, but is not limited to, at least one of the followings: the distraction action, the discomfort state, or the non-standard behavior or the like.
  • the distraction action indicates that the driver further makes an action that is not related to driving and influences the driving attention degree while driving a vehicle, for example: the distraction action includes, but is not limited to, at least one of the followings: calling, drinking water, putting on or taking off sunglasses, putting on or taking off the mask, or eating food or the like.
  • the discomfort state indicates the physical discomfort of the in-vehicle personnel caused by the influence of an in-vehicle environment or own reasons of the in-vehicle personnel in the vehicle driving process, for example: the discomfort state includes, but is not limited to, at least one of the followings: sweat wiping, rubbing an eye, or yawning or the like.
  • the non-standard behavior indicates a behavior that is made by the in-vehicle personnel and does not comply with regulations, for example, the non-standard behavior includes, but is not limited to, at least one of the followings: smoking, stretching a hand out of the vehicle, bending over a steering wheel, putting both feet on the steering wheel, leaving both hands away from the steering wheel, holding an instrument with a hand, or disturbing a driver or the like.
  • step 130 includes the followings.
  • a danger level of the predetermined dangerous action is determined.
  • Corresponding prompt information is sent according to the danger level, and/or an operation corresponding to the danger level is executed and the vehicle is controlled according to the operation.
  • danger level determination is performed on the predetermined dangerous action; optionally, the danger level of the predetermined dangerous action is determined according to a preset rule and/or correspondence, and then how to operate is determined according to the danger level. For example, operations of different degrees are performed according to the dangerous action level of the in-vehicle personnel.
  • the danger level is set to include primary level, intermediate level, and high level.
  • the sending of the corresponding prompt information according to the danger level, and/or the executing of the operation corresponding to the danger level and the controlling of the vehicle according to the operation include the followings.
  • the prompt information is sent in response to that the danger level is the primary level.
  • the operation corresponding to the danger level is executed and the vehicle is controlled according to the operation in response to that the danger level is the intermediate level.
  • the operation corresponding to the danger level is executed and the vehicle is controlled according to the operation while sending the prompt information in response to that the danger level is the high level.
  • the danger level is set as 3 levels.
  • the embodiments of the present disclosure may further set the danger level more in details, and more levels are included, for example, the danger level includes a first level, a second level, a third level, and a fourth level, where each level corresponds to different danger degrees.
  • the prompt information is sent according to different danger levels, and/or the operation corresponding to the danger level is executed and the vehicle is controlled according to the operation. By executing different operations for different danger levels, the sending of the prompt information and the controlling of the operation may be more flexible and adapt to different use requirements.
  • the determining of the danger level of the predetermined dangerous action includes the followings.
  • the frequency and/or duration of occurrence of the predetermined dangerous action in the video stream are acquired, and the danger level of the predetermined dangerous action is determined on the basis of the frequency and/or duration.
  • the embodiments of the present disclosure by performing further abstract analysis on the dangerous action obtained by action recognition, and according to a lasting degree of the action or a priori probability of the occurrence of a dangerous situation, whether a real intention of a passenger performs a dangerous action is output.
  • the embodiments of the present disclosure implement measurement of the action lasting degree by means of the frequency and/or duration of the occurrence of the predetermined dangerous action in the video stream. For example, when the driver just scratches the eye quickly, it is considered as just a quick adjustment, and alarming is not required. However, if the driver rubs the eye for a long time along with the occurrence of the action such as yawning, it is considered that the driver is relatively fatigue and should be prompted. For another example, the alarming strength for smoking may be less than that for actions such as bending over the steering wheel and calling.
  • the action includes a duration of an action
  • an early-warning condition includes: recognizing that the duration of the action exceeds a duration threshold.
  • the action may include the duration of the action; when the duration of the action exceeds the duration threshold, it is considered that the execution of the action distracts much of the attention of an action execution object; said action is considered as a dangerous action, and early-warning information needs to be sent. For example, if the duration of the smoking action of the driver exceeds 3 seconds, it is considered that the smoking action is the dangerous action and influences a driving action of the driver, and the early-warning information needs to be sent to the driver.
  • a sending condition of the prompt information and/or a control condition of the vehicle are adjusted, so that the sending of the prompt information and the controlling of the operation are more flexible and better adapt to different use requirements.
  • the action recognition result includes a duration of an action
  • a condition of belonging to the predetermined dangerous action includes: recognizing that the duration of the action exceeds the duration threshold.
  • the action recognition result includes the number of times for which an action is performed, and a condition of belonging to the predetermined dangerous action includes: recognizing that the number of times exceeds a number threshold.
  • a number threshold When the number of times exceeds the number threshold, it is considered that the action of the action execution object is frequent, and much attention is distracted; the action is considered as the dangerous action, and the early-warning information needs to be sent. For example, if the number of the smoking actions of the driver exceeds 5 times, it is considered that the smoking action is the dangerous action and influences the driving action of the driver, and the prompt information needs to be sent to the driver.
  • the action recognition result includes a duration of an action and the number of times for which the action is performed
  • the condition of belonging to the predetermined dangerous action includes: recognizing that the duration of the action exceeds the duration threshold, and the number of times exceeds the number threshold.
  • the duration of the action exceeds the duration threshold, and the number of times exceeds the number threshold, it is considered that the action of the action execution object is frequent and the duration of the action is long, and much attention is distracted; said action is considered as the dangerous action, and the prompt information needs to be sent and/or the vehicle is controlled, so that the vehicle is more flexibly controlled and adapts to different use requirements.
  • the dangerous actions corresponding to different in-vehicle personnel are different, for example, the driver at a driving seat is required not to be distracted, and the distraction action belongs to the dangerous action; while the distraction action of the in-vehicle personnel at other positions does not belong to the dangerous action. Therefore, in the embodiments of present disclosure, in order to implement more accurate alarming and intelligent control, prompting or an intelligent operation is performed by combining the action category and a category of the in-vehicle personnel, so that user experience is not influenced because of frequently alarming while improving the driving safety.
  • the embodiments of the present disclosure include: sending corresponding first prompt information and/or controlling the vehicle to execute a corresponding first predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the driver; and/or,
  • the in-vehicle personnel are divided into two categories, i. e., the driver and the non-driver. Different dangerous actions are respectively set for the driver and the non-driver, so as to implement flexible alarming and operation.
  • the distraction action of the driver may include, but is not limited to at least one of the followings: calling, drinking water, putting on or taking off sunglasses, putting on or taking off the mask, or eating food and the like.
  • the discomfort state of the driver may include, but is not limited to at least one of the followings: sweat wiping, rubbing an eye, or yawning and the like.
  • the non-standard behavior of the driver may include, but is not limited to at least one of the followings: smoking, stretching the hand out of the vehicle, bending over the steering wheel, putting both feet on the steering wheel, or leaving both hands away from the steering wheel and the like.
  • the discomfort state of the non-driver may include, but is not limited to at least one of the followings: sweat wiping or the like.
  • the non-standard behavior of the non-driver may include, but is not limited to at least one of the followings: smoking, stretching the hand out of the vehicle, holding the instrument with the hand, or disturbing the driver or the like.
  • different prompt information and predetermined operations are further respectively set for the driver and the non-driver, so as to implement flexible safety control of the vehicle, for example, when the driver makes an action of leaving both hands from the steering wheel, automatically driving (for example, the corresponding first predetermined operation) is executed while sending strong prompt information (for example, the corresponding first prompt information), so as to improve the safety of vehicle driving; moreover, for the non-driver, for example, when the non-driver makes an action of sweat wiping, weak prompt information (for example, the corresponding second prompt information) is sent, and/or an operation of adjusting an in-vehicle air conditioning temperature (for example, the corresponding second predetermined operation) is executed.
  • weak prompt information for example, the corresponding second prompt information
  • an in-vehicle air conditioning temperature for example, the corresponding second predetermined operation
  • all or some steps of implementing the forgoing embodiments of the method may be achieved by a program by instructing related hardware; the foregoing program may be stored in a computer readable storage medium; when the program is executed, steps including the foregoing embodiments of the method are performed; moreover, the foregoing storage medium includes various media capable of storing program codes such as an ROM, an RAM, a magnetic disk, or an optical disk.
  • FIG. 4 is a schematic structural diagram of a device for recognizing a dangerous action of personnel in a vehicle (i.e., in-vehicle personnel) provided in the embodiments of the present disclosure.
  • the apparatus of this embodiment is configured to implement the foregoing method embodiments of the present disclosure. As shown in FIG. 4 , the apparatus of this embodiment includes:
  • a video collection unit 41 used for obtaining at least one video stream of the in-vehicle personnel through an image capturing device
  • each video stream including information about at least one in-vehicle personnel
  • an action recognition unit 42 used for performing action recognition on the in-vehicle personnel on the basis of the video stream;
  • a danger processing unit 43 used for sending prompt information and/or executing an operation to control a vehicle in response to that an action recognition result belongs to a predetermined dangerous action
  • the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: a distraction action, a discomfort state, or a non-standard behavior and the like.
  • the action recognition unit 42 is used for detecting at least one target area included by the in-vehicle personnel in at least one frame of video image of the video stream, capturing a target image corresponding to the target area from the at least one frame of video image of the video stream according to the target area obtained through detection, and performing action recognition on the in-vehicle personnel according to the target image.
  • the target area is recognized in the video stream, the target image corresponding to the target area is captured in the video image according to the detection result of the target area, and whether the in-vehicle personnel executes the predetermined dangerous action is recognized according to the target image.
  • the target image captured according to the detection result of the target area may be applicable to human bodies having different areas in different video images.
  • the embodiments of the present disclosure have a wide application range.
  • the action recognition unit 42 is used for extracting a feature of the in-vehicle personnel included in the at least one frame of video image of the video stream when detecting at least one target area included by the in-vehicle personnel in the at least one frame of video image of the video frame, and extracting a target area from the at least one frame of video image frame on the basis of the feature, where the target area includes, but is not limited to, at least one of the followings: the face local area, the action interactive object, or the limb area or the like.
  • the face local area includes, but is not limited to, at least one of the followings: the mouth area, the ear area, or the eye area or the like.
  • the action interactive object includes, but is not limited to, at least one of the followings: a container, a cigarette, a mobile phone, food, a tool, a beverage bottle, glasses, or a mask or the like.
  • the distraction action includes, but is not limited to, at least one of the followings: calling, drinking water, putting on or taking off sunglasses, putting on or taking off the mask, or eating food or the like; and/or,
  • the discomfort state includes, but is not limited to, at least one of the followings: wiping sweat, rubbing an eye, or yawning; and/or,
  • the non-standard behavior includes, but is not limited to, at least one of the followings: smoking, stretching a hand out of the vehicle, bending over a steering wheel, putting both feet on the steering wheel, leaving both hands away from the steering wheel, holding an instrument with a hand, or disturbing a driver or the like.
  • the danger processing unit 43 includes:
  • a level determination module used for determining a danger level of the predetermined dangerous action in response to that the action recognition result belongs to the predetermined dangerous action
  • an operation processing module used for sending corresponding prompt information according to the danger level, and/or executing an operation corresponding to the danger level and controlling the vehicle according to the operation.
  • danger level determination is performed on the predetermined dangerous action; optionally, the danger level of the predetermined dangerous action is determined according to a preset rule and/or correspondence, and then how to operate is determined according to the danger level. For example, operations of different degrees are performed according to the dangerous action level of the in-vehicle personnel.
  • the dangerous action is caused by fatigue and physical discomfort of the driver, timely prompting is required, so that the driver performs adjustment and has a rest in time; and when the driver feels discomfort due to the in-vehicle environment, an adjustment of a certain degree is performed by controlling a ventilation system or an air conditioning system in the vehicle.
  • the danger level includes a primary level, an intermediate level, and a high level.
  • the operation processing module is used for sending the prompt information in response to that the danger level is the primary level, executing the operation corresponding to the danger level and controlling the vehicle according to the operation in response to that the danger level is the intermediate level, and executing the operation corresponding to the danger level and controlling the vehicle according to the operation while sending the prompt information in response to that the danger level is the high level.
  • the level determination module is used for acquiring the frequency and/or duration of occurrence of the predetermined dangerous action in the video stream, and determining the danger level of the predetermined dangerous action on the basis of the frequency and/or duration.
  • the action recognition result includes the action duration
  • the condition of belonging to the predetermined dangerous action includes: recognizing that the action duration exceeds the duration threshold.
  • the embodiments of the present disclosure by performing further abstract analysis on the dangerous action obtained by action recognition, and according to the lasting degree of the action or the priori probability of the occurrence of the dangerous situation, whether the real intention of the passenger performs the dangerous action is output.
  • the embodiments of the present disclosure implement measurement of the action lasting degree by means of the frequency and/or duration of the occurrence of the predetermined dangerous action in the video stream.
  • the action recognition result includes the action duration
  • the condition of belonging to the predetermined dangerous action includes: recognizing that the action duration exceeds the duration threshold.
  • the action recognition result includes the number of actions
  • the condition of belonging to the predetermined dangerous action includes: recognizing that the number of actions exceeds the number threshold.
  • the action recognition result includes the action duration and the number of actions
  • the condition of belonging to the predetermined dangerous action includes: recognizing that the action duration exceeds the duration threshold, and the number of actions exceeds the number threshold.
  • the in-vehicle personnel includes the driver and/or the non-driver of the vehicle.
  • the danger processing unit 43 is used for sending corresponding first prompt information and/or controlling the vehicle to execute a corresponding first predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the driver, and/or sending corresponding second prompt information and/or executing a corresponding second predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the non-driver.
  • an electronic device includes a processor, where the processor includes the in-vehicle personnel dangerous action recognition apparatus provided according to any one of the foregoing embodiments.
  • an electronic device includes: a memory used for storing executable instructions;
  • a computer readable storage medium is provided and is used for storing computer readable instructions, where when the instructions are executed, the operations of the method for recognizing the dangerous actin of the personnel in the vehicle provided according to any one of the foregoing embodiments are executed.
  • a computer program includes a computer readable code, where if the computer readable code runs on a device, a processor in the device executes an instruction for implementing the method for recognizing the dangerous actin of the personnel in the vehicle provided according to any one of the foregoing embodiments.
  • the embodiments of the present disclosure further provide an electronic device which, for example, is a mobile terminal, a Personal Computer (PC), a tablet computer, a server, and the like.
  • an electronic device which, for example, is a mobile terminal, a Personal Computer (PC), a tablet computer, a server, and the like.
  • FIG. 5 a schematic structural diagram of an electronic device 500 , which is a terminal device or a server, suitable for implementing the embodiments of the present disclosure is shown.
  • the electronic device 500 includes one or more processors, a communication part, and the like.
  • the one or more processors are, for example, one or more Central Processing Units (CPUs) 501 and/or one or more graphic processing unit (an acceleration unit) 513 , and may execute appropriate actions and processing according to executable instructions stored in a Read-Only Memory (ROM) 502 or executable instructions loaded from a storage section 508 to a Random Access Memory (RAM) 503 .
  • the communication part 512 may include, but is not limited to, a network card.
  • the network card may include, but is not limited to, an Infiniband (IB) network card.
  • the processor communicates with the ROM 502 and/or the RAM 503 to execute the executable instructions, and is connected to the communication part 512 by means of a bus 504 and communicates with other target devices by means of the communication part 512 , so as to complete the operations corresponding to any of the methods provided in the embodiments of the present disclosure, for example, obtaining at least one video stream of in-vehicle personnel by using a photographing apparatus, each video stream including information about at least one in-vehicle personnel; performing action recognition on the in-vehicle personnel on the basis of the video stream; and sending prompt information and/or executing an operation to control a vehicle in response to that an action recognition result belongs to a predetermined dangerous action, where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: a distraction action, a discomfort state, or a non-standard behavior.
  • the RAM 503 may further store various programs and data required for operations of an apparatus.
  • the CPU 501 , the ROM 502 , and the RAM 503 are connected to each other by means of the bus 504 .
  • the ROM 502 is an optional module.
  • the RAM 503 stores executable instructions, or writes the executable instructions into the ROM 502 during running, where the executable instructions cause the CPU 501 to execute corresponding operations of the foregoing communication method.
  • An Input/Output (I/O) interface 505 is also connected to the bus 504 .
  • the communication part 512 is integrated, or is also configured to have multiple sub-modules (for example, multiple IB network cards) linked on the bus.
  • the following components are connected to the I/O interface 505 : an input section 506 including a keyboard, a mouse and the like; an output section 507 including a Cathode-Ray Tube (CRT), a Liquid Crystal Display (LCD), a speaker and the like; the storage section 508 including a hard disk drive and the like; and a communication section 509 of a network interface card including an LAN card, a modem and the like.
  • the communication section 509 performs communication processing via a network such as the Internet.
  • a drive 510 is also connected to the I/O interface 505 according to requirements.
  • a removable medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like is installed on the drive 510 according to requirements, so that a computer program read from the removable medium is installed on the storage section 508 according to requirements.
  • FIG. 5 is merely an optional implementation. During specific practice, the number and types of the components in FIG. 5 may be selected, decreased, increased, or replaced according to actual requirements. Different functional components may be configured separately or integrally or the like.
  • an acceleration unit 513 and the CPU 501 may be configured separately, or the acceleration unit 513 may be integrated on the CPU 501 , and the communication part may be configured separately, and may also be configured integrally on the CPU 501 or the acceleration unit 513 or the like.
  • a process described above with reference to a flowchart according to the embodiments of the present disclosure is implemented as a computer software program.
  • the embodiments of the disclosure include a computer program product.
  • the computer program product includes a computer program tangibly included in a machine-readable medium.
  • the computer program includes a program code for performing a method shown in the flowchart.
  • the program code may include instructions for correspondingly performing steps of the method provided in the embodiments of the present disclosure, for example, obtaining at least one video stream of in-vehicle personnel by using the photographing apparatus, each video stream including information about at least one in-vehicle personnel; performing action recognition on the in-vehicle personnel on the basis of the video stream; and sending the prompt information and/or executing the operation to control the vehicle in response to that the action recognition result belongs to the predetermined dangerous action, where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: the distraction action, the discomfort state, or the non-standard behavior.
  • the computer program is downloaded and installed from the network by means of the communication section 509 , and/or is installed from the removable medium 511 .
  • the computer program when being executed by the CPU 501 , executes the operations of the foregoing functions defined in the method of the present disclosure.
  • the methods and apparatuses of the present disclosure are implemented in many manners.
  • the methods and apparatuses of the present disclosure are implemented by means of software, hardware, firmware, or any combination of software, hardware, and firmware. Unless otherwise specially stated, the foregoing sequences of steps of the methods are merely for description, and are not intended to limit the steps of the methods of the present disclosure.
  • the present disclosure may also be implemented as programs recorded in a recording medium.
  • the programs include machine-readable instructions for implementing the methods according to the present disclosure. Therefore, the present disclosure further covers the recording medium storing the programs for executing the methods according to the present disclosure.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Automation & Control Theory (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Mechanical Engineering (AREA)
  • Transportation (AREA)
  • Evolutionary Computation (AREA)
  • Ophthalmology & Optometry (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Traffic Control Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)
  • Eye Examination Apparatus (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)

Abstract

A method for recognizing a dangerous action of personnel in a vehicle, an electronic device, and a storage medium are provided. The method includes: obtaining at least one video stream of the personnel in the vehicle through an image capturing device, each video stream includes information about at least one of the personnel in the vehicle; performing action recognition on the personnel in the vehicle based on the video stream; and responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, wherein the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • The present application is a continuation application of PCT Application No. PCT/CN2019/129370, filed on Dec. 27, 2019, which claims priority to Chinese Patent Application No. CN 201910152525.X, filed on Feb. 28, 2019 and entitled “METHOD AND DEVICE FOR RECOGNIZING DANGEROUS ACTION OF PERSONNEL IN VEHICLE, ELECTRONIC DEVICE, AND STORAGE MEDIUM”. The contents of PCT Application No. PCT/CN2019/129370 and CN 201910152525.X are incorporated herein by reference in their entireties.
  • TECHNICAL FIELD
  • The present disclosure relates to computer vision technologies, and in particular, to a method and device for recognizing a dangerous action of personnel in a vehicle, an electronic device, and a storage medium.
  • BACKGROUND
  • With the rapid development of in-vehicle personnel and employee intelligence, various artificial intelligence (AI) technologies are implemented, and at present, the demand for driver monitoring in the market is increasingly urgent. The main function modules of driver monitoring may be generally summarized as modules such as in-vehicle personnel face recognition module and fatigue detection module. By monitoring the state of a driver, a danger signal may be found out in time, and a possible danger may be prevented and dealt with in advance so as to improve driving safety.
  • SUMMARY
  • In a first aspect, a method for recognizing a dangerous action is provided, including:
  • obtaining at least one video stream of personnel in a vehicle through an image capturing device, each video stream including information about at least one of the personnel in the vehicle;
  • performing action recognition on the personnel in the vehicle based on the video stream; and
  • responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
  • In a second aspect, a device for recognizing a dangerous action of personnel in a vehicle is provided, including:
  • a video collection unit, used for obtaining at least one video stream of the personnel in the vehicle through an image capturing device, each video stream including information about at least one of the personnel in the vehicle;
  • an action recognition unit, used for performing action recognition on the personnel in the vehicle based on the video stream; and
  • a danger processing unit, used for, responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
  • In a third aspect, an electronic device is provided and includes a processor, where the processor includes the device for recognizing the dangerous action of the personnel in the vehicle according to the first aspect.
  • In a fourth aspect, an electronic device is provided and includes: a processor; and a memory configured to store instructions that, when executed by the processor, cause the processor to perform the following operations including:
  • obtaining at least one video stream of personnel in a vehicle through an image capturing device, each video stream including information about at least one of the personnel in the vehicle;
  • performing action recognition on the personnel in the vehicle based on the video stream; and
  • responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior
  • In a fifth aspect, a non-transitory computer readable storage medium is provided and is used for storing computer readable instructions that, when executed by a processor of an electronic device, cause the processor to perform a method for recognizing a dangerous action is provided, including:
  • obtaining at least one video stream of personnel in a vehicle through an image capturing device, each video stream including information about at least one of the personnel in the vehicle;
  • performing action recognition on the personnel in the vehicle based on the video stream; and
  • responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, where the predetermined dangerous action includes at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
  • In a sixth aspect, a computer program product is provided and includes computer readable codes that, when being run on a device, cause a processor in the device to execute instructions for implementing the method for recognizing the dangerous action of the personnel in the vehicle according to the first aspect.
  • The following further describes in detail the technical solutions of the present disclosure with reference to the accompanying drawings and embodiments.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings constituting a part of the specification describe the embodiments of the present disclosure and are intended to explain the principles of the present disclosure together with the descriptions.
  • According to the following detailed descriptions, the present disclosure may be understood more clearly with reference to the accompanying drawings.
  • FIG. 1 is a schematic flowchart of a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 2 is a part of a schematic flowchart in an optional example of a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 3a is a part of a schematic flowchart in another optional example of a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 3b is a schematic diagram of an extracted target area in a method for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 4 is a schematic structural diagram of a device for recognizing a dangerous action of personnel in a vehicle provided in the embodiments of the present disclosure.
  • FIG. 5 is a schematic structural diagram of an electronic device, which may be a terminal device or a server, suitable for implementing the embodiments of the present disclosure.
  • DETAILED DESCRIPTION
  • Exemplary embodiments of the present disclosure are described in detail with reference to the accompany drawings now. It should be noted that, unless otherwise stated specifically, relative arrangement of the components and steps, the numerical expressions, and the values set forth in the embodiments are not intended to limit the scope of the present disclosure.
  • In addition, it should be understood that, for ease of description, the size of each part shown in the accompanying drawings is not drawn in actual proportion.
  • The following descriptions of at least one exemplary embodiment are merely illustrative actually, and are not intended to limit the present disclosure and the applications or uses thereof.
  • Technologies, methods and devices known to a person of ordinary skill in the related art may not be discussed in detail, but such technologies, methods and devices should be considered as a part of the specification in appropriate situations.
  • It should be noted that similar reference numerals and letters in the following accompanying drawings represent similar items. Therefore, once an item is defined in an accompanying drawing, the item does not need to be further discussed in the subsequent accompanying drawings.
  • The embodiments of the present disclosure may be applied to a computer system/server, which may operate with numerous other general-purpose or special-purpose computing system environments or configurations. Examples of well-known computing systems, environments, and/or configurations suitable for use together with the computer system/server include, but are not limited to, personal computer systems, server computer systems, thin clients, thick clients, handheld or laptop devices, microprocessor-based systems, set top boxes, programmable consumer electronics, network personal computers, small computer systems, large computer systems, distributed cloud computing environments that include any one of the foregoing systems, and the like.
  • The computer system/server may be described in the general context of computer system executable instructions (for example, program modules) executed by the computer system. Generally, the program modules may include routines, programs, target programs, components, logics, data structures, and the like for performing specific tasks or implementing specific abstract data types. The computer system/server may be practiced in the distributed cloud computing environments in which tasks are performed by remote processing devices that are linked through a communications network. In the distributed cloud computing environments, the program modules may be located in local or remote computing system storage media including storage devices.
  • Dangerous action recognition has a wide application prospect in in-vehicle security monitoring fields. First, a dangerous action recognition system may give a prompt when a driver makes a dangerous action, thereby performing early-warning and avoiding possible accidents; furthermore, the system may monitor some behaviors that are substandard or may cause discomfort of an in-vehicle passenger to give a prompt and restrain the behaviors; meanwhile, the monitoring of the dangerous action reflects some habits and hobbies of a driver, which facilitates the system to establish a user portrait and perform big data analysis; and an emotional state, a fatigue state, and a behavior habit of the driver is monitored by means of dangerous action recognition.
  • FIG. 1 is a schematic flowchart of a method for recognizing a dangerous action of personnel in a vehicle (i.e., in-vehicle personnel) provided in the embodiments of the present disclosure. The method is performed by any electronic device, such as a terminal device, a server, a mobile device, and a vehicle-mounted device. As shown in FIG. 1, the method of this embodiment includes the following steps.
  • At step 110, at least one video stream of in-vehicle personnel is obtained through an image capturing device (for example, a camera or the like).
  • Each video stream includes information about at least one of the personnel in the vehicle; in the embodiments of the present disclosure, an image of the in-vehicle personnel is collected by means of a photographing apparatus (also called the image capturing device, for example, one or more cameras provided in the vehicle and used for performing photographing on a vehicle seat position), so as to obtain the video stream; optionally, the video streams of multiple in-vehicle personnel (for example, all of in-vehicle personnel) in the whole vehicle are collected on the basis of one photographing apparatus, or one photographing apparatus facing one or more back row areas for image collection is provided in the vehicle, or one photographing apparatus is respectively provided in the front of each seat, so as to respectively collect the video stream of at least one in-vehicle personnel (for example, each in-vehicle personnel); and action recognition is respectively performed on the in-vehicle personnel by performing processing on the collected video stream.
  • In practical applications, a situation that video stream collection is performed on the in-vehicle personnel on the basis of an out-vehicle photographing apparatus (for example, a camera provided on a road) may further exist.
  • Optionally, the photographing apparatus includes, but is not limited to, at least one of the followings: a visible light camera, an infrared camera, or a near-infrared camera. The visible light camera is used for collecting a visible light image, the infrared camera is used for collecting an infrared image, and the near-infrared camera is used for collecting a near-infrared image.
  • In an optional example, step 110 is executed by a processor by invoking a corresponding instruction stored in a memory, or is executed by a video collection unit 41 run by the processor.
  • At step 120, action recognition is performed on the in-vehicle personnel on the basis of the video stream.
  • In a driving process of the vehicle, if one or more of in-vehicle personnel make a dangerous action, the danger of the vehicle is caused, and in particular, if the driver makes some dangerous actions, the whole vehicle may be in danger, thereby causing danger to the vehicle and the in-vehicle personnel. Therefore, it is necessary to perform recognition on the actions of the in-vehicle personnel, so as to ensure the safety of the vehicle. Some actions are determined on the basis of a single image frame in the video stream, while some actions require continuous multiple frames to be recognized. Therefore, in the embodiments of the present disclosure, recognition is performed on the actions by means of the video stream to reduce misjudgment, thereby improving the accuracy of action recognition.
  • Optionally, an action category is divided into a dangerous action and a normal action, and the dangerous action needs to be processed to exclude a possible danger, where the dangerous action includes, but is not limited to, at least one of the followings: a distraction action, a discomfort state, or a non-standard behavior or the like. The dangerous action may have same or different requirements for an ordinary non-driver and driver, for example, the requirements for the driver are relatively stricter, and meanwhile, the independence and safety of the driver need to be protected, for example, the predetermined dangerous action is divided into the dangerous action of the driver and the dangerous action of the non-driver. The embodiments of the present disclosure do not limit a specific mode for recognizing the action category.
  • In an optional example, step 120 is performed by the processor by invoking the corresponding instruction stored in the memory, or is performed by an action recognition unit 42 run by the processor.
  • At step 130, prompt information is sent and/or an operation is executed to control the vehicle in response to that an action recognition result (i.e., a result of the action recognition) belongs to a predetermined dangerous action.
  • The dangerous action in the embodiments of the present disclosure is a behavior that causes potential safety hazards to the in-vehicle personnel or others. Optionally, the predetermined dangerous action in the embodiments of the present disclosure includes, but is not limited to, at least one of the following action representations of the in-vehicle personnel: a distraction action, a discomfort state, or a non-standard behavior and the like. Optionally, the distraction action mainly aims at the driver; in a process of driving a vehicle, the driver needs to concentrate; when the distraction action (for example, actions such as eating food and smoking) occurs, the attention of the driver is influenced, and the vehicle is prone to be in danger; the discomfort state may aim at all of the in-vehicle personnel; when the discomfort state occurs in the in-vehicle personnel, on the basis of human safety considerations, some dangerous situations need to be processed in time, for example, situations such as frequent yawning of the driver or sweat wiping of a passenger; and the non-standard behavior is a behavior that do not comply with safe driving regulations, and may further be a behavior that is dangerous to the driver or other people in the vehicle, etc. In order to overcome an adverse effect of the predetermined dangerous action, in the embodiments of the present disclosure, the probability of danger is reduced by sending the prompt information or executing the operation to control the vehicle, and the safety and/or comfort level of the in-vehicle personnel are improved.
  • Optionally, a representation form of the prompt information may include, but is not limited to, at least one of the followings: voice prompt information, vibration prompt information, light prompt information, or smell prompt information and the like; for example, when the in-vehicle personnel smokes, the voice prompt information is sent to prompt that smoking in the vehicle is not allowed, so as to reduce the danger of smoking to other in-vehicle personnel; for another example: when the in-vehicle personnel wipes sweat, it means that a temperature in the vehicle is too high, and an in-vehicle air conditioning temperature is reduced by means of intelligent control, so as to solve a problem of the discomfort of the in-vehicle personnel.
  • Dangerous action recognition has an important status and high application value in driver monitoring. At present, in the driving process of the driver, many dangerous actions commonly exist, and the actions often cause the driver to be distracted, so that certain potential safety hazards exist.
  • In an optional example, step 130 is performed by the processor by invoking the corresponding instruction stored in the memory, or is performed by a danger processing unit 43 run by the processor.
  • On the basis of the method for recognizing the dangerous action of personnel in the vehicle provided in the forgoing embodiments of the present disclosure, at least one video stream of the in-vehicle personnel is obtained by using the photographing apparatus, each video stream including information about at least one in-vehicle personnel; action recognition is performed on the in-vehicle personnel on the basis of the video stream; and the prompt information is sent and/or the operation is executed to control the vehicle in response to that the action recognition result belongs to the predetermined dangerous action, where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: the distraction action, the discomfort state, or the non-standard behavior. Whether the in-vehicle personnel make the predetermined dangerous action is determined by means of action recognition, and the corresponding prompt and/or operation are made to the predetermined dangerous action so as to control the vehicle, thereby implementing early detection of the vehicle safety condition to reduce the probability of dangerous situations.
  • Optionally, the in-vehicle personnel may include a driver and/or a non-driver; the number of in-vehicle personnel generally include at least one (for example: merely including the driver); in order to respectively perform recognition on the action of each in-vehicle personnel, after an image or a video stream is acquired, optionally, the image or the video stream is segmented in terms of different in-vehicle personnel according to different positions (for example, a vehicle seat position), so as to implement performing analysis on the image or the video stream corresponding to each in-vehicle personnel. Because evaluations of dangerous actions of the driver and the non-driver are different in the driving process of the vehicle, when recognizing whether the action is the predetermined dangerous action, optionally, whether the in-vehicle personnel is the driver or the non-driver is first determined.
  • FIG. 2 is a part of a schematic flowchart in an optional example of the method for recognizing the dangerous action of personnel in the vehicle provided in the embodiments of the present disclosure. As shown in FIG. 2, step 120 includes the followings.
  • At step 202, at least one target area included by the in-vehicle personnel is detected in at least one frame of video image of the video stream.
  • In a possible implementation, in order to implement action recognition, the target area may include, but is not limited to, at least one of the followings: a face local area, an action interactive object, or a limb area and the like. For example, when the face local area serves as the target area, because a face action is usually related to five sense organs of the face, for example, a smoking or food-eating action is related to a mouth, and a phone call action is related to an ear; in said example, the target area includes, but is not limited to, one of the following parts or any combination thereof: mouth, ear, nose, eye, and eyebrow. Optionally, a target part on the face is determined according to requirements, where the target part may include one or more parts, and the target part on the face is detected by using a face detection technology.
  • At step 204, a target image corresponding to the target area is captured from the at least one frame of video image of the video stream according to the target area obtained through detection.
  • In a possible implementation, the target area is a certain area centering on the target part, for example, at least one part of the face is used as a center on the basis of the face action. An area outside the face in the video stream may include an object related to an action; for example, a smoking action is centered on the mouth, smoke may appear in areas other than the face in a detection image.
  • In a possible implementation, a position of the target area is determined in at least one frame of video image according to a detection result of the target area, and a capturing size and/or a capturing position of a target image are determined according to the position of the target area in the at least one frame of video image. In the embodiments of the present disclosure, a target image corresponding to the target area is captured according to a set condition, so that the captured target image better meets the requirements of action recognition. For example, the size of the captured target image is determined according to a distance between the target area and a set position of the face; for example, a target image of the mouth of person A is determined by using a distance between the mouth of person A and a face center point of A, and similarly, a target image of the mouth of person B is determined by using a distance between the mouth of person B and a face center point of B. Because the distance between the mouth and the face center point is related to a feature of the face, the captured target image may better meet the feature of the face. According to the target image captured according to the position of the target area in a video image, noise is reduced, and a more complete image area in which the object related to the action is located may further be included.
  • At step 206, action recognition is performed on the in-vehicle personnel on according to the target image.
  • In a possible implementation, the feature of the target image is extracted, and whether the in-vehicle personnel executes the predetermined dangerous action is determined according to the extracted feature.
  • In a possible implementation, the predetermined dangerous action includes, but is not limited to, at least one of the following action representations of the in-vehicle personnel: the distraction action, the discomfort state, or the non-standard behavior and the like. When the in-vehicle personnel executes the predetermined dangerous action, potential safety hazards may be generated. Applications such as safety analysis are performed on the in-vehicle personnel by using the action recognition result. For example, when the driver makes the smoking action in the video stream, whether the driver smokes is determined by extracting the feature of the target image of the mouth and determining whether a feature of a cigarette exists in the video stream according to the feature; and if the driver has the smoking action, it is considered that the potential safety hazards exist.
  • In the embodiments, the target area is recognized in the video stream, the target image corresponding to the target area is captured in the video image according to the detection result of the target area, and whether the in-vehicle personnel executes the predetermined dangerous action is recognized according to the target image. The target image captured according to the detection result of the target area may be applicable to human bodies having different areas in different video images. The embodiments of the present disclosure have a wide application range. On the basis of using the target image as a basis for action recognition, the embodiments of the present disclosure are conducive to obtaining feature extraction corresponding to the dangerous action more accurately, may reduce a detection interference brought by an unrelated area, and improve the accuracy of action recognition. For example, when recognizing the smoking action of the driver, the smoking action is greatly related to a mouth area, and the action of the driver is recognized by using the mouth and the vicinity of the mouth as the mouth area, so as to determine whether the driver smokes, thereby improving the accuracy of smoking action recognition.
  • FIG. 3a is a part of a schematic flowchart in another optional example of the method for recognizing the dangerous action of personnel in the vehicle provided in the embodiments of the present disclosure. As shown in FIG. 3a , in the method provided in the aforementioned embodiments, step 202 includes the following steps.
  • At step 302, a feature of the in-vehicle personnel included in the at least one frame of video image of the video stream is extracted.
  • The embodiments of the present disclosure mainly aim at performing recognition on some dangerous actions made by the in-vehicle personnel when said personnel is inside the vehicle. Moreover, the dangerous actions usually are actions related to the limb and face, and recognition on the actions cannot be implemented by means of detection of a human body key point and estimation of a human body posture. In the embodiments of the present disclosure, the feature is extracted by performing convolution operation on the video image, and action recognition in the video image is implemented according to the extracted feature. For example, the feature of the aforementioned dangerous action is: the limb and/or face local areas, and the action interactive object. Therefore, real-time photographing needs to be performed on the in-vehicle personnel by means of the photographing apparatus, and a video image including the face is obtained. Then, the convolution operation is performed on the video image, and the action feature is extracted.
  • At step 304, the target area is extracted from the at least one frame of video image based on the feature.
  • Optionally, the target area in the embodiments is a target area that may include the action.
  • The feature of the dangerous action is first defined, and then a neural network implements determination on whether the dangerous action exists in the video image according to the defined feature and the extracted video image. The neural network in the embodiments is trained, i. e., the neural network may extract the feature of a predetermined action in the video image.
  • If the extracted feature include: the limb area, the face local area, and the action interactive object, the neural network divides a feature area that simultaneously includes the limb and face local areas and the action interactive object, so as to obtain the target area, where the target area may include, but is not limited to, at least one of the followings: the face local area, the action interactive object, or the limb area and the like. Optionally, the face local area includes, but is not limited to, at least one of the followings: the mouth area, the ear area, or the eye area and the like. Optionally, the action interactive object includes, but is not limited to, at least one of the followings: a container, a cigarette, a mobile phone, food, a tool, a beverage bottle, glasses, or a mask and the like. Optionally, the limb area includes, but is not limited to, at least one of the followings: a hand area or a foot area and the like. For example, the dangerous action includes, but is not limited to: drinking water/beverage, smoking, calling, wearing glasses, wearing a mask, makeup, using a tool, eating food, and putting both feet on a steering wheel and the like. Exemplarily, action features of drinking water may include: the hand area, the face local area, and a cup; action features of smoking may include: the hand area, the face local area, and the cigarette; action features of calling may include: the hand area, the face local area, and a mobile phone; action features of wearing the glasses may include: the hand area, the face local area, and the glasses; action features of wearing the mask may include: the hand area, the face local area, and the mask; and action features of putting both feet on the steering wheel may include: the foot area and the steering wheel.
  • The recognized actions in the embodiments of the present disclosure may further include fine actions related to the face or the limb, and such fine actions at least include two features, i. e., the face local area and the action interactive object, for example, including two features, i. e., the face local area and the action interactive object, or including two of three features, i. e., the face local area, the action interactive object and the limb, and the like. Therefore, the fine actions indicate multiple actions having high similarity, for example, smoking and yawning are both recognized mainly on the basis of the mouth area, and both include actions of opening and closing the mouth, and a difference between smoking and yawning actions is merely in whether the cigarette (the action interactive object) is further included. Therefore, the embodiments of the present disclosure implement recognition on the fine actions by extracting the target area to implement action recognition. For example, for the calling action, the target area includes: the face local area, a mobile phone (i. e., the action interactive object), and a hand (i. e., the limb area). For another example, for the smoking action, a target action frame may also include: the mouth area and the cigarette (i. e., the action interactive object).
  • FIG. 3b is a schematic diagram of an extracted target area in the method for recognizing the dangerous action of personnel in the vehicle in the embodiments of the present disclosure. The method for recognizing the dangerous action of personnel in the vehicle in the embodiments of the present disclosure is used for performing target area extraction on the video image in the video stream, so as to obtain the target area for performing recognition on the actions. The action of the in-vehicle personnel in the embodiments of the present disclosure is the smoking action. Therefore, the obtained target area is based on the mouth area (the face local area) and the cigarette (the action interactive object); the target area obtained on the basis of the embodiments of the present disclosure may confirm that the in-vehicle personnel in FIG. 3b smokes. In the embodiments of the present disclosure, by obtaining the target area, action recognition is performed on the basis of the target area, noise interference in areas of the entire image that are not related to the action of the in-vehicle personnel (for example, the smoking action) is removed, and the accuracy of action recognition of the in-vehicle personnel is improved, for example, the accuracy of smoking action recognition in the embodiments.
  • Optionally, before performing action recognition on the in-vehicle personnel according to the target image, pre-processing is further performed on the target image. For example, pre-processing is performed on the target image by means of a method such as normalization and equalization; and a recognition result obtained by the pre-processed target image is more accurate.
  • Optionally, the dangerous action includes, but is not limited to, at least one of the followings: the distraction action, the discomfort state, or the non-standard behavior or the like. The distraction action indicates that the driver further makes an action that is not related to driving and influences the driving attention degree while driving a vehicle, for example: the distraction action includes, but is not limited to, at least one of the followings: calling, drinking water, putting on or taking off sunglasses, putting on or taking off the mask, or eating food or the like. The discomfort state indicates the physical discomfort of the in-vehicle personnel caused by the influence of an in-vehicle environment or own reasons of the in-vehicle personnel in the vehicle driving process, for example: the discomfort state includes, but is not limited to, at least one of the followings: sweat wiping, rubbing an eye, or yawning or the like. The non-standard behavior indicates a behavior that is made by the in-vehicle personnel and does not comply with regulations, for example, the non-standard behavior includes, but is not limited to, at least one of the followings: smoking, stretching a hand out of the vehicle, bending over a steering wheel, putting both feet on the steering wheel, leaving both hands away from the steering wheel, holding an instrument with a hand, or disturbing a driver or the like. Because multiple dangerous actions are included, when the action category of the in-vehicle personnel belongs to the dangerous action, it is necessary to first determine to which dangerous action the action category belongs, and different dangerous actions may correspond to different processing modes (for example, sending the prompt information or executing the operation to control the vehicle).
  • In one or more optional embodiments, step 130 includes the followings.
  • The fact that the action recognition result belongs to the predetermined dangerous action is responded.
  • A danger level of the predetermined dangerous action is determined.
  • Corresponding prompt information is sent according to the danger level, and/or an operation corresponding to the danger level is executed and the vehicle is controlled according to the operation.
  • Optionally, in the embodiments of the present disclosure, when the action of the in-vehicle personnel is determined to belong to the predetermined dangerous action according to the action recognition result, danger level determination is performed on the predetermined dangerous action; optionally, the danger level of the predetermined dangerous action is determined according to a preset rule and/or correspondence, and then how to operate is determined according to the danger level. For example, operations of different degrees are performed according to the dangerous action level of the in-vehicle personnel. For example, if the dangerous action is caused by fatigue and physical discomfort of the driver, timely prompting is required, so that the driver performs adjustment and has a rest in time; and when the driver feels discomfort due to the in-vehicle environment, an adjustment of a certain degree is performed by controlling a ventilation system or an air conditioning system in the vehicle. Optionally, the danger level is set to include primary level, intermediate level, and high level. In this case, the sending of the corresponding prompt information according to the danger level, and/or the executing of the operation corresponding to the danger level and the controlling of the vehicle according to the operation include the followings.
  • The prompt information is sent in response to that the danger level is the primary level.
  • The operation corresponding to the danger level is executed and the vehicle is controlled according to the operation in response to that the danger level is the intermediate level.
  • The operation corresponding to the danger level is executed and the vehicle is controlled according to the operation while sending the prompt information in response to that the danger level is the high level.
  • In the embodiments of the present disclosure, the danger level is set as 3 levels. Optionally, the embodiments of the present disclosure may further set the danger level more in details, and more levels are included, for example, the danger level includes a first level, a second level, a third level, and a fourth level, where each level corresponds to different danger degrees. The prompt information is sent according to different danger levels, and/or the operation corresponding to the danger level is executed and the vehicle is controlled according to the operation. By executing different operations for different danger levels, the sending of the prompt information and the controlling of the operation may be more flexible and adapt to different use requirements.
  • Optionally, the determining of the danger level of the predetermined dangerous action includes the followings.
  • The frequency and/or duration of occurrence of the predetermined dangerous action in the video stream are acquired, and the danger level of the predetermined dangerous action is determined on the basis of the frequency and/or duration.
  • In the embodiments of the present disclosure, by performing further abstract analysis on the dangerous action obtained by action recognition, and according to a lasting degree of the action or a priori probability of the occurrence of a dangerous situation, whether a real intention of a passenger performs a dangerous action is output. Optionally, the embodiments of the present disclosure implement measurement of the action lasting degree by means of the frequency and/or duration of the occurrence of the predetermined dangerous action in the video stream. For example, when the driver just scratches the eye quickly, it is considered as just a quick adjustment, and alarming is not required. However, if the driver rubs the eye for a long time along with the occurrence of the action such as yawning, it is considered that the driver is relatively fatigue and should be prompted. For another example, the alarming strength for smoking may be less than that for actions such as bending over the steering wheel and calling.
  • In a possible implementation, the action includes a duration of an action, an early-warning condition includes: recognizing that the duration of the action exceeds a duration threshold.
  • In a possible implementation, the action may include the duration of the action; when the duration of the action exceeds the duration threshold, it is considered that the execution of the action distracts much of the attention of an action execution object; said action is considered as a dangerous action, and early-warning information needs to be sent. For example, if the duration of the smoking action of the driver exceeds 3 seconds, it is considered that the smoking action is the dangerous action and influences a driving action of the driver, and the early-warning information needs to be sent to the driver.
  • In the embodiments, according to the duration and the duration threshold of the predetermined dangerous action, a sending condition of the prompt information and/or a control condition of the vehicle are adjusted, so that the sending of the prompt information and the controlling of the operation are more flexible and better adapt to different use requirements.
  • In a possible implementation, the action recognition result includes a duration of an action, and a condition of belonging to the predetermined dangerous action includes: recognizing that the duration of the action exceeds the duration threshold. Some actions do not cause potential safety hazards to the in-vehicle personnel and the vehicle in a short time; only when the duration of the action reaches a set duration threshold, the action is confirmed as the predetermined dangerous action, for example, an action of closing eyes of the driver is considered as a normal blink when the duration of closing eyes is short (for example, 0.5 second), and when the duration of closing eyes exceeds the duration threshold (being set according to the requirements, for example, setting to be 3 seconds), said action is considered to belong to the predetermined dangerous action, and the corresponding prompt information is sent.
  • In a possible implementation, the action recognition result includes the number of times for which an action is performed, and a condition of belonging to the predetermined dangerous action includes: recognizing that the number of times exceeds a number threshold. When the number of times exceeds the number threshold, it is considered that the action of the action execution object is frequent, and much attention is distracted; the action is considered as the dangerous action, and the early-warning information needs to be sent. For example, if the number of the smoking actions of the driver exceeds 5 times, it is considered that the smoking action is the dangerous action and influences the driving action of the driver, and the prompt information needs to be sent to the driver.
  • In a possible implementation, the action recognition result includes a duration of an action and the number of times for which the action is performed, and the condition of belonging to the predetermined dangerous action includes: recognizing that the duration of the action exceeds the duration threshold, and the number of times exceeds the number threshold.
  • In a possible implementation, when the duration of the action exceeds the duration threshold, and the number of times exceeds the number threshold, it is considered that the action of the action execution object is frequent and the duration of the action is long, and much attention is distracted; said action is considered as the dangerous action, and the prompt information needs to be sent and/or the vehicle is controlled, so that the vehicle is more flexibly controlled and adapts to different use requirements.
  • The dangerous actions corresponding to different in-vehicle personnel are different, for example, the driver at a driving seat is required not to be distracted, and the distraction action belongs to the dangerous action; while the distraction action of the in-vehicle personnel at other positions does not belong to the dangerous action. Therefore, in the embodiments of present disclosure, in order to implement more accurate alarming and intelligent control, prompting or an intelligent operation is performed by combining the action category and a category of the in-vehicle personnel, so that user experience is not influenced because of frequently alarming while improving the driving safety. Optionally, the embodiments of the present disclosure include: sending corresponding first prompt information and/or controlling the vehicle to execute a corresponding first predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the driver; and/or,
  • sending corresponding second prompt information and/or executing a corresponding second predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the non-driver.
  • Because the driver is responsible for the safety of the whole vehicle, in order to improve the driving safety of the vehicle and the freedom of the passenger, the in-vehicle personnel are divided into two categories, i. e., the driver and the non-driver. Different dangerous actions are respectively set for the driver and the non-driver, so as to implement flexible alarming and operation. Optionally, the distraction action of the driver may include, but is not limited to at least one of the followings: calling, drinking water, putting on or taking off sunglasses, putting on or taking off the mask, or eating food and the like. The discomfort state of the driver may include, but is not limited to at least one of the followings: sweat wiping, rubbing an eye, or yawning and the like. The non-standard behavior of the driver may include, but is not limited to at least one of the followings: smoking, stretching the hand out of the vehicle, bending over the steering wheel, putting both feet on the steering wheel, or leaving both hands away from the steering wheel and the like.
  • Optionally, the discomfort state of the non-driver may include, but is not limited to at least one of the followings: sweat wiping or the like. The non-standard behavior of the non-driver may include, but is not limited to at least one of the followings: smoking, stretching the hand out of the vehicle, holding the instrument with the hand, or disturbing the driver or the like.
  • In the embodiments of the present disclosure, different prompt information and predetermined operations are further respectively set for the driver and the non-driver, so as to implement flexible safety control of the vehicle, for example, when the driver makes an action of leaving both hands from the steering wheel, automatically driving (for example, the corresponding first predetermined operation) is executed while sending strong prompt information (for example, the corresponding first prompt information), so as to improve the safety of vehicle driving; moreover, for the non-driver, for example, when the non-driver makes an action of sweat wiping, weak prompt information (for example, the corresponding second prompt information) is sent, and/or an operation of adjusting an in-vehicle air conditioning temperature (for example, the corresponding second predetermined operation) is executed.
  • The in-vehicle personnel of ordinary skill in the art may understand that: all or some steps of implementing the forgoing embodiments of the method may be achieved by a program by instructing related hardware; the foregoing program may be stored in a computer readable storage medium; when the program is executed, steps including the foregoing embodiments of the method are performed; moreover, the foregoing storage medium includes various media capable of storing program codes such as an ROM, an RAM, a magnetic disk, or an optical disk.
  • FIG. 4 is a schematic structural diagram of a device for recognizing a dangerous action of personnel in a vehicle (i.e., in-vehicle personnel) provided in the embodiments of the present disclosure. The apparatus of this embodiment is configured to implement the foregoing method embodiments of the present disclosure. As shown in FIG. 4, the apparatus of this embodiment includes:
  • a video collection unit 41, used for obtaining at least one video stream of the in-vehicle personnel through an image capturing device,
  • each video stream including information about at least one in-vehicle personnel;
  • an action recognition unit 42, used for performing action recognition on the in-vehicle personnel on the basis of the video stream; and
  • a danger processing unit 43, used for sending prompt information and/or executing an operation to control a vehicle in response to that an action recognition result belongs to a predetermined dangerous action,
  • where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: a distraction action, a discomfort state, or a non-standard behavior and the like.
  • On the basis of the in-vehicle personnel dangerous action recognition apparatus provided in the foregoing embodiments of the present disclosure, whether the predetermined dangerous action is made is determined by means of action recognition, and corresponding prompt/or operation is made to the predetermined dangerous action to control the vehicle, so that early detection of vehicle safety conditions is implemented to reduce the probability of dangerous situations.
  • In one or more optional embodiments, the action recognition unit 42 is used for detecting at least one target area included by the in-vehicle personnel in at least one frame of video image of the video stream, capturing a target image corresponding to the target area from the at least one frame of video image of the video stream according to the target area obtained through detection, and performing action recognition on the in-vehicle personnel according to the target image.
  • In the embodiments, the target area is recognized in the video stream, the target image corresponding to the target area is captured in the video image according to the detection result of the target area, and whether the in-vehicle personnel executes the predetermined dangerous action is recognized according to the target image. The target image captured according to the detection result of the target area may be applicable to human bodies having different areas in different video images. The embodiments of the present disclosure have a wide application range.
  • Optionally, the action recognition unit 42 is used for extracting a feature of the in-vehicle personnel included in the at least one frame of video image of the video stream when detecting at least one target area included by the in-vehicle personnel in the at least one frame of video image of the video frame, and extracting a target area from the at least one frame of video image frame on the basis of the feature, where the target area includes, but is not limited to, at least one of the followings: the face local area, the action interactive object, or the limb area or the like.
  • Optionally, the face local area includes, but is not limited to, at least one of the followings: the mouth area, the ear area, or the eye area or the like.
  • Optionally, the action interactive object includes, but is not limited to, at least one of the followings: a container, a cigarette, a mobile phone, food, a tool, a beverage bottle, glasses, or a mask or the like.
  • In one or more optional embodiments, the distraction action includes, but is not limited to, at least one of the followings: calling, drinking water, putting on or taking off sunglasses, putting on or taking off the mask, or eating food or the like; and/or,
  • the discomfort state includes, but is not limited to, at least one of the followings: wiping sweat, rubbing an eye, or yawning; and/or,
  • the non-standard behavior includes, but is not limited to, at least one of the followings: smoking, stretching a hand out of the vehicle, bending over a steering wheel, putting both feet on the steering wheel, leaving both hands away from the steering wheel, holding an instrument with a hand, or disturbing a driver or the like.
  • In one or more optional embodiments, the danger processing unit 43 includes:
  • a level determination module, used for determining a danger level of the predetermined dangerous action in response to that the action recognition result belongs to the predetermined dangerous action; and
  • an operation processing module, used for sending corresponding prompt information according to the danger level, and/or executing an operation corresponding to the danger level and controlling the vehicle according to the operation.
  • Optionally, in the embodiments of the present disclosure, when the action of the in-vehicle personnel is determined to belong to the predetermined dangerous action according to the action recognition result, danger level determination is performed on the predetermined dangerous action; optionally, the danger level of the predetermined dangerous action is determined according to a preset rule and/or correspondence, and then how to operate is determined according to the danger level. For example, operations of different degrees are performed according to the dangerous action level of the in-vehicle personnel. For example, if the dangerous action is caused by fatigue and physical discomfort of the driver, timely prompting is required, so that the driver performs adjustment and has a rest in time; and when the driver feels discomfort due to the in-vehicle environment, an adjustment of a certain degree is performed by controlling a ventilation system or an air conditioning system in the vehicle.
  • Optionally, the danger level includes a primary level, an intermediate level, and a high level.
  • The operation processing module is used for sending the prompt information in response to that the danger level is the primary level, executing the operation corresponding to the danger level and controlling the vehicle according to the operation in response to that the danger level is the intermediate level, and executing the operation corresponding to the danger level and controlling the vehicle according to the operation while sending the prompt information in response to that the danger level is the high level.
  • Optionally, the level determination module is used for acquiring the frequency and/or duration of occurrence of the predetermined dangerous action in the video stream, and determining the danger level of the predetermined dangerous action on the basis of the frequency and/or duration.
  • In one or more optional embodiments, the action recognition result includes the action duration, and the condition of belonging to the predetermined dangerous action includes: recognizing that the action duration exceeds the duration threshold.
  • In the embodiments of the present disclosure, by performing further abstract analysis on the dangerous action obtained by action recognition, and according to the lasting degree of the action or the priori probability of the occurrence of the dangerous situation, whether the real intention of the passenger performs the dangerous action is output. Optionally, the embodiments of the present disclosure implement measurement of the action lasting degree by means of the frequency and/or duration of the occurrence of the predetermined dangerous action in the video stream.
  • Optionally, the action recognition result includes the action duration, and the condition of belonging to the predetermined dangerous action includes: recognizing that the action duration exceeds the duration threshold.
  • Optionally, the action recognition result includes the number of actions, and the condition of belonging to the predetermined dangerous action includes: recognizing that the number of actions exceeds the number threshold.
  • Optionally, the action recognition result includes the action duration and the number of actions, and the condition of belonging to the predetermined dangerous action includes: recognizing that the action duration exceeds the duration threshold, and the number of actions exceeds the number threshold.
  • Optionally, the in-vehicle personnel includes the driver and/or the non-driver of the vehicle.
  • Optionally, the danger processing unit 43 is used for sending corresponding first prompt information and/or controlling the vehicle to execute a corresponding first predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the driver, and/or sending corresponding second prompt information and/or executing a corresponding second predetermined operation according to the predetermined dangerous action in response to that the in-vehicle personnel is the non-driver.
  • For the working process, the setting mode, and corresponding technical effect of any embodiment of the in-vehicle personnel dangerous action recognition apparatus provided by the embodiments of the present disclosure, reference may be made to the specific descriptions of the corresponding method embodiments of the present disclosure, and details are not described herein again due to space limitation.
  • According to still another aspect of the embodiments of the present disclosure, an electronic device is provided and includes a processor, where the processor includes the in-vehicle personnel dangerous action recognition apparatus provided according to any one of the foregoing embodiments.
  • According to yet another aspect of the embodiments of the present disclosure, an electronic device is provided and includes: a memory used for storing executable instructions;
  • and a processor, used for communicating with the memory to execute the executable instructions so as to complete operations of the method for recognizing the dangerous actin of the personnel in the vehicle provided according to any one of the foregoing embodiments.
  • According to a further aspect of the embodiments of the present application, a computer readable storage medium is provided and is used for storing computer readable instructions, where when the instructions are executed, the operations of the method for recognizing the dangerous actin of the personnel in the vehicle provided according to any one of the foregoing embodiments are executed.
  • According to still another aspect of the embodiments of the present disclosure, a computer program is provided and includes a computer readable code, where if the computer readable code runs on a device, a processor in the device executes an instruction for implementing the method for recognizing the dangerous actin of the personnel in the vehicle provided according to any one of the foregoing embodiments.
  • The embodiments of the present disclosure further provide an electronic device which, for example, is a mobile terminal, a Personal Computer (PC), a tablet computer, a server, and the like. Referring to FIG. 5 below, a schematic structural diagram of an electronic device 500, which is a terminal device or a server, suitable for implementing the embodiments of the present disclosure is shown. As shown in FIG. 5, the electronic device 500 includes one or more processors, a communication part, and the like. The one or more processors are, for example, one or more Central Processing Units (CPUs) 501 and/or one or more graphic processing unit (an acceleration unit) 513, and may execute appropriate actions and processing according to executable instructions stored in a Read-Only Memory (ROM) 502 or executable instructions loaded from a storage section 508 to a Random Access Memory (RAM) 503. The communication part 512 may include, but is not limited to, a network card. The network card may include, but is not limited to, an Infiniband (IB) network card.
  • The processor communicates with the ROM 502 and/or the RAM 503 to execute the executable instructions, and is connected to the communication part 512 by means of a bus 504 and communicates with other target devices by means of the communication part 512, so as to complete the operations corresponding to any of the methods provided in the embodiments of the present disclosure, for example, obtaining at least one video stream of in-vehicle personnel by using a photographing apparatus, each video stream including information about at least one in-vehicle personnel; performing action recognition on the in-vehicle personnel on the basis of the video stream; and sending prompt information and/or executing an operation to control a vehicle in response to that an action recognition result belongs to a predetermined dangerous action, where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: a distraction action, a discomfort state, or a non-standard behavior.
  • In addition, the RAM 503 may further store various programs and data required for operations of an apparatus. The CPU 501, the ROM 502, and the RAM 503 are connected to each other by means of the bus 504. In the presence of the RAM 503, the ROM 502 is an optional module. The RAM 503 stores executable instructions, or writes the executable instructions into the ROM 502 during running, where the executable instructions cause the CPU 501 to execute corresponding operations of the foregoing communication method. An Input/Output (I/O) interface 505 is also connected to the bus 504. The communication part 512 is integrated, or is also configured to have multiple sub-modules (for example, multiple IB network cards) linked on the bus.
  • The following components are connected to the I/O interface 505: an input section 506 including a keyboard, a mouse and the like; an output section 507 including a Cathode-Ray Tube (CRT), a Liquid Crystal Display (LCD), a speaker and the like; the storage section 508 including a hard disk drive and the like; and a communication section 509 of a network interface card including an LAN card, a modem and the like. The communication section 509 performs communication processing via a network such as the Internet. A drive 510 is also connected to the I/O interface 505 according to requirements. A removable medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like is installed on the drive 510 according to requirements, so that a computer program read from the removable medium is installed on the storage section 508 according to requirements.
  • It should be noted that, the architecture shown in FIG. 5 is merely an optional implementation. During specific practice, the number and types of the components in FIG. 5 may be selected, decreased, increased, or replaced according to actual requirements. Different functional components may be configured separately or integrally or the like. For example, an acceleration unit 513 and the CPU 501 may be configured separately, or the acceleration unit 513 may be integrated on the CPU 501, and the communication part may be configured separately, and may also be configured integrally on the CPU 501 or the acceleration unit 513 or the like. These alternative implementations all fall within the scope of protection of the present disclosure.
  • Particularly, a process described above with reference to a flowchart according to the embodiments of the present disclosure is implemented as a computer software program. For example, the embodiments of the disclosure include a computer program product. The computer program product includes a computer program tangibly included in a machine-readable medium. The computer program includes a program code for performing a method shown in the flowchart. The program code may include instructions for correspondingly performing steps of the method provided in the embodiments of the present disclosure, for example, obtaining at least one video stream of in-vehicle personnel by using the photographing apparatus, each video stream including information about at least one in-vehicle personnel; performing action recognition on the in-vehicle personnel on the basis of the video stream; and sending the prompt information and/or executing the operation to control the vehicle in response to that the action recognition result belongs to the predetermined dangerous action, where the predetermined dangerous action includes at least one of the following action representations of the in-vehicle personnel: the distraction action, the discomfort state, or the non-standard behavior. In such embodiments, the computer program is downloaded and installed from the network by means of the communication section 509, and/or is installed from the removable medium 511. The computer program, when being executed by the CPU 501, executes the operations of the foregoing functions defined in the method of the present disclosure.
  • The embodiments in the specification are all described in a progressive manner, for same or similar parts in the embodiments, refer to these embodiments, and each embodiment focuses on a difference from other embodiments. The system embodiments correspond to the method embodiments substantially and therefore are only described briefly, and for the associated part, refer to the descriptions of the method embodiments.
  • The methods and apparatuses of the present disclosure are implemented in many manners. For example, the methods and apparatuses of the present disclosure are implemented by means of software, hardware, firmware, or any combination of software, hardware, and firmware. Unless otherwise specially stated, the foregoing sequences of steps of the methods are merely for description, and are not intended to limit the steps of the methods of the present disclosure. In addition, in some embodiments, the present disclosure may also be implemented as programs recorded in a recording medium. The programs include machine-readable instructions for implementing the methods according to the present disclosure. Therefore, the present disclosure further covers the recording medium storing the programs for executing the methods according to the present disclosure.
  • The descriptions of the present disclosure are provided for the purpose of examples and description, and are not intended to be exhaustive or limit the present disclosure to the disclosed form. Many modifications and changes are obvious to a person of ordinary skill in the art. The embodiments are selected and described to better describe a principle and an actual application of the present disclosure, and to make a person of ordinary skill in the art understand the present disclosure, so as to design various embodiments with various modifications applicable to particular use.

Claims (20)

1. A method for recognizing a dangerous action of personnel in a vehicle, comprising:
obtaining at least one video stream of the personnel in the vehicle through an image capturing device, each video stream comprising information about at least one of the personnel in the vehicle;
performing action recognition on the personnel in the vehicle based on the video stream; and
responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, wherein the predetermined dangerous action comprises at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
2. The method according to claim 1, wherein performing the action recognition on the personnel in the vehicle based on the video stream comprises:
detecting at least one target area comprised by the personnel in the vehicle, in at least one frame of video image of the video stream;
capturing a target image corresponding to the target area from the at least one frame of video image of the video stream according to the target area obtained through detection; and
performing action recognition on the personnel in the vehicle according to the target image.
3. The method according to claim 2, wherein detecting the at least one target area comprised by the personnel in the vehicle, in the at least one frame of video image of the video stream comprises:
extracting a feature, comprised in the at least one frame of video image of the video stream, of the personnel in the vehicle; and
extracting a target area from the at least one frame of video image based on the feature, wherein the target area comprises at least one of: a face local area, an action interactive object, or a limb area.
4. The method according to claim 3, wherein the face local area comprises at least one of: a mouth area, an ear area, or an eye area.
5. The method according to claim 3, wherein the action interactive object comprises at least one of: a container, a cigarette, a mobile phone, food, a tool, a beverage bottle, glasses, or a mask.
6. The method according to claim 1, wherein the distraction action comprises at least one of: calling, drinking water, putting on or taking off sunglasses, putting on or taking off a mask, or eating food;
the discomfort state comprises at least one of: wiping sweat, rubbing an eye, or yawning;
the non-standard behavior comprises at least one of: smoking, stretching a hand out of the vehicle, bending over a steering wheel, putting both feet on the steering wheel, leaving both hands away from the steering wheel, holding an instrument with a hand, or disturbing a driver.
7. The method according to claim 1, wherein responsive to that the result of the action recognition belongs to the predetermined dangerous action, performing the at least one of: sending the prompt information, or executing the operation to control the vehicle comprises:
responsive to that the result of the action recognition belongs to the predetermined dangerous action;
determining a danger level of the predetermined dangerous action; and
performing at least one of: sending corresponding prompt information according to the danger level, or executing an operation corresponding to the danger level and controlling the vehicle according to the operation.
8. The method according to claim 7, wherein the danger level comprises a primary level, an intermediate level, and a high level;
wherein performing the at least one of: sending the corresponding prompt information according to the danger level, or executing the operation corresponding to the danger level and controlling the vehicle according to the operation comprises:
sending the prompt information responsive to that the danger level is the primary level;
executing the operation corresponding to the danger level and controlling the vehicle according to the operation, responsive to that the danger level is the intermediate level; and
executing the operation corresponding to the danger level and controlling the vehicle according to the operation while sending the prompt information, responsive to that the danger level is the high level.
9. The method according to claim 7, wherein determining the danger level of the predetermined dangerous action comprises:
acquiring at least one of a frequency or a duration of occurrence of the predetermined dangerous action in the video stream, and determining the danger level of the predetermined dangerous action based on the at least one of the frequency or the duration.
10. The method according to claim 1, wherein the result of the action recognition comprises a duration of an action, and a condition of belonging to the predetermined dangerous action comprises: recognizing that the duration of the action exceeds a duration threshold.
11. The method according to claim 1, wherein the result of the action recognition comprises a number of times for which an action is performed, and a condition of belonging to the predetermined dangerous action comprises: recognizing that the number of times exceeds a number threshold.
12. The method according to claim 1, wherein the result of the action recognition comprises a duration of an action and a number of times for which the action is performed, and a condition of belonging to the predetermined dangerous action comprises: recognizing that the duration of the action exceeds a duration threshold, and the number of times exceeds a number threshold.
13. The method according to claim 1, wherein the personnel in the vehicle comprises at least one of a driver or a non-driver of the vehicle.
14. The method according to claim 13, wherein responsive to that the result of the action recognition belongs to the predetermined dangerous action, performing the at least one of: sending the prompt information, or executing the operation to control the vehicle comprises at least one of:
responsive to that the personnel in the vehicle is the driver, performing at least one of: sending corresponding first prompt information according to the predetermined dangerous action, or controlling the vehicle to execute a corresponding first predetermined operation according to the predetermined dangerous action; or
responsive to that the personnel in the vehicle is the non-driver, performing at least one of: sending corresponding second prompt information according to the predetermined dangerous action, or executing a corresponding second predetermined operation according to the predetermined dangerous action.
15. An electronic device, comprising:
a processor; and
a memory configured to store instructions that, when executed by the processor, cause the processor to perform the following operations comprising:
obtaining at least one video stream of personnel in a vehicle through an image capturing device, each video stream comprising information about at least one of the personnel in the vehicle;
performing action recognition on the personnel in the vehicle based on the video stream; and
responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, wherein the predetermined dangerous action comprises at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
16. The device according to claim 15, wherein the processor is configured to: detect at least one target area comprised by the personnel in the vehicle in at least one frame of video image of the video stream, capture a target image corresponding to the target area from the at least one frame of video image of the video stream according to the target area obtained through detection, and perform action recognition on the personnel in the vehicle according to the target image.
17. The device according to claim 16, wherein the processor is configured to: extract a feature, comprised in the at least one frame of video image of the video stream, of the personnel in the vehicle when detecting the at least one target area comprised by the personnel in the vehicle in the at least one frame of video image of the video stream, and extract a target area from the at least one frame of video image based on the feature, wherein the target area comprises at least one of: a face local area, an action interactive object, or a limb area.
18. The device according to claim 15, wherein the processor is configured to:
determine a danger level of the predetermined dangerous action responsive to that the result of the action recognition belongs to the predetermined dangerous action; and
perform at least one of: sending corresponding prompt information according to the danger level, or executing an operation corresponding to the danger level and controlling the vehicle according to the operation.
19. The device according to claim 18, wherein the danger level comprises a primary level, an intermediate level, and a high level; and
the processor is configured to:
send prompt information responsive to that the danger level is the primary level;
execute the operation corresponding to the danger level and control the vehicle according to the operation, responsive to that the danger level is the intermediate level; and
execute the operation corresponding to the danger level and control the vehicle according to the operation while sending the prompt information, responsive to that the danger level is the high level.
20. A non-transitory computer readable storage medium configured to store computer readable instructions that, when executed by a processor of an electronic device, cause the processor to perform a method for method for recognizing a dangerous action of personnel in a vehicle, comprising:
obtaining at least one video stream of the personnel in the vehicle through an image capturing device, each video stream comprising information about at least one of the personnel in the vehicle;
performing action recognition on the personnel in the vehicle based on the video stream; and
responsive to that a result of the action recognition belongs to a predetermined dangerous action, performing at least one of: sending prompt information, or executing an operation to control the vehicle, wherein the predetermined dangerous action comprises at least one of the following action representations of the personnel in the vehicle: a distraction action, a discomfort state, or a non-standard behavior.
US17/034,290 2017-08-10 2020-09-28 Method for recognizing dangerous action of personnel in vehicle, electronic device and storage medium Abandoned US20210009150A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
PCT/CN2017/096957 WO2019028798A1 (en) 2017-08-10 2017-08-10 Method and device for monitoring driving condition, and electronic device
CN201910152525.X 2019-02-28
CN201910152525.XA CN110399767A (en) 2017-08-10 2019-02-28 Occupant's dangerous play recognition methods and device, electronic equipment, storage medium
PCT/CN2019/129370 WO2020173213A1 (en) 2017-08-10 2019-12-27 Method and apparatus for identifying dangerous actions of persons in vehicle, electronic device and storage medium

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/129370 Continuation WO2020173213A1 (en) 2017-08-10 2019-12-27 Method and apparatus for identifying dangerous actions of persons in vehicle, electronic device and storage medium

Publications (1)

Publication Number Publication Date
US20210009150A1 true US20210009150A1 (en) 2021-01-14

Family

ID=65273075

Family Applications (5)

Application Number Title Priority Date Filing Date
US16/177,198 Active 2037-11-29 US10853675B2 (en) 2017-08-10 2018-10-31 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles
US17/034,290 Abandoned US20210009150A1 (en) 2017-08-10 2020-09-28 Method for recognizing dangerous action of personnel in vehicle, electronic device and storage medium
US17/085,972 Abandoned US20210049387A1 (en) 2017-08-10 2020-10-30 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles
US17/085,953 Abandoned US20210049386A1 (en) 2017-08-10 2020-10-30 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles
US17/085,989 Abandoned US20210049388A1 (en) 2017-08-10 2020-10-30 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/177,198 Active 2037-11-29 US10853675B2 (en) 2017-08-10 2018-10-31 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles

Family Applications After (3)

Application Number Title Priority Date Filing Date
US17/085,972 Abandoned US20210049387A1 (en) 2017-08-10 2020-10-30 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles
US17/085,953 Abandoned US20210049386A1 (en) 2017-08-10 2020-10-30 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles
US17/085,989 Abandoned US20210049388A1 (en) 2017-08-10 2020-10-30 Driving state monitoring methods and apparatuses, driver monitoring systems, and vehicles

Country Status (8)

Country Link
US (5) US10853675B2 (en)
EP (1) EP3666577A4 (en)
JP (2) JP6933668B2 (en)
KR (2) KR102391279B1 (en)
CN (3) CN109803583A (en)
SG (2) SG11202002549WA (en)
TW (1) TWI758689B (en)
WO (3) WO2019028798A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11037348B2 (en) * 2016-08-19 2021-06-15 Beijing Sensetime Technology Development Co., Ltd Method and apparatus for displaying business object in video image and electronic device
US20220230474A1 (en) * 2019-05-08 2022-07-21 Jaguar Land Rover Limited Activity identification method and apparatus
US20220265881A1 (en) * 2021-02-25 2022-08-25 Toyota Jidosha Kabushiki Kaisha Taxi vehicle and taxi system

Families Citing this family (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019028798A1 (en) * 2017-08-10 2019-02-14 北京市商汤科技开发有限公司 Method and device for monitoring driving condition, and electronic device
JP6888542B2 (en) * 2017-12-22 2021-06-16 トヨタ自動車株式会社 Drowsiness estimation device and drowsiness estimation method
US10850746B2 (en) * 2018-07-24 2020-12-01 Harman International Industries, Incorporated Coordinating delivery of notifications to the driver of a vehicle to reduce distractions
JP6906717B2 (en) * 2018-12-12 2021-07-21 三菱電機株式会社 Status determination device, status determination method, and status determination program
US11170240B2 (en) * 2019-01-04 2021-11-09 Cerence Operating Company Interaction system and method
US10657396B1 (en) * 2019-01-30 2020-05-19 StradVision, Inc. Method and device for estimating passenger statuses in 2 dimension image shot by using 2 dimension camera with fisheye lens
CN111626087A (en) * 2019-02-28 2020-09-04 北京市商汤科技开发有限公司 Neural network training and eye opening and closing state detection method, device and equipment
CN111661059B (en) * 2019-03-08 2022-07-08 虹软科技股份有限公司 Method and system for monitoring distracted driving and electronic equipment
CN110001652B (en) * 2019-03-26 2020-06-23 深圳市科思创动科技有限公司 Driver state monitoring method and device and terminal equipment
CN111845749A (en) * 2019-04-28 2020-10-30 郑州宇通客车股份有限公司 Control method and system for automatically driving vehicle
CN109977930B (en) * 2019-04-29 2021-04-02 中国电子信息产业集团有限公司第六研究所 Fatigue driving detection method and device
CN110263641A (en) * 2019-05-17 2019-09-20 成都旷视金智科技有限公司 Fatigue detection method, device and readable storage medium storing program for executing
US11281920B1 (en) * 2019-05-23 2022-03-22 State Farm Mutual Automobile Insurance Company Apparatuses, systems and methods for generating a vehicle driver signature
CN110188655A (en) * 2019-05-27 2019-08-30 上海蔚来汽车有限公司 Driving condition evaluation method, system and computer storage medium
WO2020237664A1 (en) * 2019-05-31 2020-12-03 驭势(上海)汽车科技有限公司 Driving prompt method, driving state detection method and computing device
CN112241645A (en) * 2019-07-16 2021-01-19 广州汽车集团股份有限公司 Fatigue driving detection method and system and electronic equipment
JP7047821B2 (en) * 2019-07-18 2022-04-05 トヨタ自動車株式会社 Driving support device
US10991130B2 (en) * 2019-07-29 2021-04-27 Verizon Patent And Licensing Inc. Systems and methods for implementing a sensor based real time tracking system
FR3100640B1 (en) * 2019-09-10 2021-08-06 Faurecia Interieur Ind Method and device for detecting yawns of a driver of a vehicle
CN112758098B (en) * 2019-11-01 2022-07-22 广州汽车集团股份有限公司 Vehicle driving authority take-over control method and device based on driver state grade
CN110942591B (en) * 2019-11-12 2022-06-24 博泰车联网科技(上海)股份有限公司 Driving safety reminding system and method
CN110826521A (en) * 2019-11-15 2020-02-21 爱驰汽车有限公司 Driver fatigue state recognition method, system, electronic device, and storage medium
CN110837815A (en) * 2019-11-15 2020-02-25 济宁学院 Driver state monitoring method based on convolutional neural network
CN110968718B (en) * 2019-11-19 2023-07-14 北京百度网讯科技有限公司 Target detection model negative sample mining method and device and electronic equipment
CN110909715B (en) * 2019-12-06 2023-08-04 重庆商勤科技有限公司 Method, device, server and storage medium for identifying smoking based on video image
CN111160126B (en) * 2019-12-11 2023-12-19 深圳市锐明技术股份有限公司 Driving state determining method, driving state determining device, vehicle and storage medium
JP2021096530A (en) * 2019-12-13 2021-06-24 トヨタ自動車株式会社 Operation support device, operation support program, and operation support system
CN111160237A (en) * 2019-12-27 2020-05-15 智车优行科技(北京)有限公司 Head pose estimation method and apparatus, electronic device, and storage medium
CN111191573A (en) * 2019-12-27 2020-05-22 中国电子科技集团公司第十五研究所 Driver fatigue detection method based on blink rule recognition
CN113128295A (en) * 2019-12-31 2021-07-16 湖北亿咖通科技有限公司 Method and device for identifying dangerous driving state of vehicle driver
CN113126296B (en) * 2020-01-15 2023-04-07 未来(北京)黑科技有限公司 Head-up display equipment capable of improving light utilization rate
CN111243236A (en) * 2020-01-17 2020-06-05 南京邮电大学 Fatigue driving early warning method and system based on deep learning
US11873000B2 (en) 2020-02-18 2024-01-16 Toyota Motor North America, Inc. Gesture detection for transport control
EP4099293B1 (en) * 2020-02-25 2024-04-24 Huawei Technologies Co., Ltd. Method and apparatus for identifying special road conditions, electronic device, and storage medium
JP7402084B2 (en) * 2020-03-05 2023-12-20 本田技研工業株式会社 Occupant behavior determination device
US11912307B2 (en) 2020-03-18 2024-02-27 Waymo Llc Monitoring head movements of drivers tasked with monitoring a vehicle operating in an autonomous driving mode
CN111783515A (en) * 2020-03-18 2020-10-16 北京沃东天骏信息技术有限公司 Behavior action recognition method and device
CN111460950B (en) * 2020-03-25 2023-04-18 西安工业大学 Cognitive distraction method based on head-eye evidence fusion in natural driving conversation behavior
JP7380380B2 (en) * 2020-03-26 2023-11-15 いすゞ自動車株式会社 Driving support device
CN111626101A (en) * 2020-04-13 2020-09-04 惠州市德赛西威汽车电子股份有限公司 Smoking monitoring method and system based on ADAS
WO2021240671A1 (en) * 2020-05-27 2021-12-02 三菱電機株式会社 Gesture detection device and gesture detection method
WO2021240668A1 (en) * 2020-05-27 2021-12-02 三菱電機株式会社 Gesture detection device and gesture detection method
CN111611970B (en) * 2020-06-01 2023-08-22 城云科技(中国)有限公司 Urban management monitoring video-based random garbage throwing behavior detection method
CN111652128B (en) * 2020-06-02 2023-09-01 浙江大华技术股份有限公司 High-altitude power operation safety monitoring method, system and storage device
CN111767823A (en) * 2020-06-23 2020-10-13 京东数字科技控股有限公司 Sleeping post detection method, device, system and storage medium
JP7359087B2 (en) * 2020-07-02 2023-10-11 トヨタ自動車株式会社 Driver monitoring device and driver monitoring method
CN111785008A (en) * 2020-07-04 2020-10-16 苏州信泰中运物流有限公司 Logistics monitoring management method and device based on GPS and Beidou positioning and computer readable storage medium
CN113920576A (en) * 2020-07-07 2022-01-11 奥迪股份公司 Method, device, equipment and storage medium for identifying object loss behavior of personnel on vehicle
US20220414796A1 (en) * 2020-07-08 2022-12-29 Pilot Travel Centers, LLC Computer implemented oil field logistics
CN111797784B (en) * 2020-07-09 2024-03-05 斑马网络技术有限公司 Driving behavior monitoring method and device, electronic equipment and storage medium
US11776319B2 (en) * 2020-07-14 2023-10-03 Fotonation Limited Methods and systems to predict activity in a sequence of images
CN111860280A (en) * 2020-07-15 2020-10-30 南通大学 Deep learning-based driver violation behavior recognition system
CN111832526A (en) * 2020-07-23 2020-10-27 浙江蓝卓工业互联网信息技术有限公司 Behavior detection method and device
CN112061065B (en) * 2020-07-27 2022-05-10 大众问问(北京)信息科技有限公司 In-vehicle behavior recognition alarm method, device, electronic device and storage medium
US11651599B2 (en) * 2020-08-17 2023-05-16 Verizon Patent And Licensing Inc. Systems and methods for identifying distracted driver behavior from video
CN112069931A (en) * 2020-08-20 2020-12-11 深圳数联天下智能科技有限公司 State report generation method and state monitoring system
CN112016457A (en) * 2020-08-27 2020-12-01 青岛慕容信息科技有限公司 Driver distraction and dangerous driving behavior recognition method, device and storage medium
CN112084919A (en) * 2020-08-31 2020-12-15 广州小鹏汽车科技有限公司 Target detection method, target detection device, vehicle and storage medium
CN114201985A (en) * 2020-08-31 2022-03-18 魔门塔(苏州)科技有限公司 Method and device for detecting key points of human body
CN112163470A (en) * 2020-09-11 2021-01-01 高新兴科技集团股份有限公司 Fatigue state identification method, system and storage medium based on deep learning
CN112307920B (en) * 2020-10-22 2024-03-22 东云睿连(武汉)计算技术有限公司 High-risk worker behavior early warning device and method
CN112149641A (en) * 2020-10-23 2020-12-29 北京百度网讯科技有限公司 Method, device, equipment and storage medium for monitoring driving state
CN112347891B (en) * 2020-10-30 2022-02-22 南京佑驾科技有限公司 Method for detecting drinking water state in cabin based on vision
CN112270283A (en) * 2020-11-04 2021-01-26 北京百度网讯科技有限公司 Abnormal driving behavior determination method, device, equipment, vehicle and medium
CN112356839A (en) * 2020-11-06 2021-02-12 广州小鹏自动驾驶科技有限公司 Driving state monitoring method and system and automobile
JP2022077282A (en) * 2020-11-11 2022-05-23 株式会社コムテック Warning system
KR102443980B1 (en) * 2020-11-17 2022-09-21 주식회사 아르비존 Vehicle control method
TWI739675B (en) * 2020-11-25 2021-09-11 友達光電股份有限公司 Image recognition method and apparatus
CN112455452A (en) * 2020-11-30 2021-03-09 恒大新能源汽车投资控股集团有限公司 Method, device and equipment for detecting driving state
CN112766050B (en) * 2020-12-29 2024-04-16 富泰华工业(深圳)有限公司 Dressing and operation checking method, computer device and storage medium
CN112660141A (en) * 2020-12-29 2021-04-16 长安大学 Method for identifying driver driving distraction behavior through driving behavior data
CN112754498B (en) * 2021-01-11 2023-05-26 一汽解放汽车有限公司 Driver fatigue detection method, device, equipment and storage medium
CN114005104A (en) * 2021-03-23 2022-02-01 深圳市创乐慧科技有限公司 Intelligent driving method and device based on artificial intelligence and related products
CN113313019A (en) * 2021-05-27 2021-08-27 展讯通信(天津)有限公司 Distracted driving detection method, system and related equipment
CN113298041A (en) * 2021-06-21 2021-08-24 黑芝麻智能科技(上海)有限公司 Method and system for calibrating driver distraction reference direction
CN113139531A (en) * 2021-06-21 2021-07-20 博泰车联网(南京)有限公司 Drowsiness state detection method and apparatus, electronic device, and readable storage medium
CN113486759B (en) * 2021-06-30 2023-04-28 上海商汤临港智能科技有限公司 Dangerous action recognition method and device, electronic equipment and storage medium
CN113537135A (en) * 2021-07-30 2021-10-22 三一重机有限公司 Driving monitoring method, device and system and readable storage medium
CN113734173B (en) * 2021-09-09 2023-06-20 东风汽车集团股份有限公司 Intelligent vehicle monitoring method, device and storage medium
KR102542683B1 (en) * 2021-09-16 2023-06-14 국민대학교산학협력단 Method and apparatus for classifying action based on hand tracking
FR3127355B1 (en) * 2021-09-20 2023-09-29 Renault Sas method of selecting an operating mode of an image capture device for facial recognition
KR102634012B1 (en) * 2021-10-12 2024-02-07 경북대학교 산학협력단 Apparatus for detecting driver behavior using object classification based on deep running
CN114162130B (en) * 2021-10-26 2023-06-20 东风柳州汽车有限公司 Driving assistance mode switching method, device, equipment and storage medium
CN114187581B (en) * 2021-12-14 2024-04-09 安徽大学 Driver distraction fine granularity detection method based on unsupervised learning
CN114005105B (en) * 2021-12-30 2022-04-12 青岛以萨数据技术有限公司 Driving behavior detection method and device and electronic equipment
CN114582090A (en) * 2022-02-27 2022-06-03 武汉铁路职业技术学院 Rail vehicle drives monitoring and early warning system
CN114666378A (en) * 2022-03-03 2022-06-24 武汉科技大学 Vehicle-mounted remote monitoring system of heavy-duty diesel vehicle
KR20230145614A (en) 2022-04-07 2023-10-18 한국기술교육대학교 산학협력단 System and method for monitoring driver safety
CN115035502A (en) * 2022-07-08 2022-09-09 北京百度网讯科技有限公司 Driver behavior monitoring method and device, electronic equipment and storage medium
CN114898341B (en) * 2022-07-14 2022-12-20 苏州魔视智能科技有限公司 Fatigue driving early warning method and device, electronic equipment and storage medium
CN115601709B (en) * 2022-11-07 2023-10-27 北京万理软件开发有限公司 Colliery staff violation statistics system, method, device and storage medium
CN116311181B (en) * 2023-03-21 2023-09-12 重庆利龙中宝智能技术有限公司 Method and system for rapidly detecting abnormal driving
CN116052136B (en) * 2023-03-27 2023-09-05 中国科学技术大学 Distraction detection method, vehicle-mounted controller, and computer storage medium
CN116645732B (en) * 2023-07-19 2023-10-10 厦门工学院 Site dangerous activity early warning method and system based on computer vision

Family Cites Families (113)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2546415B2 (en) * 1990-07-09 1996-10-23 トヨタ自動車株式会社 Vehicle driver monitoring device
US7738678B2 (en) * 1995-06-07 2010-06-15 Automotive Technologies International, Inc. Light modulation techniques for imaging objects in or around a vehicle
JP3843513B2 (en) * 1996-12-24 2006-11-08 トヨタ自動車株式会社 Vehicle alarm device
JPH11161798A (en) * 1997-12-01 1999-06-18 Toyota Motor Corp Vehicle driver monitoring device
JP3383594B2 (en) * 1998-09-29 2003-03-04 沖電気工業株式会社 Eye opening measurement device
JP3495934B2 (en) * 1999-01-08 2004-02-09 矢崎総業株式会社 Accident prevention system
US20120231773A1 (en) * 1999-08-27 2012-09-13 Lipovski Gerald John Jack Cuboid-based systems and methods for safe mobile texting.
AU2001258867A1 (en) * 2000-05-04 2001-11-12 Jin-Ho Song Automatic vehicle management apparatus and method using wire and wireless communication network
JP2003131785A (en) * 2001-10-22 2003-05-09 Toshiba Corp Interface device, operation control method and program product
US6926429B2 (en) * 2002-01-30 2005-08-09 Delphi Technologies, Inc. Eye tracking/HUD system
US6873714B2 (en) * 2002-02-19 2005-03-29 Delphi Technologies, Inc. Auto calibration and personalization of eye tracking system using larger field of view imager with higher resolution
BR0307760A (en) * 2002-02-19 2006-04-04 Volvo Technology Corp system and method for driver attention monitor monitoring and management
JP2004017939A (en) * 2002-06-20 2004-01-22 Denso Corp Apparatus and program for transmitting information for vehicle
JP3951231B2 (en) * 2002-12-03 2007-08-01 オムロン株式会社 Safe travel information mediation system, safe travel information mediation device used therefor, and safe travel information confirmation method
US7639148B2 (en) * 2003-06-06 2009-12-29 Volvo Technology Corporation Method and arrangement for controlling vehicular subsystems based on interpreted driver activity
KR100494848B1 (en) * 2004-04-16 2005-06-13 에이치케이이카 주식회사 Method for sensing if person sleeps inside vehicle and device therefor
DE102005018697A1 (en) * 2004-06-02 2005-12-29 Daimlerchrysler Ag Method and device for warning a driver in the event of leaving the traffic lane
JP4564320B2 (en) 2004-09-29 2010-10-20 アイシン精機株式会社 Driver monitor system
CN1680779A (en) * 2005-02-04 2005-10-12 江苏大学 Fatigue monitoring method and device for driver
US7253739B2 (en) * 2005-03-10 2007-08-07 Delphi Technologies, Inc. System and method for determining eye closure state
WO2006131926A2 (en) * 2005-06-09 2006-12-14 Drive Diagnostics Ltd. System and method for displaying a driving profile
US20070041552A1 (en) * 2005-06-13 2007-02-22 Moscato Jonathan D Driver-attentive notification system
JP2007237919A (en) * 2006-03-08 2007-09-20 Toyota Motor Corp Input operation device for vehicle
WO2008007781A1 (en) * 2006-07-14 2008-01-17 Panasonic Corporation Visual axis direction detection device and visual line direction detection method
US20130150004A1 (en) * 2006-08-11 2013-06-13 Michael Rosen Method and apparatus for reducing mobile phone usage while driving
CN100462047C (en) * 2007-03-21 2009-02-18 汤一平 Safe driving auxiliary device based on omnidirectional computer vision
CN101030316B (en) * 2007-04-17 2010-04-21 北京中星微电子有限公司 Safety driving monitoring system and method for vehicle
JP2008302741A (en) * 2007-06-05 2008-12-18 Toyota Motor Corp Driving support device
US20130275899A1 (en) * 2010-01-18 2013-10-17 Apple Inc. Application Gateway for Providing Different User Interfaces for Limited Distraction and Non-Limited Distraction Contexts
JP5208711B2 (en) 2008-12-17 2013-06-12 アイシン精機株式会社 Eye open / close discrimination device and program
CN101540090B (en) * 2009-04-14 2011-06-15 华南理工大学 Driver fatigue monitoring method based on multivariate information fusion
US10019634B2 (en) * 2010-06-04 2018-07-10 Masoud Vaziri Method and apparatus for an eye tracking wearable computer
US9460601B2 (en) * 2009-09-20 2016-10-04 Tibet MIMAR Driver distraction and drowsiness warning and sleepiness reduction for accident avoidance
CN101692980B (en) 2009-10-30 2011-06-08 深圳市汉华安道科技有限责任公司 Method and device for detecting fatigue driving
CN101877051A (en) * 2009-10-30 2010-11-03 江苏大学 Driver attention state monitoring method and device
US20110224875A1 (en) * 2010-03-10 2011-09-15 Cuddihy Mark A Biometric Application of a Polymer-based Pressure Sensor
US10592757B2 (en) * 2010-06-07 2020-03-17 Affectiva, Inc. Vehicular cognitive data collection using multiple devices
US10074024B2 (en) * 2010-06-07 2018-09-11 Affectiva, Inc. Mental state analysis using blink rate for vehicles
CN101950355B (en) * 2010-09-08 2012-09-05 中国人民解放军国防科学技术大学 Method for detecting fatigue state of driver based on digital video
JP5755012B2 (en) 2011-04-21 2015-07-29 キヤノン株式会社 Information processing apparatus, processing method thereof, program, and imaging apparatus
US11270699B2 (en) * 2011-04-22 2022-03-08 Emerging Automotive, Llc Methods and vehicles for capturing emotion of a human driver and customizing vehicle response
JP5288045B2 (en) * 2011-07-11 2013-09-11 トヨタ自動車株式会社 Emergency vehicle evacuation device
US8744642B2 (en) * 2011-09-16 2014-06-03 Lytx, Inc. Driver identification based on face data
CN102436715B (en) * 2011-11-25 2013-12-11 大连海创高科信息技术有限公司 Detection method for fatigue driving
KR20140025812A (en) * 2012-08-22 2014-03-05 삼성전기주식회사 Apparatus and method for sensing drowsy driving
JP2014048760A (en) * 2012-08-29 2014-03-17 Denso Corp Information presentation system, information presentation device, and information center for presenting information to vehicle driver
JP6036065B2 (en) * 2012-09-14 2016-11-30 富士通株式会社 Gaze position detection device and gaze position detection method
US9405982B2 (en) * 2013-01-18 2016-08-02 GM Global Technology Operations LLC Driver gaze detection system
US20140272811A1 (en) * 2013-03-13 2014-09-18 Mighty Carma, Inc. System and method for providing driving and vehicle related assistance to a driver
US10210761B2 (en) * 2013-09-30 2019-02-19 Sackett Solutions & Innovations, LLC Driving assistance systems and methods
JP5939226B2 (en) 2013-10-16 2016-06-22 トヨタ自動車株式会社 Driving assistance device
KR101537936B1 (en) * 2013-11-08 2015-07-21 현대자동차주식회사 Vehicle and control method for the same
US10417486B2 (en) * 2013-12-30 2019-09-17 Alcatel Lucent Driver behavior monitoring systems and methods for driver behavior monitoring
JP6150258B2 (en) * 2014-01-15 2017-06-21 みこらった株式会社 Self-driving car
JP6213282B2 (en) * 2014-02-12 2017-10-18 株式会社デンソー Driving assistance device
US20150310758A1 (en) * 2014-04-26 2015-10-29 The Travelers Indemnity Company Systems, methods, and apparatus for generating customized virtual reality experiences
US20160001785A1 (en) * 2014-07-07 2016-01-07 Chin-Jung Hsu Motion sensing system and method
US9714037B2 (en) * 2014-08-18 2017-07-25 Trimble Navigation Limited Detection of driver behaviors using in-vehicle systems and methods
US9796391B2 (en) * 2014-10-13 2017-10-24 Verizon Patent And Licensing Inc. Distracted driver prevention systems and methods
TW201615457A (en) * 2014-10-30 2016-05-01 鴻海精密工業股份有限公司 Safety recognition and reaction system and method utilized in automobile
CN104408879B (en) * 2014-11-19 2017-02-01 湖南工学院 Method, device and system for processing fatigue driving early warning
US10614726B2 (en) * 2014-12-08 2020-04-07 Life Long Driver, Llc Behaviorally-based crash avoidance system
CN104574817A (en) * 2014-12-25 2015-04-29 清华大学苏州汽车研究院(吴江) Machine vision-based fatigue driving pre-warning system suitable for smart phone
JP2016124364A (en) * 2014-12-26 2016-07-11 本田技研工業株式会社 Awakener
US10705521B2 (en) 2014-12-30 2020-07-07 Visteon Global Technologies, Inc. Autonomous driving interface
DE102015200697A1 (en) * 2015-01-19 2016-07-21 Robert Bosch Gmbh Method and apparatus for detecting microsleep of a driver of a vehicle
CN104688251A (en) * 2015-03-02 2015-06-10 西安邦威电子科技有限公司 Method for detecting fatigue driving and driving in abnormal posture under multiple postures
FR3033303B1 (en) * 2015-03-03 2017-02-24 Renault Sas DEVICE AND METHOD FOR PREDICTING A LEVEL OF VIGILANCE IN A DRIVER OF A MOTOR VEHICLE.
EP3286057B1 (en) * 2015-04-20 2021-06-09 Bayerische Motoren Werke Aktiengesellschaft Apparatus and method for controlling a user situation awareness modification of a user of a vehicle, and a user situation awareness modification processing system
CN105139583A (en) * 2015-06-23 2015-12-09 南京理工大学 Vehicle danger prompting method based on portable intelligent equipment
CN106327801B (en) * 2015-07-07 2019-07-26 北京易车互联信息技术有限公司 Method for detecting fatigue driving and device
CN204915314U (en) * 2015-07-21 2015-12-30 戴井之 Car safe driving device
CN105096528B (en) * 2015-08-05 2017-07-11 广州云从信息科技有限公司 A kind of method for detecting fatigue driving and system
JP7011578B2 (en) * 2015-08-31 2022-01-26 エスアールアイ インターナショナル Methods and systems for monitoring driving behavior
CN105261153A (en) * 2015-11-03 2016-01-20 北京奇虎科技有限公司 Vehicle running monitoring method and device
CN105354985B (en) * 2015-11-04 2018-01-12 中国科学院上海高等研究院 Fatigue driving monitoring apparatus and method
JP6641916B2 (en) * 2015-11-20 2020-02-05 オムロン株式会社 Automatic driving support device, automatic driving support system, automatic driving support method, and automatic driving support program
CN105574487A (en) * 2015-11-26 2016-05-11 中国第一汽车股份有限公司 Facial feature based driver attention state detection method
CN105654753A (en) * 2016-01-08 2016-06-08 北京乐驾科技有限公司 Intelligent vehicle-mounted safe driving assistance method and system
CN105769120B (en) * 2016-01-27 2019-01-22 深圳地平线机器人科技有限公司 Method for detecting fatigue driving and device
FR3048544B1 (en) * 2016-03-01 2021-04-02 Valeo Comfort & Driving Assistance DEVICE AND METHOD FOR MONITORING A DRIVER OF A MOTOR VEHICLE
US10108260B2 (en) * 2016-04-01 2018-10-23 Lg Electronics Inc. Vehicle control apparatus and method thereof
WO2017208529A1 (en) * 2016-06-02 2017-12-07 オムロン株式会社 Driver state estimation device, driver state estimation system, driver state estimation method, driver state estimation program, subject state estimation device, subject state estimation method, subject state estimation program, and recording medium
US20180012090A1 (en) * 2016-07-07 2018-01-11 Jungo Connectivity Ltd. Visual learning system and method for determining a driver's state
JP2018022229A (en) * 2016-08-01 2018-02-08 株式会社デンソーテン Safety driving behavior notification system and safety driving behavior notification method
CN106218405A (en) * 2016-08-12 2016-12-14 深圳市元征科技股份有限公司 Fatigue driving monitoring method and cloud server
CN106446811A (en) * 2016-09-12 2017-02-22 北京智芯原动科技有限公司 Deep-learning-based driver's fatigue detection method and apparatus
JP6940612B2 (en) * 2016-09-14 2021-09-29 ナウト, インコーポレイテッドNauto, Inc. Near crash judgment system and method
CN106355838A (en) * 2016-10-28 2017-01-25 深圳市美通视讯科技有限公司 Fatigue driving detection method and system
WO2018085804A1 (en) * 2016-11-07 2018-05-11 Nauto Global Limited System and method for driver distraction determination
CN106709420B (en) * 2016-11-21 2020-07-10 厦门瑞为信息技术有限公司 Method for monitoring driving behavior of commercial vehicle driver
US10467488B2 (en) * 2016-11-21 2019-11-05 TeleLingo Method to analyze attention margin and to prevent inattentive and unsafe driving
CN106585629B (en) * 2016-12-06 2019-07-12 广东泓睿科技有限公司 A kind of control method for vehicle and device
CN106585624B (en) * 2016-12-07 2019-07-26 深圳市元征科技股份有限公司 Driver status monitoring method and device
CN106781282A (en) * 2016-12-29 2017-05-31 天津中科智能识别产业技术研究院有限公司 A kind of intelligent travelling crane driver fatigue early warning system
CN106909879A (en) * 2017-01-11 2017-06-30 开易(北京)科技有限公司 A kind of method for detecting fatigue driving and system
CN106985750A (en) * 2017-01-17 2017-07-28 戴姆勒股份公司 In-car safety monitoring system and automobile for vehicle
FR3063557B1 (en) * 2017-03-03 2022-01-14 Valeo Comfort & Driving Assistance DEVICE FOR DETERMINING THE STATE OF ATTENTION OF A VEHICLE DRIVER, ON-BOARD SYSTEM COMPRISING SUCH A DEVICE, AND ASSOCIATED METHOD
DE112017007252T5 (en) * 2017-03-14 2019-12-19 Omron Corporation DRIVER MONITORING DEVICE, DRIVER MONITORING METHOD, LEARNING DEVICE AND LEARNING METHOD
US10922566B2 (en) * 2017-05-09 2021-02-16 Affectiva, Inc. Cognitive state evaluation for vehicle navigation
US10289938B1 (en) * 2017-05-16 2019-05-14 State Farm Mutual Automobile Insurance Company Systems and methods regarding image distification and prediction models
US10402687B2 (en) * 2017-07-05 2019-09-03 Perceptive Automata, Inc. System and method of predicting human interaction with vehicles
US10592785B2 (en) * 2017-07-12 2020-03-17 Futurewei Technologies, Inc. Integrated system for detection of driver condition
WO2019028798A1 (en) * 2017-08-10 2019-02-14 北京市商汤科技开发有限公司 Method and device for monitoring driving condition, and electronic device
JP6666892B2 (en) * 2017-11-16 2020-03-18 株式会社Subaru Driving support device and driving support method
CN107933471B (en) * 2017-12-04 2019-12-20 惠州市德赛西威汽车电子股份有限公司 Accident active calling rescue method and vehicle-mounted automatic help-seeking system
CN108407813A (en) * 2018-01-25 2018-08-17 惠州市德赛西威汽车电子股份有限公司 A kind of antifatigue safe driving method of vehicle based on big data
US10322728B1 (en) * 2018-02-22 2019-06-18 Futurewei Technologies, Inc. Method for distress and road rage detection
US10776644B1 (en) * 2018-03-07 2020-09-15 State Farm Mutual Automobile Insurance Company Image analysis technologies for assessing safety of vehicle operation
US10970571B2 (en) * 2018-06-04 2021-04-06 Shanghai Sensetime Intelligent Technology Co., Ltd. Vehicle control method and system, vehicle-mounted intelligent system, electronic device, and medium
US10915769B2 (en) * 2018-06-04 2021-02-09 Shanghai Sensetime Intelligent Technology Co., Ltd Driving management methods and systems, vehicle-mounted intelligent systems, electronic devices, and medium
JP6870660B2 (en) * 2018-06-08 2021-05-12 トヨタ自動車株式会社 Driver monitoring device
CN108961669A (en) * 2018-07-19 2018-12-07 上海小蚁科技有限公司 The safe early warning method and device, storage medium, server of net about vehicle

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11037348B2 (en) * 2016-08-19 2021-06-15 Beijing Sensetime Technology Development Co., Ltd Method and apparatus for displaying business object in video image and electronic device
US20220230474A1 (en) * 2019-05-08 2022-07-21 Jaguar Land Rover Limited Activity identification method and apparatus
US20220265881A1 (en) * 2021-02-25 2022-08-25 Toyota Jidosha Kabushiki Kaisha Taxi vehicle and taxi system

Also Published As

Publication number Publication date
JP6933668B2 (en) 2021-09-08
US20210049387A1 (en) 2021-02-18
CN109803583A (en) 2019-05-24
SG11202002549WA (en) 2020-04-29
US20190065873A1 (en) 2019-02-28
CN110399767A (en) 2019-11-01
EP3666577A4 (en) 2020-08-19
TWI758689B (en) 2022-03-21
WO2019029195A1 (en) 2019-02-14
SG11202009720QA (en) 2020-10-29
CN109937152A (en) 2019-06-25
US20210049386A1 (en) 2021-02-18
KR20200124278A (en) 2020-11-02
JP2019536673A (en) 2019-12-19
EP3666577A1 (en) 2020-06-17
KR20200051632A (en) 2020-05-13
TW202033395A (en) 2020-09-16
WO2020173213A1 (en) 2020-09-03
JP2021517313A (en) 2021-07-15
US10853675B2 (en) 2020-12-01
CN109937152B (en) 2022-03-25
WO2019028798A1 (en) 2019-02-14
US20210049388A1 (en) 2021-02-18
KR102391279B1 (en) 2022-04-26

Similar Documents

Publication Publication Date Title
US20210009150A1 (en) Method for recognizing dangerous action of personnel in vehicle, electronic device and storage medium
KR102305914B1 (en) Driving management methods and systems, in-vehicle intelligent systems, electronic devices, media
JP7146959B2 (en) DRIVING STATE DETECTION METHOD AND DEVICE, DRIVER MONITORING SYSTEM AND VEHICLE
US10915769B2 (en) Driving management methods and systems, vehicle-mounted intelligent systems, electronic devices, and medium
TWI741512B (en) Method, device and electronic equipment for monitoring driver's attention
US11386676B2 (en) Passenger state analysis method and apparatus, vehicle, electronic device and storage medium
KR102469234B1 (en) Driving condition analysis method and device, driver monitoring system and vehicle
US20210012127A1 (en) Action recognition method and apparatus, driving action analysis method and apparatus, and storage medium
CN110956061B (en) Action recognition method and device, and driver state analysis method and device
US11403879B2 (en) Method and apparatus for child state analysis, vehicle, electronic device, and storage medium
US20220254063A1 (en) Gaze point estimation processing apparatus, gaze point estimation model generation apparatus, gaze point estimation processing system, and gaze point estimation processing method
CN111616718B (en) Method and system for detecting fatigue state of driver based on attitude characteristics
KR20160062521A (en) System and method for preventing neck disc
CN114663863A (en) Image processing method, image processing device, electronic equipment and computer storage medium
CN109657550B (en) Fatigue degree detection method and device
CN115761638A (en) Online real-time intelligent analysis method based on image data and terminal equipment
Ujir et al. Real-Time Driver’s Monitoring Mobile Application through Head Pose, Drowsiness and Angry Detection
Chinta et al. Driver Distraction Detection and Recognition
Kaur et al. Driver’s Drowsiness Detection System Using Machine Learning
Syahputra et al. EYE ASPECT RATIO ADJUSTMENT DETECTION FOR STRONG BLINKING SLEEPINESS BASED ON FACIAL LANDMARKS WITH EYE-BLINK DATASET
Swetha et al. Vehicle Accident Prevention System Using Artificial Intelligence
CN113989894A (en) Anti-tailing method, device and computer readable storage medium
Gill et al. INTELLIGENT DROWSY EYE DETECTION USING CONTOURLET TRANSFORM AND WEB LOCAL DESCRIPTORS.

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

AS Assignment

Owner name: BEIJING SENSETIME TECHNOLOGY DEVELOPMENT CO., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHEN, YANJIE;WANG, FEI;QIAN, CHEN;REEL/FRAME:054748/0547

Effective date: 20200624

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION