WO2022027894A1 - Driver behavior detection method and apparatus, electronic device, storage medium and program - Google Patents
Driver behavior detection method and apparatus, electronic device, storage medium and program Download PDFInfo
- Publication number
- WO2022027894A1 WO2022027894A1 PCT/CN2020/135501 CN2020135501W WO2022027894A1 WO 2022027894 A1 WO2022027894 A1 WO 2022027894A1 CN 2020135501 W CN2020135501 W CN 2020135501W WO 2022027894 A1 WO2022027894 A1 WO 2022027894A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- steering wheel
- detection result
- human hand
- driver
- pooling
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 392
- 238000000034 method Methods 0.000 claims abstract description 57
- 238000004590 computer program Methods 0.000 claims abstract description 14
- 238000011176 pooling Methods 0.000 claims description 146
- 230000006399 behavior Effects 0.000 claims description 97
- 230000008569 process Effects 0.000 claims description 23
- 230000006870 function Effects 0.000 claims description 11
- 230000004913 activation Effects 0.000 claims description 6
- 238000006243 chemical reaction Methods 0.000 claims description 5
- 230000003542 behavioural effect Effects 0.000 claims 1
- 230000004044 response Effects 0.000 description 16
- 238000010586 diagram Methods 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 206010039203 Road traffic accident Diseases 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
- B60W40/09—Driving style or behaviour
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W30/00—Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units, or advanced driver assistance systems for ensuring comfort, stability and safety or drive control systems for propelling or retarding the vehicle
- B60W30/08—Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/59—Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
- G06V20/597—Recognising the driver's state or behaviour, e.g. attention or drowsiness
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W2050/0001—Details of the control system
- B60W2050/0002—Automatic control, details of type of controller or control system architecture
- B60W2050/0004—In digital systems, e.g. discrete-time systems involving sampling
- B60W2050/0005—Processor details or data handling, e.g. memory registers or chip architecture
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
- B60W2050/143—Alarm means
Definitions
- the present application relates to the technical field of deep learning, and in particular, to a driver behavior detection method, device, electronic device, computer storage medium and computer program.
- the safe driving of a vehicle is determined by many factors, such as the driver's driving behavior, road conditions, and weather conditions.
- the embodiments of the present application provide at least one driver behavior detection method, device, electronic device, computer storage medium, and computer program.
- the embodiment of the present application provides a driver behavior detection method, including:
- the target detection result is obtained by detecting the to-be-detected image corresponding to the obtained driving position area, and the target detection result includes the steering wheel detection result and the hand detection result, and the driving behavior category of the driver is determined through the target detection result. , and when the driver's driving behavior category is dangerous driving, a warning message is issued to realize the detection of the driver's driving behavior, so as to facilitate the safety reminder to the driver and improve the safety of vehicle driving.
- the driving behavior category of the driver is determined according to the target detection result, including:
- the detection result of the steering wheel and the detection result of the human hand determine the positional relationship between the steering wheel and the human hand
- the driving behavior category of the driver is determined.
- the driving behavior category of the driver is determined according to the position relationship, including:
- the driving behavior category of the driver is determined according to the position relationship, including:
- determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result including:
- the human hand detection result includes a human hand
- the detection frame corresponding to the human hand in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result have an overlapping area
- the positional relationship between the human hands is that the driver holds the steering wheel; if there is no overlapping area between the detection frame corresponding to the human hand in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result, it is determined that the steering wheel and the The positional relationship between the human hands is that the driver's hands are off the steering wheel.
- determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result including:
- the human hand detection result includes two human hands
- determine the The positional relationship between the steering wheel and the human hand is that the driver's hands are off the steering wheel; if there is an overlapping area between the detection frame corresponding to at least one human hand in the detection result of the human hand and the detection frame corresponding to the steering wheel in the detection result of the steering wheel, determine The positional relationship between the steering wheel and the human hand is that the driver holds the steering wheel.
- determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result including:
- each channel feature map in the two-channel classification feature map corresponds to a type of human hand. category;
- two feature values at the feature positions matching the center point position information are extracted from the classification feature map;
- the maximum eigenvalue is selected from the value, and the category of the channel feature map corresponding to the maximum eigenvalue in the classification feature map is determined as the category corresponding to the center point position information;
- the positional relationship between the steering wheel and the human hand is determined.
- the detection result of the steering wheel is determined, and the classification feature map is generated by performing at least one convolution process on the intermediate feature map, and then combined with the generated center point position information of the driver's hand, the steering wheel and the steering wheel can be more accurately determined.
- the positional relationship between the hands is determined, and the classification feature map is generated by performing at least one convolution process on the intermediate feature map, and then combined with the generated center point position information of the driver's hand, the steering wheel and the steering wheel can be more accurately determined. The positional relationship between the hands.
- determining the positional relationship between the steering wheel and the human hand based on the category corresponding to each center point position information indicated by the detection frame information corresponding to the human hand including:
- the category corresponding to the center point position information is determined as the positional relationship between the steering wheel and the human hand;
- the detection frame information corresponding to the human hand includes two center point position information
- the category corresponding to the two center point position information is that the driver's hand is off the steering wheel
- the positional relationship is that the driver's hands are off the steering wheel; in the categories corresponding to the two center point position information, there is at least one category corresponding to the center point position information is that the driver is holding the steering wheel.
- the positional relationship between the hands is that the driver holds the steering wheel.
- the steering wheel detection result includes a steering wheel
- determining the driving behavior category of the driver according to the target detection result including:
- the target detection result it is determined that the driving behavior category of the driver is dangerous driving.
- the image to be detected is detected to obtain a target detection result, including:
- the converted target channel feature map is subjected to maximum pooling processing to obtain multiple pooling values and the position corresponding to each pooling value in the multiple pooling values. index; the position index is used to identify the position of the pooled value in the converted target channel feature map;
- target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values
- the target detection result is determined according to the target detection frame information.
- generating the target detection frame information based on multiple pooling values and a position index corresponding to each of the multiple pooling values includes:
- At least one pooling value is greater than the set pooling threshold value among the plurality of pooling values, based on the plurality of pooling values and the pooling threshold, it is determined from the plurality of pooling values that the The target pooling value of the center point of the target detection frame;
- the target detection frame information is generated based on the position index corresponding to the target pooling value.
- the pooling value greater than the pooling threshold among the plurality of pooling values is determined as the target pooling value belonging to the center point of the target detection frame of the steering wheel or the driver's hand, based on the corresponding target pooling value.
- the position index generates at least one target detection frame information of the steering wheel or the driver's hand more accurately.
- generating the target detection frame information based on multiple pooling values and a position index corresponding to each of the multiple pooling values includes:
- the target detection frame information is empty.
- the embodiment of the present application provides a driver behavior detection device, including:
- an acquisition module configured to acquire a to-be-detected image of the driving position area in the vehicle cabin
- a detection module configured to detect the to-be-detected image to obtain a target detection result, where the target detection result includes a steering wheel detection result and a hand detection result;
- a determining module configured to determine the driving behavior category of the driver according to the target detection result
- the warning module is configured to issue warning information when the driving behavior category of the driver is dangerous driving.
- An embodiment of the present application provides an electronic device, including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and when the electronic device runs, the processor and the memory The machine-readable instructions are executed by the processor to execute the driver behavior detection method according to any one of the above embodiments.
- An embodiment of the present application provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the driver behavior detection method described in any of the foregoing embodiments is executed.
- Embodiments of the present application further provide a computer program, including computer-readable codes.
- the processor in the electronic device executes the program for implementing the above-mentioned methods in any of the foregoing embodiments.
- the described driver behavior detection method is not limited to:
- FIG. 1a shows a schematic diagram of an application scenario of an embodiment of the present application
- Fig. 1b shows a schematic flowchart of a driver behavior detection method provided by an embodiment of the present application
- FIG. 2 shows a schematic flowchart of a specific method for detecting an image to be detected and obtaining a target detection result in a driver behavior detection method provided by an embodiment of the present application
- FIG. 3 shows a schematic structural diagram of a driver behavior detection device provided by an embodiment of the present application
- FIG. 4 shows a schematic structural diagram of an electronic device provided by an embodiment of the present application.
- the embodiment of the present application provides a driver behavior detection method.
- the driver behavior detection method may be performed by a driver behavior detection device, and the driver behavior detection device may be a user equipment (User Equipment, UE), a mobile device, a user terminal, a terminal, a cellular Phone, cordless phone, Personal Digital Assistant (PDA), handheld device, computing device, vehicle-mounted device, wearable device, etc.
- the method can be implemented by the processor calling the computer-readable instructions stored in the memory .
- the method can be performed by a server.
- FIG. 1a is a schematic diagram of an application scenario of the embodiment of the present application, as shown in FIG. 10.
- Acquire an image to be detected 11 of the driving position area, and the image to be detected 11 includes a steering wheel 110 and a human hand 111; in the driver behavior detection device 10, the steering wheel 110 and the human hand can be obtained by performing processing through the driver behavior detection described in the foregoing embodiment.
- the driving behavior category of the driver can be determined, and when the driving behavior category of the driver is dangerous driving, the driver behavior detection device 10 issues a warning message, The detection of the driving behavior of the driver is realized, thereby facilitating the safety reminder to the driver and improving the safety of vehicle driving.
- Fig. 1b is a schematic flowchart of a driver behavior detection method provided by an embodiment of the present application, the method includes S101-S104, wherein:
- S102 Detect the image to be detected to obtain a target detection result, where the target detection result includes a steering wheel detection result and a human hand detection result.
- the target detection result is obtained by detecting the acquired image to be detected in the driving position area, and the target detection result includes the steering wheel detection result and the hand detection result, and the driving behavior category of the driver is determined by the target detection result, And when the driver's driving behavior category is dangerous driving, a warning message is issued to realize the detection of the driver's driving behavior, so as to facilitate the safety reminder to the driver and improve the safety of vehicle driving.
- a camera device may be provided in the vehicle cabin, and an image to be detected of the driving position area may be acquired in real time through the camera device provided in the vehicle cabin.
- the installation position of the imaging device may be a position where the steering wheel and the driver's seat area in the driving position area can be photographed.
- the images to be detected may be input into the trained neural network, and the images to be detected are detected respectively to obtain target detection results, wherein the target detection results include steering wheel detection results and human hand detection results.
- the steering wheel detection result includes the information of whether there is a steering wheel in the image to be detected.
- the steering wheel detection result includes the detection frame information of the steering wheel
- the human hand detection result includes the detection frame information of whether there is a human hand in the image to be detected.
- the hand detection result includes the detection frame information of the hand.
- the image to be detected is detected to obtain a target detection result, which may include:
- S202 Perform at least one target convolution process on the intermediate feature map to generate detection feature maps of multiple channels corresponding to the intermediate feature map.
- S204 perform maximum pooling processing on the converted target channel feature map to obtain multiple pooling values and each pooling value corresponding to the multiple pooling values
- the position index of ; the position index is used to identify the position of the pooled value in the transformed target channel feature map.
- the above embodiment obtains multiple pooled values and a position index corresponding to each of the multiple pooled values by performing maximum pooling processing on the target channel feature map, and generates target detection frame information, which is Generating object detection results provides data support.
- the image to be detected can be input into the trained neural network, and the backbone network in the trained neural network performs multiple convolution processing on the image to be detected to generate an intermediate feature map corresponding to the image to be detected.
- the structure of the backbone network in the neural network can be set according to actual needs.
- the intermediate feature map can be input into the steering wheel detection branch network and the hand detection branch network of the neural network respectively, to generate the steering wheel detection result and the human hand detection result.
- the generation of the steering wheel detection result will be described in detail below.
- At least one first convolution process may be performed on the intermediate feature map to generate detection feature maps of multiple channels corresponding to the steering wheel, and the number of channels corresponding to the detection feature map may be three channels.
- the detection feature map includes a first channel feature map representing the position (the first channel feature map is the target channel feature map), a second channel feature map representing the length information of the detection frame, and a feature map representing the width information of the detection frame.
- the third channel feature map is a first channel feature map representing the position (the first channel feature map is the target channel feature map), a second channel feature map representing the length information of the detection frame, and a feature map representing the width information of the detection frame.
- the activation function can be used to transform the feature value of the target channel feature map representing the position in the detection feature map of multiple channels to generate the converted target channel feature map.
- Each feature value in the converted target channel feature map is A numeric value between 0-1.
- the activation function may be a sigmoid function. For the feature value of any feature point in the converted target channel feature map, if the feature value is closer to 1, the probability that the feature point corresponding to the feature value belongs to the center point of the detection frame of the steering wheel is greater.
- the maximum pooling process can be performed on the converted target channel feature map to obtain the pooling value corresponding to each feature position in the target channel feature map and each pool.
- the location index corresponding to the pooled value; the location index can be used to identify the location of the pooled value in the transformed target channel feature map.
- the same position index in the corresponding position index at each feature position can be merged to obtain the target channel feature map corresponding to multiple pooling values and the position index corresponding to each pooling value in the multiple pooling values.
- the preset pooling size and pooling step size may be set according to actual needs. For example, the preset pooling size may be 3 ⁇ 3, and the preset pooling step size may be 1.
- the first detection frame information (ie, target detection frame information) corresponding to the steering wheel may be generated based on the plurality of pooled values and the position index corresponding to each of the plurality of pooled values.
- a 3 ⁇ 3 maximum pooling process with a step size of 1 may be performed on the target channel feature map; during pooling, for every 3 ⁇ 3 feature points in the target channel feature map
- the feature value of determine the maximum response value (that is, the pooling value) of the 3 ⁇ 3 feature points and the position index of the maximum response value on the feature map of the target channel.
- the number of maximum response values is related to the size of the target channel feature map; for example, if the size of the target channel feature map is 80 ⁇ 60 ⁇ 3, the maximum response obtained after the maximum pooling process is performed on the target channel feature map There are 80 ⁇ 60 values in total; and for each maximum response value, there may be at least one other maximum response value with the same position index.
- the maximum response values with the same position index are combined to obtain M maximum response values and a position index corresponding to each of the M maximum response values.
- the first detection frame information corresponding to the steering wheel is generated.
- the process of determining the information of the second detection frame corresponding to the human hand may refer to the process of determining the information of the first detection frame corresponding to the steering wheel, which will not be repeated here.
- the first detection frame information After obtaining the first detection frame information corresponding to the steering wheel, the first detection frame information may be determined as the steering wheel detection result. When the first detection frame information corresponding to the steering wheel is not obtained, it is determined that the steering wheel detection result does not include the steering wheel. And after obtaining the second detection frame information corresponding to the human hand, the second detection frame information may be determined as the human hand detection result. When the second detection frame information corresponding to the human hand is not obtained, it is determined that the human hand detection result does not include the human hand.
- generating target detection frame information based on a plurality of pooled values and a position index corresponding to each of the plurality of pooled values may include:
- Step A1 In the case where at least one pooling value is greater than the set pooling threshold in the multiple pooling values, determine the center of the target detection frame from the multiple pooling values based on the multiple pooling values and the pooling threshold The target pooling value for the point.
- Step A2 Generate target detection frame information based on the position index corresponding to the target pooling value.
- a pooling threshold can be set.
- multiple pooling values are set based on the set pooling threshold.
- the value is filtered to obtain a target pooling value that is greater than the pooling threshold among the multiple pooling values.
- there is no target pooling value that is, there is no first detection frame information of the steering wheel.
- the center point position information of the first detection frame corresponding to the steering wheel may be generated based on the position index corresponding to the target pooling value.
- the pooling threshold corresponding to the steering wheel and the pooling threshold corresponding to the driver's hand may be the same or different. Specifically, the pooling threshold corresponding to the steering wheel and the pooling threshold corresponding to the driver's hand may be determined according to the actual situation. For example, multi-frame sample images collected by the camera device corresponding to the image to be detected can be obtained, and an adaptive algorithm can be used to generate a pooling threshold corresponding to the steering wheel and a pooling threshold corresponding to the driver's hand according to the collected multi-frame sample images.
- each of the M maximum response values can be combined with the pooling threshold. Compare; when a certain maximum response value is greater than the pooling threshold, the maximum response value is determined as the target pooling value.
- the position index corresponding to the target pooling value that is, the position information of the center point of the first detection frame of the steering wheel.
- a feature matching the center point position information may be selected from the second channel feature map
- the second feature value at the position, the selected second feature value is determined as the length corresponding to the first detection frame of the steering wheel, and the third channel feature map at the feature position matching the center point position information is selected from the third channel.
- feature value, and the selected third feature value is determined as the width corresponding to the first detection frame of the steering wheel, and the size information of the first detection frame of the steering wheel is obtained.
- one or two second detection frame information can be obtained, that is, the second detection frame information corresponding to the left hand and/or the right hand respectively can be obtained.
- the process of determining the second detection frame information corresponding to the driver's hand reference may be made to the above-mentioned process of determining the first detection frame information of the steering wheel, which will not be repeated here.
- the pooling value greater than the pooling threshold among the plurality of pooling values is determined as the target pooling value belonging to the center point of the target detection frame of the steering wheel or the driver's hand, based on the corresponding target pooling value.
- the position index generates at least one target detection frame information of the steering wheel or the driver's hand more accurately.
- generating target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values may include: when the plurality of pooling values are less than or When it is equal to the set pooling threshold, it is determined that the target detection frame information is empty.
- the multiple pooling values corresponding to the steering wheel are less than or equal to the set pooling threshold, it is determined that the first detection frame information of the steering wheel is empty; there is at least one pooling value greater than the set pooling value among the multiple pooling values corresponding to the steering wheel
- the pooling threshold is , it is determined that the information of the first detection frame of the steering wheel is not empty.
- the driving behavior category of the driver can be determined based on the steering wheel detection result and the human hand detection result.
- the driving behavior category of the driver is determined according to the target detection result, which may include:
- the detection result of the steering wheel and the detection result of the human hand determine the positional relationship between the steering wheel and the human hand
- the driving behavior category of the driver is determined.
- the positional relationship between the steering wheel and the human hand can be determined first according to the steering wheel detection result and the human hand detection result, and the driver's driving behavior category can be determined according to the determined positional relationship, that is, whether the driver is driving safely or dangerously.
- determining the driving behavior category of the driver according to the positional relationship may include: when the positional relationship indicates that the driver is holding the steering wheel, determining that the driving behavior category of the driver is safe drive.
- the driver's behavior category is safe driving.
- the case where the driver holds the steering wheel includes the driver holding the steering wheel with the left hand, the driver holding the steering wheel with the right hand, or the driver holding the steering wheel with both hands.
- determining the driving behavior category of the driver according to the positional relationship may include: when the positional relationship indicates that the driver's hands are off the steering wheel, determining that the driver's driving behavior category is dangerous driving.
- the category of the driving behavior of the driver is determined to be dangerous driving.
- determining the driving behavior category of the driver according to the target detection result may include: determining the driver according to the target detection result. is classified as dangerous driving.
- the driving behavior category of the driver is determined to be abnormal.
- determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result may include: if the human hand detection result includes a human hand, if the human hand detection result includes a human hand
- the corresponding detection frame and the detection frame corresponding to the steering wheel in the steering wheel detection result have overlapping areas, and the positional relationship between the steering wheel and the human hand is determined to be that the driver holds the steering wheel; There is no overlapping area in the detection frame corresponding to the steering wheel, and it is determined that the positional relationship between the steering wheel and the human hand is that the driver's hands are off the steering wheel.
- determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result may include: if the human hand detection result includes two human hands, if two human hands are included in the human hand detection result There is no overlapping area between the detection frame corresponding to only the human hand and the detection frame corresponding to the steering wheel in the steering wheel detection result, and the positional relationship between the steering wheel and the human hand is determined to be that the driver's hands are off the steering wheel; if there is at least one human hand in the human hand detection result. There is an overlapping area between the detection frame and the detection frame corresponding to the steering wheel in the steering wheel detection result, and it is determined that the positional relationship between the steering wheel and the human hand is that the driver holds the steering wheel.
- the positional relationship between the steering wheel and the human hand can be determined by using the detection frame corresponding to the steering wheel in the steering wheel detection result and the detection frame corresponding to the human hand in the human hand detection result.
- the positional relationship between the steering wheel and the human hand is determined as holding the steering wheel.
- a non-overlapping area exists between the detection frame corresponding to the human hand and the detection frame corresponding to the steering wheel, it is determined that the positional relationship between the steering wheel and the human hand is that the hand is disengaged from the steering wheel.
- the human hand detection result includes two human hands
- the detection frame corresponding to at least one human hand and the detection frame corresponding to the steering wheel have an overlapping area
- the positional relationship between the steering wheel and the human hand is determined as holding the steering wheel.
- the detection frames corresponding to the two hands and the detection frames corresponding to the steering wheel both have non-overlapping regions, it is determined that the positional relationship between the steering wheel and the human hand is that the hand is separated from the steering wheel.
- determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result may include:
- an intermediate feature map corresponding to the image to be detected is generated
- two eigenvalues at the feature positions matching the center point position information are extracted from the classification feature map; the largest eigenvalue is selected from the two eigenvalues , the category of the channel feature map corresponding to the largest feature value in the classification feature map is determined as the category corresponding to the center point position information.
- the positional relationship between the steering wheel and the human hand is determined.
- each channel feature map in the two-channel classification feature map corresponds to a category of human hands.
- the category corresponding to the channel feature map of the 0th channel may be that the driver's hand is off the steering wheel; the category corresponding to the channel feature map of the first channel may be that the driver is holding the steering wheel.
- two eigenvalues at the feature positions matching the center point position information can be extracted from the classification feature map, and the largest eigenvalue is selected from the two eigenvalues.
- the category of the channel feature map corresponding to the maximum feature value in the classification feature map is determined as the category corresponding to the center point position information.
- the detection frame information corresponding to the human hand includes two center point position information (that is, including the center point position information corresponding to the left hand and the center point position information corresponding to the right hand), for each center point position information, determine the center point position information corresponding category.
- the category corresponding to the channel feature map of the 0th channel can be that the driver's hand is off the steering wheel;
- the category corresponding to the channel feature map of the first channel can be that the driver is holding the steering wheel, then the center corresponding to the left hand Point position information, two feature values are extracted from the classification feature map, namely 0.8 and 0.2, then the classification feature map, the category of the 0th channel feature map corresponding to 0.8 is determined as the center point position information corresponding to the left hand
- the category of the center point position information corresponding to the left hand is that the driver's hand is off the steering wheel.
- the category of the position information of the center point corresponding to the right hand can be obtained.
- the detection result of the steering wheel is determined, and the classification feature map is generated by performing at least one convolution process on the intermediate feature map, and then combined with the generated center point position information of the driver's hand, the steering wheel and the steering wheel can be more accurately determined.
- the positional relationship between the hands is determined, and the classification feature map is generated by performing at least one convolution process on the intermediate feature map, and then combined with the generated center point position information of the driver's hand, the steering wheel and the steering wheel can be more accurately determined. The positional relationship between the hands.
- determining the positional relationship between the steering wheel and the human hand may include:
- Manner 1 When the detection frame information corresponding to the human hand includes a center point position information, the category corresponding to the center point position information is determined as the positional relationship between the steering wheel and the human hand.
- Mode 2 When the detection frame information corresponding to the human hand includes two center point position information, and the category corresponding to the two center point position information is that the driver's hand is detached from the steering wheel, it is determined that the positional relationship between the steering wheel and the human hand is driving.
- the driver's hand is off the steering wheel; in the categories corresponding to the two center point position information, if there is at least one category corresponding to the center point position information is that the driver holds the steering wheel, the positional relationship between the steering wheel and the human hand is determined as the driver's hand. Hold the steering wheel.
- the position information corresponding to the human hand can be In the detection frame information, the category corresponding to the position information of the center point is determined as the positional relationship between the steering wheel and the human hand.
- the detection frame information corresponding to the human hand includes the center point position information corresponding to the left hand, and the type of the center point position information corresponding to the left hand is that when the driver holds the steering wheel, the positional relationship between the steering wheel and the human hand is the driver's hand. Hold the steering wheel.
- the detection frame information corresponding to the human hand includes two center point position information, that is, the detection frame information corresponding to the human hand includes the center point position information corresponding to the left hand and the center point position information corresponding to the right hand, then the two center The categories corresponding to the point position information are all when the driver's hand is off the steering wheel, and the positional relationship between the steering wheel and the human hand is determined to be that the driver's hand is off the steering wheel; the category of the center point position information corresponding to the left hand is that the driver is holding the steering wheel, and/ Or, when the type of the center point position information corresponding to the right hand is when the driver holds the steering wheel, the positional relationship between the steering wheel and the human hand is determined to be that the driver holds the steering wheel.
- warning information for the driver may be generated based on the driving behavior category of the driver.
- the warning information can be played in the form of voice.
- the generated warning message can be "Danger, please hold the steering wheel".
- the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
- an embodiment of the present application also provides a driver behavior detection device.
- a schematic structural diagram of a driver behavior detection device provided by an embodiment of the present application includes an acquisition module 301, a detection Module 302, determination module 303, and warning module 304, specifically:
- the obtaining module 301 is configured to obtain the to-be-detected image corresponding to the driving position area in the vehicle cabin;
- the detection module 302 is configured to detect the to-be-detected image to obtain a target detection result, where the target detection result includes a steering wheel detection result and a human hand detection result;
- a determination module 303 configured to determine the driving behavior category of the driver according to the target detection result
- the warning module 304 is configured to issue warning information when the driving behavior category of the driver is dangerous driving.
- the determining module 303 determines the driving behavior category of the driver according to the target detection result , the configuration is:
- the detection result of the steering wheel and the detection result of the human hand determine the positional relationship between the steering wheel and the human hand
- the driving behavior category of the driver is determined.
- the determining module 303 when determining the driving behavior category of the driver according to the positional relationship, is configured as follows:
- the determining module 303 when determining the driving behavior category of the driver according to the positional relationship, is configured as follows:
- the determining module 303 determines the driving behavior of the driver according to the target detection result category, configure as:
- the target detection result it is determined that the driving behavior category of the driver is dangerous driving.
- the detection module 302 when detecting the to-be-detected image to obtain a target detection result, is configured as follows:
- the converted target channel feature map is subjected to maximum pooling processing to obtain multiple pooling values and the position corresponding to each pooling value in the multiple pooling values. index; the position index is used to identify the position of the pooled value in the converted target channel feature map;
- target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values
- the target detection result is determined according to the target detection frame information.
- the detection module 302 when generating target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values, is configured as :
- determining the pooling value from the plurality of pooling values based on the plurality of pooling values and the pooling threshold The target pooling value of the center point of the target detection frame;
- the target detection frame information is generated based on the position index corresponding to the target pooling value.
- the detection module 302 when the detection module 302 generates the target detection frame information based on a plurality of pooling values and a position index corresponding to each pooling value in the plurality of pooling values, Configured as:
- the target detection frame information is empty.
- the determining module 303 when determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result, is configured to:
- the human hand detection result includes a human hand
- the detection frame corresponding to the human hand in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result have an overlapping area
- the positional relationship between the human hands is that the driver holds the steering wheel; if there is no overlapping area between the detection frame corresponding to the human hand in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result, it is determined that the steering wheel and the The positional relationship between the human hands is that the driver's hands are off the steering wheel.
- the determining module 303 when determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result, is configured to:
- the human hand detection result includes two human hands
- determine the The positional relationship between the steering wheel and the human hand is that the driver's hands are off the steering wheel; if there is an overlapping area between the detection frame corresponding to at least one human hand in the detection result of the human hand and the detection frame corresponding to the steering wheel in the detection result of the steering wheel, determine The positional relationship between the steering wheel and the human hand is that the driver holds the steering wheel.
- the determining module 303 when determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result, is configured to:
- each channel feature map in the two-channel classification feature map corresponds to a type of human hand. category;
- two feature values at the feature positions matching the center point position information are extracted from the classification feature map;
- the maximum eigenvalue is selected from the value, and the category of the channel feature map corresponding to the maximum eigenvalue in the classification feature map is determined as the category corresponding to the center point position information;
- the positional relationship between the steering wheel and the human hand is determined.
- the determining module 303 determines the positional relationship between the steering wheel and the human hand based on the category corresponding to each center point position information indicated by the detection frame information corresponding to the human hand, Configured as:
- the category corresponding to the center point position information is determined as the positional relationship between the steering wheel and the human hand;
- the detection frame information corresponding to the human hand includes two center point position information
- the category corresponding to the two center point position information is that the driver's hand is off the steering wheel
- determine the position between the steering wheel and the human hand The relationship is that the driver's hands are off the steering wheel; in the categories corresponding to the two center point position information, if there is at least one category corresponding to the center point position information is the driver holding the steering wheel, determine the steering wheel and the human hand The positional relationship between them is that the driver holds the steering wheel.
- the functions or templates included in the apparatuses provided in the embodiments of the present application may be used to execute the methods described in the above method embodiments.
- a schematic structural diagram of an electronic device provided in an embodiment of the present application includes a processor 401 , a memory 402 , and a bus 403 .
- the memory 402 is configured to store execution instructions, including the memory 4021 and the external memory 4022; the memory 4021 here is also called internal memory, and is configured to temporarily store the operation data in the processor 401 and the data exchanged with the external memory 4022 such as the hard disk,
- the processor 401 exchanges data with the external memory 4022 through the memory 4021.
- the processor 401 communicates with the memory 402 through the bus 403, so that the processor 401 executes the following instructions:
- the embodiments of the present application also provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the driver behavior detection method described in the above method embodiments is executed A step of.
- the computer program product of the driver behavior detection method provided by the embodiments of the present application includes a computer-readable storage medium storing program codes, and the instructions included in the program codes can be used to execute the driver behavior described in the above method embodiments.
- the steps of the detection method reference may be made to the foregoing method embodiments, which will not be repeated here.
- the units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
- each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
- the technical solution of the present application can be embodied in the form of a software product in essence, or the part that contributes to the prior art or the part of the technical solution.
- the computer software product is stored in a storage medium, including Several instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in the various embodiments of the present application.
- the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
- the embodiment of the present application discloses a driver behavior detection method, device, electronic device, computer storage medium and computer program.
- the method includes: obtaining a program code for controlling the operation of a first controlled device through a local agent program;
- the local agent program sends the program code to the first controlled device, so that the first controlled device runs the program code.
- the program code can be obtained and pushed based on the local agent program, and then the control of the first controlled device can be realized based on the program code, and the program code can be flexibly edited. Control the first controlled device.
Abstract
Description
Claims (16)
- 一种驾驶员行为检测方法,包括:A driver behavior detection method, comprising:获取车舱内的驾驶位置区域的待检测图像;Obtain the to-be-detected image of the driving position area in the cabin;对所述待检测图像进行检测,得到目标检测结果,所述目标检测结果包括方向盘检测结果和人手检测结果;Detecting the to-be-detected image to obtain a target detection result, where the target detection result includes a steering wheel detection result and a human hand detection result;根据所述目标检测结果,确定驾驶员的驾驶行为类别;According to the target detection result, determine the driving behavior category of the driver;在所述驾驶员的驾驶行为类别为危险驾驶时,发出警示信息。When the driving behavior category of the driver is dangerous driving, a warning message is issued.
- 根据权利要求1所述的方法,其中,在所述方向盘检测结果中包括方向盘,所述人手检测结果中包括人手时,所述根据所述目标检测结果,确定驾驶员的驾驶行为类别,包括:The method according to claim 1, wherein, when the steering wheel detection result includes a steering wheel and the human hand detection result includes a human hand, the determining the driving behavior category of the driver according to the target detection result includes:根据所述方向盘检测结果和所述人手检测结果,确定方向盘与人手之间的位置关系;According to the detection result of the steering wheel and the detection result of the human hand, determine the positional relationship between the steering wheel and the human hand;根据所述位置关系,确定驾驶员的驾驶行为类别。According to the position relationship, the driving behavior category of the driver is determined.
- 根据权利要求2所述的方法,其中,所述根据位置关系,确定驾驶员的驾驶行为类别,包括:The method according to claim 2, wherein the determining the driving behavior category of the driver according to the position relationship comprises:在所述位置关系指示驾驶员手握方向盘的情况下,确定所述驾驶员的驾驶行为类别为安全驾驶。When the positional relationship indicates that the driver is holding the steering wheel, it is determined that the driving behavior category of the driver is safe driving.
- 根据权利要求2所述的方法,其中,所述根据位置关系,确定驾驶员的驾驶行为类别,包括:The method according to claim 2, wherein the determining the driving behavior category of the driver according to the position relationship comprises:在所述位置关系指示驾驶员双手脱离方向盘的情况下,确定所述驾驶员的驾驶行为类别为危险驾驶。When the positional relationship indicates that the driver's hands are off the steering wheel, it is determined that the driver's driving behavior is classified as dangerous driving.
- 根据权利要求2所述的方法,其中,所述根据所述方向盘检测结果和所述人手检测结果,确定方向盘与人手之间的位置关系,包括:The method according to claim 2, wherein the determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result comprises:在所述人手检测结果中包括一只人手的情况下,若所述人手检测结果中人手对应的检测框、与所述方向盘检测结果中方向盘对应的检测框存在重合区域,确定所述方向盘与所述人手之间的位置关系为驾驶员手握方向盘;若所述人手检测结果中人手对应的检测框与所述方向盘检测结果中方向盘对应的检测框不存在重合区域,确定所述方向盘与所述人手之间的位置关系为驾驶员双手脱离方向盘。In the case where the human hand detection result includes a human hand, if the detection frame corresponding to the human hand in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result have an overlapping area, it is determined that the steering wheel and the detection frame corresponding to the steering wheel in the steering wheel detection result overlap. The positional relationship between the human hands is that the driver holds the steering wheel; if there is no overlapping area between the detection frame corresponding to the human hand in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result, it is determined that the steering wheel and the The positional relationship between the human hands is that the driver's hands are off the steering wheel.
- 根据权利要求2所述的方法,其中,所述根据所述方向盘检测结果和所述人手检测结果,确定方向盘与人手之间的位置关系,包括:The method according to claim 2, wherein the determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result comprises:在所述人手检测结果中包括两只人手的情况下,若所述人手检测结果中两只人手对应的检测框分别与所述方向盘检测结果中方向盘对应的检测框不存在重合区域,确定所述方向盘与所述人手之间的位置关系为驾驶员双手脱离方向盘;若所述人手检测结果中存在至少一只人手对应的检测框与所述方向盘检测结果中方向盘对应的检测框存在重合区域,确定所述方向盘与所述人手之间的位置关系为驾驶员手握方向盘。In the case where the human hand detection result includes two human hands, if there is no overlapping area between the detection frames corresponding to the two human hands in the human hand detection result and the detection frame corresponding to the steering wheel in the steering wheel detection result, determine the The positional relationship between the steering wheel and the human hand is that the driver's hands are off the steering wheel; if there is an overlapping area between the detection frame corresponding to at least one human hand in the detection result of the human hand and the detection frame corresponding to the steering wheel in the detection result of the steering wheel, determine The positional relationship between the steering wheel and the human hand is that the driver holds the steering wheel.
- 根据权利要求2所述的方法,其中,所述根据所述方向盘检测结果和所述人手检测结果,确定方向盘与人手之间的位置关系,包括:The method according to claim 2, wherein the determining the positional relationship between the steering wheel and the human hand according to the steering wheel detection result and the human hand detection result comprises:基于所述待检测图像,生成所述待检测图像对应的中间特征图;generating an intermediate feature map corresponding to the to-be-detected image based on the to-be-detected image;对所述中间特征图进行至少一次卷积处理,生成所述中间特征图对应的二通道的分类特征图;其中,所述二通道的分类特征图中的每个通道特征图对应一种人手的类别;Perform at least one convolution process on the intermediate feature map to generate a two-channel classification feature map corresponding to the intermediate feature map; wherein, each channel feature map in the two-channel classification feature map corresponds to a type of human hand. category;基于所述人手检测结果中人手对应的检测框信息指示的中心点位置信息,从所述分类特征图中提取与所述中心点位置信息匹配的特征位置处的两个特征值;从两个特征值中选取最大特征值,将所述分类特征图中,与所述最大特征值对应的通道特征图的类别,确定为所述中心点位置信息对应的类别;Based on the center point position information indicated by the detection frame information corresponding to the human hand in the hand detection result, two feature values at the feature positions matching the center point position information are extracted from the classification feature map; The maximum eigenvalue is selected from the value, and the category of the channel feature map corresponding to the maximum eigenvalue in the classification feature map is determined as the category corresponding to the center point position information;基于所述人手对应的检测框信息指示的每个中心点位置信息对应的类别,确定所述方向盘与所述人手之间的位置关系。The positional relationship between the steering wheel and the human hand is determined based on the category corresponding to each center point position information indicated by the detection frame information corresponding to the human hand.
- 根据权利要求7所述的方法,其中,基于所述人手对应的检测框信息指示的每个中心点位置信息对应的类别,确定所述方向盘与人手之间的位置关系,包括:The method according to claim 7, wherein determining the positional relationship between the steering wheel and the human hand based on the category corresponding to each center point position information indicated by the detection frame information corresponding to the human hand, comprising:在所述人手对应的检测框信息中包括一个中心点位置信息的情况下,将所述中心点位置信息对应的类别,确定为方向盘与人手之间的位置关系;In the case that the detection frame information corresponding to the human hand includes a center point position information, the category corresponding to the center point position information is determined as the positional relationship between the steering wheel and the human hand;在所述人手对应的检测框信息中包括两个中心点位置信息、且所述两个中心点位置信息对应的类别为驾驶员手脱离方向盘的情况下,确定所述方向盘与所述人手之间的位置关系为驾驶员双手脱离方向盘;在所述两个中心点位置信息对应的类别中,存在至少一个中心点位置信息对应的类别为驾驶员手握方向盘的情况下,确定所述方向盘与所述人手之间的位置关系为驾驶员手握方向盘。In the case where the detection frame information corresponding to the human hand includes two center point position information, and the category corresponding to the two center point position information is that the driver's hand is off the steering wheel, determine the distance between the steering wheel and the human hand. The positional relationship is that the driver's hands are off the steering wheel; in the categories corresponding to the two center point position information, there is at least one category corresponding to the center point position information that the driver holds the steering wheel, determine the steering wheel and all The positional relationship between the hands is that the driver holds the steering wheel.
- 根据权利要求1所述的方法,其中,所述在所述方向盘检测结果中包括方向盘,所述人手检测结果中未包括人手时,根据所述目标检测结果,确定驾驶员的驾驶行为类别,包括:The method according to claim 1, wherein when the steering wheel detection result includes a steering wheel and the human hand detection result does not include a human hand, determining the driving behavior category of the driver according to the target detection result, including :根据所述目标检测结果,确定所述驾驶员的驾驶行为类别为危险驾驶。According to the target detection result, it is determined that the driving behavior category of the driver is dangerous driving.
- 根据权利要求1至9任一项所述的方法,其中,所述对所述待检测图像进行检测,得到目标检测结果,包括:The method according to any one of claims 1 to 9, wherein the detecting the to-be-detected image to obtain a target detection result comprises:基于所述待检测图像,生成所述待检测图像对应的中间特征图;generating an intermediate feature map corresponding to the to-be-detected image based on the to-be-detected image;对所述中间特征图进行至少一次目标卷积处理,生成所述中间特征图对应的多个通道的检测特征图;Perform at least one target convolution process on the intermediate feature map to generate detection feature maps of multiple channels corresponding to the intermediate feature map;利用激活函数对多个通道的所述检测特征图中表征位置的目标通道特征图的每个特征值进行特征值转换处理,生成转换后的目标通道特征图;Using the activation function to perform feature value conversion processing on each feature value of the target channel feature map representing the position in the detection feature maps of multiple channels, to generate a converted target channel feature map;按照预设的池化尺寸和池化步长,对转换后的目标通道特征图进行最大池化处理,得到多个池化值以及与多个池化值中的每个池化值对应的位置索引;所述位置索引用于标识所述池化值在所述转换后的目标通道特征图中的位置;According to the preset pooling size and pooling step size, the converted target channel feature map is subjected to maximum pooling processing to obtain multiple pooling values and the position corresponding to each pooling value in the multiple pooling values. index; the position index is used to identify the position of the pooled value in the converted target channel feature map;基于多个池化值以及与多个池化值中的每个池化值对应的位置索引,生成目标检测框信息;generating target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values;根据所述目标检测框信息,确定所述目标检测结果。The target detection result is determined according to the target detection frame information.
- 根据权利要求10所述的方法,其中,所述基于多个池化值以及与多个池化值中的每个池化值对应的位置索引,生成所述目标检测框信息,包括:The method according to claim 10, wherein the generating the target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values comprises:在所述多个池化值中存在至少一个池化值大于设置的池化阈值的情况下,基于所述多个池化值以及池化阈值,从所述多个池化值中确定所述目标检测框的中心点的目标池化值;In the case where at least one pooling value is greater than a set pooling threshold value among the plurality of pooling values, determining the pooling value from the plurality of pooling values based on the plurality of pooling values and the pooling threshold The target pooling value of the center point of the target detection frame;基于所述目标池化值对应的位置索引,生成所述目标检测框信息。The target detection frame information is generated based on the position index corresponding to the target pooling value.
- 根据权利要求10所述的方法,其中,所述基于多个池化值以及与多个池化值中的每个池化值对应的位置索引,生成所述目标检测框信息,包括:The method according to claim 10, wherein the generating the target detection frame information based on a plurality of pooling values and a position index corresponding to each of the plurality of pooling values comprises:在所述多个池化值小于或等于设置的池化阈值的情况下,确定所述目标检测框信息为空。In the case that the multiple pooling values are less than or equal to the set pooling threshold, it is determined that the target detection frame information is empty.
- 一种驾驶员行为检测装置,包括:A driver behavior detection device, comprising:获取模块,配置为获取车舱内的驾驶位置区域的待检测图像;an acquisition module, configured to acquire a to-be-detected image of the driving position area in the vehicle cabin;检测模块,配置为对所述待检测图像进行检测,得到目标检测结果,所述目标检测结果包括方向盘检测结果和人手检测结果;a detection module, configured to detect the to-be-detected image to obtain a target detection result, where the target detection result includes a steering wheel detection result and a human hand detection result;确定模块,配置为根据所述目标检测结果,确定驾驶员的驾驶行为类别;a determining module, configured to determine the driving behavior category of the driver according to the target detection result;警示模块,配置为在所述驾驶员的驾驶行为类别为危险驾驶时,发出警示信息。The warning module is configured to issue warning information when the driving behavior category of the driver is dangerous driving.
- 一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,当电子设备运行时,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如权利要求1至12任一所述的驾驶员行为检测方法。An electronic device, comprising: a processor, a memory and a bus, the memory stores machine-readable instructions executable by the processor, and when the electronic device is running, the processor and the memory communicate through the bus , when the machine-readable instructions are executed by the processor, the driver behavior detection method according to any one of claims 1 to 12 is executed.
- 一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行如权利要求1至12任一所述的驾驶员行为检测方法。A computer-readable storage medium storing a computer program on the computer-readable storage medium, the computer program executing the driver behavior detection method according to any one of claims 1 to 12 when the computer program is run by a processor.
- 一种计算机程序,包括计算机可读代码,当所述计算机可读代码在电子设备中运行时,所述电子设备中的处理器执行配置为实现权利要求1至12中任一所述的驾驶员行为检测方法。A computer program comprising computer readable code, when the computer readable code is run in an electronic device, a processor in the electronic device executes a driver configured to implement any one of claims 1 to 12 Behavioral detection methods.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020227003906A KR20220032074A (en) | 2020-08-07 | 2020-12-10 | Driver behavior detection method, apparatus, electronic device, storage medium and program |
JP2022523602A JP2023500218A (en) | 2020-08-07 | 2020-12-10 | DRIVER ACTION DETECTION METHOD, APPARATUS, ELECTRONIC DEVICE, STORAGE MEDIUM AND PROGRAM |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010790208.3 | 2020-08-07 | ||
CN202010790208.3A CN111931639A (en) | 2020-08-07 | 2020-08-07 | Driver behavior detection method and device, electronic equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022027894A1 true WO2022027894A1 (en) | 2022-02-10 |
Family
ID=73307528
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/135501 WO2022027894A1 (en) | 2020-08-07 | 2020-12-10 | Driver behavior detection method and apparatus, electronic device, storage medium and program |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP2023500218A (en) |
KR (1) | KR20220032074A (en) |
CN (1) | CN111931639A (en) |
WO (1) | WO2022027894A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115471826A (en) * | 2022-08-23 | 2022-12-13 | 中国航空油料集团有限公司 | Method and device for judging safe driving behavior of aircraft refueling truck and safe operation and maintenance system |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111931639A (en) * | 2020-08-07 | 2020-11-13 | 上海商汤临港智能科技有限公司 | Driver behavior detection method and device, electronic equipment and storage medium |
CN112528910B (en) * | 2020-12-18 | 2023-04-07 | 上海高德威智能交通系统有限公司 | Hand-off steering wheel detection method and device, electronic equipment and storage medium |
CN113486759B (en) * | 2021-06-30 | 2023-04-28 | 上海商汤临港智能科技有限公司 | Dangerous action recognition method and device, electronic equipment and storage medium |
CN115171082B (en) * | 2022-06-29 | 2024-01-19 | 北京百度网讯科技有限公司 | Driving behavior detection method and device, electronic equipment and readable storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102263937A (en) * | 2011-07-26 | 2011-11-30 | 华南理工大学 | Driver's driving behavior monitoring device and monitoring method based on video detection |
CN102289660A (en) * | 2011-07-26 | 2011-12-21 | 华南理工大学 | Method for detecting illegal driving behavior based on hand gesture tracking |
CN109034111A (en) * | 2018-08-17 | 2018-12-18 | 北京航空航天大学 | A kind of driver's hand based on deep learning is from steering wheel detection method and system |
CN109086662A (en) * | 2018-06-19 | 2018-12-25 | 浙江大华技术股份有限公司 | A kind of anomaly detection method and device |
US20200143563A1 (en) * | 2017-11-22 | 2020-05-07 | Beijing Sensetime Technology Development Co., Ltd. | Methods and apparatuses for object detection, and devices |
CN111931639A (en) * | 2020-08-07 | 2020-11-13 | 上海商汤临港智能科技有限公司 | Driver behavior detection method and device, electronic equipment and storage medium |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5018926B2 (en) * | 2010-04-19 | 2012-09-05 | 株式会社デンソー | Driving assistance device and program |
CN104276080B (en) * | 2014-10-16 | 2016-04-06 | 北京航空航天大学 | Bus man hand detects forewarn system and method for early warning from bearing circle |
CN107766865A (en) * | 2017-11-06 | 2018-03-06 | 北京旷视科技有限公司 | Pond method, object detecting method, device, system and computer-readable medium |
CN111439170B (en) * | 2020-03-30 | 2021-09-17 | 上海商汤临港智能科技有限公司 | Child state detection method and device, electronic equipment and storage medium |
-
2020
- 2020-08-07 CN CN202010790208.3A patent/CN111931639A/en active Pending
- 2020-12-10 WO PCT/CN2020/135501 patent/WO2022027894A1/en active Application Filing
- 2020-12-10 JP JP2022523602A patent/JP2023500218A/en not_active Withdrawn
- 2020-12-10 KR KR1020227003906A patent/KR20220032074A/en not_active Application Discontinuation
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102263937A (en) * | 2011-07-26 | 2011-11-30 | 华南理工大学 | Driver's driving behavior monitoring device and monitoring method based on video detection |
CN102289660A (en) * | 2011-07-26 | 2011-12-21 | 华南理工大学 | Method for detecting illegal driving behavior based on hand gesture tracking |
US20200143563A1 (en) * | 2017-11-22 | 2020-05-07 | Beijing Sensetime Technology Development Co., Ltd. | Methods and apparatuses for object detection, and devices |
CN109086662A (en) * | 2018-06-19 | 2018-12-25 | 浙江大华技术股份有限公司 | A kind of anomaly detection method and device |
CN109034111A (en) * | 2018-08-17 | 2018-12-18 | 北京航空航天大学 | A kind of driver's hand based on deep learning is from steering wheel detection method and system |
CN111931639A (en) * | 2020-08-07 | 2020-11-13 | 上海商汤临港智能科技有限公司 | Driver behavior detection method and device, electronic equipment and storage medium |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115471826A (en) * | 2022-08-23 | 2022-12-13 | 中国航空油料集团有限公司 | Method and device for judging safe driving behavior of aircraft refueling truck and safe operation and maintenance system |
CN115471826B (en) * | 2022-08-23 | 2024-03-26 | 中国航空油料集团有限公司 | Method and device for judging safe driving behavior of aviation fueller and safe operation and maintenance system |
Also Published As
Publication number | Publication date |
---|---|
CN111931639A (en) | 2020-11-13 |
KR20220032074A (en) | 2022-03-15 |
JP2023500218A (en) | 2023-01-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022027894A1 (en) | Driver behavior detection method and apparatus, electronic device, storage medium and program | |
US20220084316A1 (en) | Method and electronic device for recognizing abnormal sitting posture, and storage medium | |
KR20190060817A (en) | Image based vehicle damage determination method and apparatus, and electronic device | |
CN111259737B (en) | Method and device for predicting failure of steering wheel of vehicle, electronic equipment and storage medium | |
JP7288097B2 (en) | Seat belt wearing detection method, device, electronic device, storage medium and program | |
CN111523489A (en) | Generation method of age classification network, and vehicle-mounted person detection method and device | |
CN111582272A (en) | Double-row license plate recognition method, device and equipment and computer readable storage medium | |
CN110598562A (en) | Vehicle image acquisition guiding method and device | |
CN107867295B (en) | Accident early warning method based on vehicle risk probability, storage device and vehicle-mounted terminal | |
CN112818839A (en) | Method, device, equipment and medium for identifying violation behaviors of driver | |
CN114373189A (en) | Behavior detection method and apparatus, terminal device and storage medium | |
CN111310650A (en) | Vehicle riding object classification method and device, computer equipment and storage medium | |
CN112926544A (en) | Driving state determination method, device, equipment and storage medium | |
CN112883843A (en) | Driver visual salient region detection method and device and computer equipment | |
US10266132B2 (en) | Method for operating driver assistance systems in a motor vehicle, and motor vehicle | |
CN113051958A (en) | Driver state detection method, system, device and medium based on deep learning | |
CN112277956B (en) | Method and device for detecting drunk driving behavior by using steering wheel | |
CN115115531A (en) | Image denoising method and device, vehicle and storage medium | |
CN113920310A (en) | Method and device for detecting steering wheel hands off | |
JP2021096710A (en) | Agent device, agent system, and agent program | |
CN110751008B (en) | Processing method, device, equipment and storage medium based on driving behavior | |
CN114241453B (en) | Driver distraction driving monitoring method utilizing key point attention | |
CN111797784B (en) | Driving behavior monitoring method and device, electronic equipment and storage medium | |
CN112215840A (en) | Image detection method, image detection device, driving control method, driving control device, electronic equipment and storage medium | |
CN113569844A (en) | License plate detection method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 20227003906 Country of ref document: KR Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20947943 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022523602 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 10.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20947943 Country of ref document: EP Kind code of ref document: A1 |