CN113569817A - Driver attention dispersion detection method based on image area positioning mechanism - Google Patents

Driver attention dispersion detection method based on image area positioning mechanism Download PDF

Info

Publication number
CN113569817A
CN113569817A CN202111110059.2A CN202111110059A CN113569817A CN 113569817 A CN113569817 A CN 113569817A CN 202111110059 A CN202111110059 A CN 202111110059A CN 113569817 A CN113569817 A CN 113569817A
Authority
CN
China
Prior art keywords
driver
image
state
neural network
behavior
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111110059.2A
Other languages
Chinese (zh)
Other versions
CN113569817B (en
Inventor
赵磊
孙浩然
罗映
徐楠
刘建华
闫法义
贝太学
李新海
张宗喜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong Jianzhu University
Original Assignee
Shandong Jianzhu University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong Jianzhu University filed Critical Shandong Jianzhu University
Priority to CN202111110059.2A priority Critical patent/CN113569817B/en
Publication of CN113569817A publication Critical patent/CN113569817A/en
Application granted granted Critical
Publication of CN113569817B publication Critical patent/CN113569817B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Biophysics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Biology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

A driver attention dispersion detection method based on an image area positioning mechanism is characterized in that areas needing attention under different behavior states of a driver in an image are obtained through a manual calibration method, the areas are combined with neural network activation mapping, a model optimization function based on area enhancement driving is established, the neural network is trained through the optimization function, so that a detection model can automatically obtain key areas in the image of the driver according to different behavior characteristics of the driver in the detection process, the problem of automatic extraction of key positions and features in a driver behavior detection method based on image information is solved, and the detection precision of the model is improved.

Description

Driver attention dispersion detection method based on image area positioning mechanism
Technical Field
The invention relates to the technical field of driver state recognition, in particular to a driver attention dispersion detection method based on an image area positioning mechanism.
Background
With the development of science and technology, intelligent electronic devices such as smart phones, tablet computers and vehicle information systems greatly improve the probability of driver distraction, easily generate potential safety hazards to cause traffic accidents and harm life and property safety. Statistically, nearly 125 million people die of traffic accidents each year. Nearly one fifth of accidents are caused by distractions of the driver. With the progress of artificial intelligence technology, the automatic driving technology is rapidly developed. However, current conditional autonomous driving systems still require a driver who is ready to take over in time. The american national traffic safety committee counted 37 car accidents in 18 months for the Uber autopilot test car between 2018 and 2019. Therefore, the accurate and effective driver distraction behavior detection system is designed to have important significance for improving traffic safety.
The driver distraction detection methods can be classified into the following three categories: based on the driver physiological information, the driving operation information, and the visual information. When the mental state of a driver changes, the physiological signal of the driver also changes, however, most of the physiological acquisition sensors need to be worn to the corresponding position of the body of the driver, and the driving experience is influenced. The driver state identification method based on the operation behaviors mainly utilizes the driver to acquire operation information of a steering wheel, an accelerator and a brake pedal, analyzes the driving behaviors of the driver in different states and presumes whether the driver is in a dangerous driving state. However, the recognition accuracy of the method is often influenced by the operating habits, skills, traffic road conditions and other factors of the driver. The vision-based detection method can non-invasively extract visual image information of the driver and is not affected by external interference. Therefore, the visual characteristics are the most widely used information in the driver distraction detection method. Vision-based detection methods can be divided into two categories: the first method directly classifies the original image to detect the state and behavior of the driver, and the method is often interfered by other factors in the image besides the driver in the image; the second method utilizes a target detection or image segmentation model to extract key areas or features such as hands, heads, upper bodies and the like from a driver image, and then inputs the extracted information into an identification model to obtain a detection result, however, the positioning of the areas or the features is often limited by the accuracy of an algorithm and is often subjected to false detection.
Disclosure of Invention
In order to overcome the defects of the technologies, the invention provides the driver attention dispersion detection method based on the image area positioning mechanism, solves the problem of automatic extraction of key positions and features in the driver behavior detection method based on the image information on the premise of not increasing the complexity of the model, and improves the detection precision of the model.
The technical scheme adopted by the invention for overcoming the technical problems is as follows:
a driver distraction detection method based on an image area positioning mechanism comprises the following steps:
a) acquiring visual images of different behaviors of a driver, and determining key areas needing attention in the different behavior states in an automatic positioning and manual adjusting mode according to the different behavior states of the driver in each visual image;
b) establishing a probability heat map of a key area in a visual image of a driver by using a Gaussian model, and establishing a driver behavior detection data set based on area positioning;
c) establishing a neural network model, constructing cost functions driven by class activation mapping and key region probability heat maps, and training a neural network by using the cost functions to obtain an optimized neural network model;
d) and installing a camera in the vehicle, acquiring a real-time image of the side part of the driver, inputting the image into the optimized neural network model, and extracting the output probability of the model to obtain the behavior state of the driver.
Further, a camera is installed in the vehicle in the step a), videos of different behaviors of the driver are collected through the camera, the videos are converted into visual images frame by frame, and the visual images are stored to obtain sample images.
Further, the behavior states of the driver in step a) are respectively defined as: a normal driving state, a state of using a smart phone or a tablet computer, a calling state, a conversation with a co-driver state, a drinking state and an operation center control electronic equipment state; when the driver is in a normal driving state, key areas needing attention are positioned on the hands and the upper arms of the driver in the visual image; when the driver is in a state of using the smart phone or the tablet personal computer, the key area needing attention is located in the mobile phone or the tablet personal computer of the hand of the driver in the visual image; when the driver is in a calling state, the key area needing attention is positioned at the mouth and the mobile phone position of the driver in the visual image; when the driver is in a conversation state with the co-driver, key areas needing attention are positioned at the mouth and the face of the driver in the visual image; when the driver is in a drinking state, the key area needing attention is positioned in a container held by the driver in the visual image; when the driver is in the state of operating the central control electronic equipment, the key area needing attention is positioned at the hand of the driver and the central control equipment in the visual image.
Further, the step of step a) comprises:
a-1) finding out the limb movement area of a driver in the process of executing different behaviors in a sample image, and establishing a key area based on different behavior states in the driver image;
a-2) based on the established key area, automatically acquiring the position information of an upper arm skeleton point and a head skeleton point of a driver in a sample image by a skeleton point positioning method, drawing a rectangular frame based on the upper arm skeleton point and the head skeleton point, wherein the skeleton point is positioned at the center of the rectangular frame, and obtaining the initial position of the key area of the image;
and a-3) manually correcting to obtain a final key area of the image according to the position and the size of the rectangular frame.
Further, step b) comprises the following steps:
b-1) based on the key region, by formula
Figure 524261DEST_PATH_IMAGE001
Establishing a two-dimensional Gaussian model
Figure 571851DEST_PATH_IMAGE002
In the formula
Figure 694659DEST_PATH_IMAGE003
In order to normalize the factors, the method comprises the steps of,
Figure 562121DEST_PATH_IMAGE004
in the form of a covariance matrix,
Figure 195621DEST_PATH_IMAGE005
in order to be transposed, the device is provided with a plurality of groups of parallel connection terminals,
Figure 379609DEST_PATH_IMAGE006
are variables of the two-dimensional gaussian model,
Figure 922586DEST_PATH_IMAGE007
for key locations in the driver behavior image,
Figure 57770DEST_PATH_IMAGE008
Figure 977184DEST_PATH_IMAGE009
is the abscissa of the critical area and is,
Figure 750099DEST_PATH_IMAGE010
is the ordinate of the key area and is,
Figure 463977DEST_PATH_IMAGE011
is composed of
Figure 89387DEST_PATH_IMAGE009
The medium maximum value is the maximum value of the average,
Figure 281334DEST_PATH_IMAGE012
is composed of
Figure 908756DEST_PATH_IMAGE009
The minimum value of the sum of the average values,
Figure 793535DEST_PATH_IMAGE013
is composed of
Figure 637732DEST_PATH_IMAGE010
The medium maximum value is the maximum value of the average,
Figure 915261DEST_PATH_IMAGE014
is composed of
Figure 380877DEST_PATH_IMAGE010
A medium to minimum value;
b-2) two-dimensional Gaussian model
Figure 930897DEST_PATH_IMAGE002
Converting the image into a two-dimensional image to obtain a probability heat map of a key area in a visual image of the driver;
b-3) traversing all image samples in the driver behavior detection data set based on the area positioning, and repeatedly executing the steps b-1) to b-2), and storing probability heat maps of key areas in the visual images of all drivers to obtain the driver behavior detection data set based on the area positioning.
Further, step c) comprises the steps of:
c-1) establishing a ResNeXt neural network model, and adopting a global average pooling layer on the top layer of the neural network;
c-2) establishing a SoftMax classifier at the top layer of the global pooling layer to output a driver behavior prediction probability value;
c-3) by the formula
Figure 481964DEST_PATH_IMAGE015
Calculating the class activation function of each driver behavior state class output by the top layer of the neural network,
Figure 828763DEST_PATH_IMAGE016
is composed of
Figure 414465DEST_PATH_IMAGE017
A heat map of the class is generated,
Figure 890314DEST_PATH_IMAGE018
the number of neurons in the top layer is,
Figure 741727DEST_PATH_IMAGE019
as a top-level weight parameter,
Figure 344747DEST_PATH_IMAGE020
mapping values of a previous layer of the global mean pooling layer;
c-4) extracting the driver behavior prediction probability value and the class activation mapping of the neural network model through a formula
Figure 37152DEST_PATH_IMAGE021
Calculating to obtain a region enhanced optimization function
Figure 700215DEST_PATH_IMAGE022
In the formula (I), wherein,
Figure 507765DEST_PATH_IMAGE023
in order to be a function of the non-linear transformation,
Figure 694902DEST_PATH_IMAGE024
is a neural network
Figure 989617DEST_PATH_IMAGE017
The class activation mapping of the class behavior state class,
Figure 574313DEST_PATH_IMAGE025
as of the same class as the true behavioral state class
Figure 321689DEST_PATH_IMAGE017
The class activates the mapping and the class activates the mapping,
Figure 784288DEST_PATH_IMAGE026
in the same class as the real category of behavior,
Figure 199089DEST_PATH_IMAGE027
as a function of the number of the coefficients,
Figure 954686DEST_PATH_IMAGE028
the matrix is a product of the hadamard,
Figure 454938DEST_PATH_IMAGE029
is a predefined one
Figure 718298DEST_PATH_IMAGE017
Class activation mapping;
c-5) by the formula
Figure 722026DEST_PATH_IMAGE030
Calculating a cost function
Figure 914104DEST_PATH_IMAGE031
Figure 636072DEST_PATH_IMAGE032
For the cost function based on the driver state value,
Figure 440474DEST_PATH_IMAGE033
Figure 130529DEST_PATH_IMAGE034
as a function of the number of the coefficients,
Figure 605373DEST_PATH_IMAGE035
for the output values of the resenext neural network model,
Figure 476115DEST_PATH_IMAGE036
is a calibration value;
c-6) passing a cost function
Figure 1905DEST_PATH_IMAGE031
Training the ResNeXt neural network model until convergence, and establishing the hyper-parameters of the model through cross validation.
Further, a camera is arranged at the position of the roof above the right of the driver in the step d).
Further, step d) comprises the following steps:
d-1) reading the ResNeXt neural network model trained in c-6) as a detection model;
d-2) inputting each frame of image of the driver acquired by the camera into the detection model;
d-3) obtaining the prediction probability value in a SoftMax classifier at the top layer of the ResNeXt neural network model, and identifying the current behavior state of the driver.
The invention has the beneficial effects that: the method comprises the steps of obtaining regions needing attention of a driver in different behavior states in an image through a manual calibration method, combining the regions with neural network type activation mapping, establishing a model optimization function based on region enhancement driving, training a neural network through the optimization function, enabling a detection model to automatically obtain key regions in the image of the driver according to different behavior characteristics of the driver in the detection process, solving the problem of automatic extraction of key features and positions in a detection method based on visual features, and improving the identification precision of the model.
Drawings
FIG. 1 is a flow chart of a method of the present invention;
FIG. 2 is a flow chart of an embodiment of the present invention.
Detailed Description
The invention will be further explained with reference to fig. 1 and 2.
As shown in the attached drawings, a method for detecting distraction of a driver based on an image area positioning mechanism comprises the following steps:
a) the method comprises the steps of collecting visual images of different behaviors of a driver, and determining key areas needing attention in the different behavior states in an automatic positioning and manual adjusting mode according to different behavior states of the driver in each visual image.
b) And establishing a probability heat map of a key area in a visual image of the driver by using a Gaussian model, and establishing a driver behavior detection data set based on area positioning.
c) Establishing a neural network model, constructing cost functions driven by class activation mapping and key region probability heat maps, and training a neural network by using the cost functions to obtain the optimized neural network model.
d) And installing a camera in the vehicle, acquiring a real-time image of the side part of the driver, inputting the image into the optimized neural network model, and extracting the output probability of the model to obtain the behavior state of the driver.
As shown in the attached figure 2, the regions needing attention in different behavior states of the driver in the image are obtained through a manual calibration method, the regions are combined with neural network activation mapping, a model optimization function based on region enhancement driving is established, the neural network is trained through the optimization function, so that the detection model can automatically obtain key regions in the image of the driver according to different behavior characteristics of the driver in the detection process, the problem of automatic extraction of key features and positions in the detection method based on visual features is solved, and the identification precision of the model is improved.
Specifically, a camera is installed in the vehicle in the step a), videos of different behaviors of the driver are collected through the camera, the videos are converted into visual images frame by frame, and the visual images are stored to obtain sample images.
Specifically, the behavior states of the driver in step a) are respectively defined as: a normal driving state, a state of using a smart phone or a tablet computer, a calling state, a conversation with a co-driver state, a drinking state and an operation center control electronic equipment state; when the driver is in a normal driving state, key areas needing attention are positioned on the hands and the upper arms of the driver in the visual image; when the driver is in a state of using the smart phone or the tablet personal computer, the key area needing attention is located in the mobile phone or the tablet personal computer of the hand of the driver in the visual image; when the driver is in a calling state, the key area needing attention is positioned at the mouth and the mobile phone position of the driver in the visual image; when the driver is in a conversation state with the co-driver, key areas needing attention are positioned at the mouth and the face of the driver in the visual image; when the driver is in a drinking state, the key area needing attention is positioned in a container held by the driver in the visual image; when the driver is in the state of operating the central control electronic equipment, the key area needing attention is positioned at the hand of the driver and the central control equipment in the visual image.
Specifically, the step a) is as follows:
a-1) finding out the limb movement area of the driver in the process of executing different behaviors in the sample image, and establishing key areas based on different behavior states in the driver image.
a-2) based on the established key area, automatically obtaining the position information of the upper arm skeleton point and the head skeleton point of the driver in the sample image by a skeleton point positioning method, drawing a rectangular frame based on the upper arm skeleton point and the head skeleton point, and obtaining the initial position of the key area of the image, wherein the skeleton point is positioned at the center of the rectangular frame. Let the size of the image behÍbThe dimensions of the initial rectangular frame are set as follows: the height of the rectangular frame in the normal driving image ishA width ofb(ii)/4; the height of the rectangular frame of the driver in the image of the smart phone or the tablet computer ishA width ofb(iii)/5; the height of the rectangular frame in the image of the driver's call ishA width ofb(iii)/5; the rectangular frame of the driver in the image of the conversation with the copilot is h/3 in height and h/3 in widthb(iii)/5; the height of the rectangular frame of the driver in the drinking image ishA width ofb(iii)/5; the height of the rectangular frame in the image of the electronic equipment controlled in the driving operation ishA width ofb/5。
a-3) manually correcting to obtain a rectangular frame of which the image finally contains the key area according to the position and the size of the rectangular frame. Preferably, the ratio of the finally corrected rectangular frame in the key area is equal to or greater than 90% and the rectangular frame is equal to or less than 1/2 of the sample image. Based on the above principle, the size range of the final rectangular frame after manual correction is as follows: the range of the rectangular box height in the normal driving image is:h/9 ~ h/7, wide range ofb/6 ~ bA/3; the range of the height of the rectangular frame of the image of the driver using the smart phone or the tablet computer is as follows:h/6 ~ ha broad range ofb/6 ~ bA/3; the range of the height of the rectangular frame in the image of the driver making a call is:h/6 ~ h/4, wide range ofb/8 ~ b6; the range of the height of the rectangular frame in the image of the driver talking with the co-driver is: h4 to h/3, the wide range isb/6 ~ b(iii)/5; the range of the height of the rectangular frame of the driver in the drinking image is as follows:h/7 ~ ha broad range ofb/8 ~ b(iii)/5; the range of the height of the rectangular frame in the driving operation central control electronic equipment image is as follows: h/3 to h/2, the wide range isb/6 ~ b/4。
Specifically, the step b) comprises the following steps:
b-1) based on the key region, by formula
Figure 519474DEST_PATH_IMAGE001
Establishing a two-dimensional Gaussian model
Figure 950455DEST_PATH_IMAGE002
In the formula
Figure 831080DEST_PATH_IMAGE003
In order to normalize the factors, the method comprises the steps of,
Figure 195065DEST_PATH_IMAGE004
in the form of a covariance matrix,
Figure 368689DEST_PATH_IMAGE005
in order to be transposed, the device is provided with a plurality of groups of parallel connection terminals,
Figure 552545DEST_PATH_IMAGE006
are variables of the two-dimensional gaussian model,
Figure 968352DEST_PATH_IMAGE007
for key locations in the driver behavior image,
Figure 186844DEST_PATH_IMAGE008
Figure 796948DEST_PATH_IMAGE009
is the abscissa of the critical area and is,
Figure 468101DEST_PATH_IMAGE010
is the ordinate of the key area and is,
Figure 690528DEST_PATH_IMAGE011
is composed of
Figure 763526DEST_PATH_IMAGE009
The medium maximum value is the maximum value of the average,
Figure 278952DEST_PATH_IMAGE012
is composed of
Figure 437401DEST_PATH_IMAGE009
The minimum value of the sum of the average values,
Figure 460590DEST_PATH_IMAGE013
is composed of
Figure 388094DEST_PATH_IMAGE010
The medium maximum value is the maximum value of the average,
Figure 323689DEST_PATH_IMAGE014
is composed of
Figure 720167DEST_PATH_IMAGE010
The medium minimum value.
b-2) two-dimensional Gaussian model
Figure 32199DEST_PATH_IMAGE002
And converting the image into a two-dimensional image to obtain a probability heat map of a key area in the visual image of the driver.
b-3) traversing all image samples in the driver behavior detection data set based on the area positioning, and repeatedly executing the steps b-1) to b-2), and storing probability heat maps of key areas in the visual images of all drivers to obtain the driver behavior detection data set based on the area positioning.
Specifically, the step c) comprises the following steps:
c-1) establishing a ResNeXt neural network model with the network layer number of 50, and adopting a global average pooling layer at the top layer of the neural network.
And c-2) establishing a SoftMax classifier at the top layer of the global pooling layer to output the predicted probability value of the driver behavior.
c-3) by the formula
Figure 331987DEST_PATH_IMAGE015
Calculating the class activation function of each driver behavior state class output by the top layer of the neural network,
Figure 704063DEST_PATH_IMAGE016
is composed of
Figure 587836DEST_PATH_IMAGE017
A heat map of the class is generated,
Figure 703560DEST_PATH_IMAGE018
the number of neurons in the top layer is,
Figure 589345DEST_PATH_IMAGE019
as a top-level weight parameter,
Figure 132322DEST_PATH_IMAGE020
and the mapping value of the previous layer of the global mean pooling layer.
c-4) extracting the driver behavior prediction probability value and the class activation mapping of the neural network model through a formula
Figure 503391DEST_PATH_IMAGE021
Calculating to obtain a region enhanced optimization function
Figure 891647DEST_PATH_IMAGE022
In the formula (I), wherein,
Figure 166027DEST_PATH_IMAGE023
in order to be a function of the non-linear transformation,
Figure 879905DEST_PATH_IMAGE024
is a neural network
Figure 738271DEST_PATH_IMAGE017
The class activation mapping of the class behavior state class,
Figure 195797DEST_PATH_IMAGE025
as of the same class as the true behavioral state class
Figure 321754DEST_PATH_IMAGE017
The class activates the mapping and the class activates the mapping,
Figure 940954DEST_PATH_IMAGE026
in the same class as the real category of behavior,
Figure 21037DEST_PATH_IMAGE027
as a function of the number of the coefficients,
Figure 282254DEST_PATH_IMAGE028
the matrix is a product of the hadamard,
Figure 289084DEST_PATH_IMAGE029
is a predefined one
Figure 79186DEST_PATH_IMAGE017
The class activates the mapping.
c-5) by the formula
Figure 646564DEST_PATH_IMAGE030
Calculating a cost function
Figure 445893DEST_PATH_IMAGE031
Figure 766016DEST_PATH_IMAGE032
For a cost function based on the driver state value, the traditional ResNeXt network uses a cost function of
Figure 773024DEST_PATH_IMAGE032
Figure 811387DEST_PATH_IMAGE033
Figure 165139DEST_PATH_IMAGE034
As a function of the number of the coefficients,
Figure 123124DEST_PATH_IMAGE035
for the output values of the resenext neural network model,
Figure 520608DEST_PATH_IMAGE036
is a calibrated value.
c-6) passing a cost function
Figure 46267DEST_PATH_IMAGE031
Training the ResNeXt neural network model until convergence, and establishing the hyper-parameters of the model through cross validation. The final defined hyper-parameters are mainly: the neural network trains the learning rate, the number of batch training samples (batch size), the coefficients in the loss function, the number of batch samples, the momentum parameter β of the momentum optimizer.
Preferably, in this patent, the learning rate r = 0.001 for neural network training, and the number of batch training samples: blocksize =32, coefficients in loss function
Figure 672551DEST_PATH_IMAGE034
=0.3, momentum parameter β =0.9 of the momentum optimizer.
Preferably, a camera is installed at the position of the roof right above the driver in the step d).
Specifically, the step d) comprises the following steps:
d-1) reading the trained ResNeXt neural network model in c-6) as a detection model.
d-2) inputting each frame of image of the driver acquired by the camera into the detection model.
d-3) obtaining the prediction probability value in a SoftMax classifier at the top layer of the ResNeXt neural network model, and identifying the current behavior state of the driver.
In order to verify that the detection precision is improved by the driver distraction detection method based on the image area positioning mechanism, a driver behavior data set is constructed through a real-vehicle experiment, and the data set comprises 12688 driver images of 6 behaviors including a normal driving state, a state of using a smart phone or a tablet personal computer, a state of making a call, a state of talking with a co-driver, a water drinking state and a state of operating a central control electronic device, wherein 40 drivers, 10 females among 40 drivers and 30 males. If the data set is input into the ResNeXt model trained by the traditional training method, the model recognition accuracy is only 89.75% taking the ResNeXt model with 50 layers as an example. If the data set is input into the ResNeXt model trained by the invention, the model identification accuracy can reach 95.59% by taking a 50-layer ResNeXt model as an example.
In order to verify the accuracy of the driver distraction detection method based on the image area positioning mechanism in the patent, a driver behavior data set is constructed through a real-vehicle experiment, and the data set comprises 12688 driver images of 6 behaviors including a normal driving state, a state of using a smart phone or a tablet computer, a state of making a call, a state of talking with a copilot, a water drinking state and a state of operating central control electronic equipment, wherein 40 drivers, 10 females and 30 males among 40 drivers. The experiment takes a ResNeXt model with 50 layers as an example, training and verification are carried out by relying on the data set, the experimental result is shown in table 1, and C0-C5 in table 1 respectively represent 6 behaviors of a normal driving state, a state of using a smart phone or a tablet computer, a state of calling, a state of talking with a co-driver, a state of drinking water and a state of operating central control electronic equipment. Compared with the traditional training method, the recognition accuracy of the ResNeXt model can be effectively improved by the training method provided by the patent through experimental results.
TABLE 1
Figure 232846DEST_PATH_IMAGE037
Finally, it should be noted that: although the present invention has been described in detail with reference to the foregoing embodiments, it will be apparent to those skilled in the art that changes may be made in the embodiments and/or equivalents thereof without departing from the spirit and scope of the invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (8)

1. A driver distraction detection method based on an image area positioning mechanism is characterized by comprising the following steps:
a) acquiring visual images of different behaviors of a driver, and determining key areas needing attention in the different behavior states in an automatic positioning and manual adjusting mode according to the different behavior states of the driver in each visual image;
b) establishing a probability heat map of a key area in a visual image of a driver by using a Gaussian model, and establishing a driver behavior detection data set based on area positioning;
c) establishing a neural network model, constructing cost functions driven by class activation mapping and key region probability heat maps, and training a neural network by using the cost functions to obtain an optimized neural network model;
d) and installing a camera in the vehicle, acquiring a real-time image of the side part of the driver, inputting the image into the optimized neural network model, and extracting the output probability of the model to obtain the behavior state of the driver.
2. The method for detecting the distraction of the driver based on the image area localization mechanism according to claim 1, wherein: in the step a), a camera is installed in the vehicle, videos of different behaviors of a driver are collected through the camera, the videos are converted into visual images frame by frame, and the visual images are stored to obtain sample images.
3. The method for detecting the distraction of the driver based on the image area positioning mechanism according to claim 1, wherein the behavior states of the driver in the step a) are respectively defined as: a normal driving state, a state of using a smart phone or a tablet computer, a calling state, a conversation with a co-driver state, a drinking state and an operation center control electronic equipment state; when the driver is in a normal driving state, key areas needing attention are positioned on the hands and the upper arms of the driver in the visual image; when the driver is in a state of using the smart phone or the tablet personal computer, the key area needing attention is located in the mobile phone or the tablet personal computer of the hand of the driver in the visual image; when the driver is in a calling state, the key area needing attention is positioned at the mouth and the mobile phone position of the driver in the visual image; when the driver is in a conversation state with the co-driver, key areas needing attention are positioned at the mouth and the face of the driver in the visual image; when the driver is in a drinking state, the key area needing attention is positioned in a container held by the driver in the visual image; when the driver is in the state of operating the central control electronic equipment, the key area needing attention is positioned at the hand of the driver and the central control equipment in the visual image.
4. The method for detecting the distraction of the driver based on the image area positioning mechanism according to claim 3, wherein the step a) comprises:
a-1) finding out the limb movement area of a driver in the process of executing different behaviors in a sample image, and establishing a key area based on different behavior states in the driver image;
a-2) based on the established key area, automatically acquiring the position information of an upper arm skeleton point and a head skeleton point of a driver in a sample image by a skeleton point positioning method, drawing a rectangular frame based on the upper arm skeleton point and the head skeleton point, wherein the skeleton point is positioned at the center of the rectangular frame, and obtaining the initial position of the key area of the image;
a-3) manually correcting to obtain a rectangular frame of which the image finally contains the key area according to the position and the size of the rectangular frame.
5. The method for detecting the distraction of the driver based on the image area positioning mechanism according to claim 1, wherein the step b) comprises the following steps:
b-1) based on the key region, by formula
Figure 132274DEST_PATH_IMAGE001
Establishing a two-dimensional Gaussian model
Figure 72549DEST_PATH_IMAGE002
In the formula
Figure 782885DEST_PATH_IMAGE003
In order to normalize the factors, the method comprises the steps of,
Figure 150412DEST_PATH_IMAGE004
in the form of a covariance matrix,
Figure 313409DEST_PATH_IMAGE005
in order to be transposed, the device is provided with a plurality of groups of parallel connection terminals,
Figure 791795DEST_PATH_IMAGE006
are variables of the two-dimensional gaussian model,
Figure 93988DEST_PATH_IMAGE007
for key locations in the driver behavior image,
Figure 632416DEST_PATH_IMAGE008
Figure 282709DEST_PATH_IMAGE009
is the abscissa of the critical area and is,
Figure 830365DEST_PATH_IMAGE010
is the ordinate of the key area and is,
Figure 249714DEST_PATH_IMAGE011
is composed of
Figure 959044DEST_PATH_IMAGE009
The medium maximum value is the maximum value of the average,
Figure 362213DEST_PATH_IMAGE012
is composed of
Figure 447980DEST_PATH_IMAGE009
The minimum value of the sum of the average values,
Figure 724765DEST_PATH_IMAGE013
is composed of
Figure 339418DEST_PATH_IMAGE010
The medium maximum value is the maximum value of the average,
Figure 698724DEST_PATH_IMAGE014
is composed of
Figure 322603DEST_PATH_IMAGE010
A medium to minimum value;
b-2) two-dimensional Gaussian model
Figure 716544DEST_PATH_IMAGE002
Converting the image into a two-dimensional image to obtain a probability heat map of a key area in a visual image of the driver;
b-3) traversing all image samples in the driver behavior detection data set based on the area positioning, and repeatedly executing the steps b-1) to b-2), and storing probability heat maps of key areas in the visual images of all drivers to obtain the driver behavior detection data set based on the area positioning.
6. The method for detecting the distraction of the driver based on the image area positioning mechanism according to claim 1, wherein the step c) comprises the following steps:
c-1) establishing a ResNeXt neural network model, and adopting a global average pooling layer on the top layer of the neural network;
c-2) establishing a SoftMax classifier at the top layer of the global pooling layer to output a driver behavior prediction probability value;
c-3) by the formula
Figure 502097DEST_PATH_IMAGE015
Calculating the class activation function of each driver behavior state class output by the top layer of the neural network,
Figure 879858DEST_PATH_IMAGE016
is composed of
Figure 41849DEST_PATH_IMAGE017
A heat map of the class is generated,
Figure 293226DEST_PATH_IMAGE018
the number of neurons in the top layer is,
Figure 780840DEST_PATH_IMAGE019
as a top-level weight parameter,
Figure 380317DEST_PATH_IMAGE020
mapping values of a previous layer of the global mean pooling layer;
c-4) extracting the driver behavior prediction probability value and the class activation mapping of the neural network model through a formula
Figure 1791DEST_PATH_IMAGE021
Calculating to obtain a region enhanced optimization function
Figure 121057DEST_PATH_IMAGE022
In the formula (I), wherein,
Figure 28839DEST_PATH_IMAGE023
in order to be a function of the non-linear transformation,
Figure 397503DEST_PATH_IMAGE024
is a neural network
Figure 989432DEST_PATH_IMAGE017
The class activation mapping of the class behavior state class,
Figure 697625DEST_PATH_IMAGE025
as of the same class as the true behavioral state class
Figure 510729DEST_PATH_IMAGE017
The class activates the mapping and the class activates the mapping,
Figure 569952DEST_PATH_IMAGE026
in the same class as the real category of behavior,
Figure 861125DEST_PATH_IMAGE027
as a function of the number of the coefficients,
Figure 220562DEST_PATH_IMAGE028
the matrix is a product of the hadamard,
Figure 611092DEST_PATH_IMAGE029
is a predefined one
Figure 409808DEST_PATH_IMAGE017
Class activation mapping;
c-5) by the formula
Figure 114459DEST_PATH_IMAGE030
Calculating a cost function
Figure 453036DEST_PATH_IMAGE031
Figure 358676DEST_PATH_IMAGE032
For the cost function based on the driver state value,
Figure 641758DEST_PATH_IMAGE033
Figure 150100DEST_PATH_IMAGE034
as a function of the number of the coefficients,
Figure 218550DEST_PATH_IMAGE035
for the output values of the resenext neural network model,
Figure 544358DEST_PATH_IMAGE036
is a calibration value;
c-6) passing a cost function
Figure 799890DEST_PATH_IMAGE031
Training the ResNeXt neural network model until convergence, and establishing the hyper-parameters of the model through cross validation.
7. The method for detecting the distraction of the driver based on the image area localization mechanism according to claim 1, wherein: and d) mounting a camera at the position of the roof above the right of the driver.
8. The method for detecting the distraction of the driver based on the image area positioning mechanism according to claim 6, wherein the step d) comprises the following steps:
d-1) reading the ResNeXt neural network model trained in c-6) as a detection model;
d-2) inputting each frame of image of the driver acquired by the camera into the detection model;
d-3) obtaining the prediction probability value in a SoftMax classifier at the top layer of the ResNeXt neural network model, and identifying the current behavior state of the driver.
CN202111110059.2A 2021-09-23 2021-09-23 Driver attention dispersion detection method based on image area positioning mechanism Active CN113569817B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111110059.2A CN113569817B (en) 2021-09-23 2021-09-23 Driver attention dispersion detection method based on image area positioning mechanism

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111110059.2A CN113569817B (en) 2021-09-23 2021-09-23 Driver attention dispersion detection method based on image area positioning mechanism

Publications (2)

Publication Number Publication Date
CN113569817A true CN113569817A (en) 2021-10-29
CN113569817B CN113569817B (en) 2021-12-21

Family

ID=78173969

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111110059.2A Active CN113569817B (en) 2021-09-23 2021-09-23 Driver attention dispersion detection method based on image area positioning mechanism

Country Status (1)

Country Link
CN (1) CN113569817B (en)

Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014167811A1 (en) * 2013-04-10 2014-10-16 株式会社デンソー Drowsiness prediction device and drowsiness prediction system
CN105809152A (en) * 2016-04-06 2016-07-27 清华大学 Monitoring method for cognitive distraction of driver on basis of multi-source information fusion
CN106778677A (en) * 2016-12-30 2017-05-31 东北农业大学 Feature based selection and driver's fatigue state recognition method and device of facial multizone combining classifiers
CN108227912A (en) * 2017-11-30 2018-06-29 北京市商汤科技开发有限公司 Apparatus control method and device, electronic equipment, computer storage media
CN108960065A (en) * 2018-06-01 2018-12-07 浙江零跑科技有限公司 A kind of driving behavior detection method of view-based access control model
CN108985259A (en) * 2018-08-03 2018-12-11 百度在线网络技术(北京)有限公司 Human motion recognition method and device
CN109583338A (en) * 2018-11-19 2019-04-05 山东派蒙机电技术有限公司 Driver Vision decentralized detection method based on depth integration neural network
CN109711463A (en) * 2018-12-25 2019-05-03 广东顺德西安交通大学研究院 Important object detection method based on attention
CN110119676A (en) * 2019-03-28 2019-08-13 广东工业大学 A kind of Driver Fatigue Detection neural network based
CN110288597A (en) * 2019-07-01 2019-09-27 哈尔滨工业大学 Wireless capsule endoscope saliency detection method based on attention mechanism
CN110298257A (en) * 2019-06-04 2019-10-01 东南大学 A kind of driving behavior recognition methods based on human body multiple location feature
CN110728185A (en) * 2019-09-10 2020-01-24 西安工业大学 Detection method for judging existence of handheld mobile phone conversation behavior of driver
WO2020122986A1 (en) * 2019-06-10 2020-06-18 Huawei Technologies Co.Ltd. Driver attention detection using heat maps
CN111563468A (en) * 2020-05-13 2020-08-21 电子科技大学 Driver abnormal behavior detection method based on attention of neural network
CN111860525A (en) * 2020-08-06 2020-10-30 宁夏宁电电力设计有限公司 Bottom-up optical character recognition method suitable for terminal block
CN111914902A (en) * 2020-07-08 2020-11-10 南京航空航天大学 Traditional Chinese medicine identification and surface defect detection method based on deep neural network
CN112069988A (en) * 2020-09-04 2020-12-11 徐尔灵 Gun-ball linkage-based driver safe driving behavior detection method
CN112241679A (en) * 2020-09-14 2021-01-19 浙江理工大学 Automatic garbage classification method
CN112418261A (en) * 2020-09-17 2021-02-26 电子科技大学 Human body image multi-attribute classification method based on prior prototype attention mechanism
CN112419670A (en) * 2020-09-15 2021-02-26 深圳市点创科技有限公司 Method, device and medium for detecting fatigue driving of driver by fusing key point positioning and image classification
CN113051958A (en) * 2019-12-26 2021-06-29 深圳市光鉴科技有限公司 Driver state detection method, system, device and medium based on deep learning
CN113313199A (en) * 2021-06-21 2021-08-27 北京工业大学 Brain CT medical report automatic generation method based on weak supervision attention

Patent Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014167811A1 (en) * 2013-04-10 2014-10-16 株式会社デンソー Drowsiness prediction device and drowsiness prediction system
CN105809152A (en) * 2016-04-06 2016-07-27 清华大学 Monitoring method for cognitive distraction of driver on basis of multi-source information fusion
CN106778677A (en) * 2016-12-30 2017-05-31 东北农业大学 Feature based selection and driver's fatigue state recognition method and device of facial multizone combining classifiers
CN108227912A (en) * 2017-11-30 2018-06-29 北京市商汤科技开发有限公司 Apparatus control method and device, electronic equipment, computer storage media
CN108960065A (en) * 2018-06-01 2018-12-07 浙江零跑科技有限公司 A kind of driving behavior detection method of view-based access control model
CN108985259A (en) * 2018-08-03 2018-12-11 百度在线网络技术(北京)有限公司 Human motion recognition method and device
CN109583338A (en) * 2018-11-19 2019-04-05 山东派蒙机电技术有限公司 Driver Vision decentralized detection method based on depth integration neural network
CN109711463A (en) * 2018-12-25 2019-05-03 广东顺德西安交通大学研究院 Important object detection method based on attention
CN110119676A (en) * 2019-03-28 2019-08-13 广东工业大学 A kind of Driver Fatigue Detection neural network based
CN110298257A (en) * 2019-06-04 2019-10-01 东南大学 A kind of driving behavior recognition methods based on human body multiple location feature
WO2020122986A1 (en) * 2019-06-10 2020-06-18 Huawei Technologies Co.Ltd. Driver attention detection using heat maps
CN110288597A (en) * 2019-07-01 2019-09-27 哈尔滨工业大学 Wireless capsule endoscope saliency detection method based on attention mechanism
CN110728185A (en) * 2019-09-10 2020-01-24 西安工业大学 Detection method for judging existence of handheld mobile phone conversation behavior of driver
CN113051958A (en) * 2019-12-26 2021-06-29 深圳市光鉴科技有限公司 Driver state detection method, system, device and medium based on deep learning
CN111563468A (en) * 2020-05-13 2020-08-21 电子科技大学 Driver abnormal behavior detection method based on attention of neural network
CN111914902A (en) * 2020-07-08 2020-11-10 南京航空航天大学 Traditional Chinese medicine identification and surface defect detection method based on deep neural network
CN111860525A (en) * 2020-08-06 2020-10-30 宁夏宁电电力设计有限公司 Bottom-up optical character recognition method suitable for terminal block
CN112069988A (en) * 2020-09-04 2020-12-11 徐尔灵 Gun-ball linkage-based driver safe driving behavior detection method
CN112241679A (en) * 2020-09-14 2021-01-19 浙江理工大学 Automatic garbage classification method
CN112419670A (en) * 2020-09-15 2021-02-26 深圳市点创科技有限公司 Method, device and medium for detecting fatigue driving of driver by fusing key point positioning and image classification
CN112418261A (en) * 2020-09-17 2021-02-26 电子科技大学 Human body image multi-attribute classification method based on prior prototype attention mechanism
CN113313199A (en) * 2021-06-21 2021-08-27 北京工业大学 Brain CT medical report automatic generation method based on weak supervision attention

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
BOLEI ZHOU 等: "Learning Deep Features for Discriminative Localization", 《2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
HESHAMM. ERAQI 等: "Driver Distraction Identification with an Ensemble of Convolutional Neural Networks", 《JOURNAL OF ADVANCED TRANSPORTATION》 *
LEI ZHAO 等: "Driver behavior detection via adaptive spatial attention mechanism", 《ADVANCED ENGINEERING INFORMATICS》 *
SAINING XIE 等: "Aggregated Residual Transformations for Deep Neural Networks", 《2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
廖南星 等: "基于类激活映射-注意力机制的图像描述方法", 《山东大学学报(工学版)》 *
王超 等: "表征驾驶风格和驾驶员能力的驾驶员模型", 《北京理工大学学报》 *

Also Published As

Publication number Publication date
CN113569817B (en) 2021-12-21

Similar Documents

Publication Publication Date Title
CN111723748B (en) Infrared remote sensing image ship detection method
CN110059582B (en) Driver behavior identification method based on multi-scale attention convolution neural network
CN107563372B (en) License plate positioning method based on deep learning SSD frame
CN111460919B (en) Monocular vision road target detection and distance estimation method based on improved YOLOv3
CN110363182B (en) Deep learning-based lane line detection method
CN107133974B (en) Gaussian Background models the vehicle type classification method combined with Recognition with Recurrent Neural Network
CN101859382B (en) License plate detection and identification method based on maximum stable extremal region
CN113065578B (en) Image visual semantic segmentation method based on double-path region attention coding and decoding
CN110033002A (en) Detection method of license plate based on multitask concatenated convolutional neural network
CN108537197A (en) A kind of lane detection prior-warning device and method for early warning based on deep learning
CN111325146B (en) Truck type and axle type identification method and system
CN105354568A (en) Convolutional neural network based vehicle logo identification method
CN109345770A (en) A kind of child leaves in-vehicle alarm system and child leaves interior alarm method
CN111753674A (en) Fatigue driving detection and identification method based on deep learning
CN108694408B (en) Driving behavior recognition method based on deep sparse filtering convolutional neural network
CN109635784A (en) Traffic sign recognition method based on improved convolutional neural networks
CN110261329A (en) A kind of Minerals identification method based on full spectral coverage high-spectrum remote sensing data
CN111256693B (en) Pose change calculation method and vehicle-mounted terminal
CN110363093A (en) A kind of driver's action identification method and device
CN110852358A (en) Vehicle type distinguishing method based on deep learning
CN112766273A (en) License plate recognition method
CN110263836B (en) Bad driving state identification method based on multi-feature convolutional neural network
KR101050687B1 (en) Pedestrian recognition device and method using posterior probability and pedestrian protection device using it
CN114548216A (en) Online abnormal driving behavior identification method based on Encoder-Decoder attention network and LSTM
CN112052829B (en) Pilot behavior monitoring method based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant