CN108205649A - Driver drives to take the state identification method and device of phone - Google Patents

Driver drives to take the state identification method and device of phone Download PDF

Info

Publication number
CN108205649A
CN108205649A CN201611185468.8A CN201611185468A CN108205649A CN 108205649 A CN108205649 A CN 108205649A CN 201611185468 A CN201611185468 A CN 201611185468A CN 108205649 A CN108205649 A CN 108205649A
Authority
CN
China
Prior art keywords
driver
phone
head
region
shoulder region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201611185468.8A
Other languages
Chinese (zh)
Other versions
CN108205649B (en
Inventor
陈鑫嘉
张震
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Uniview Technologies Co Ltd
Original Assignee
Zhejiang Uniview Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Uniview Technologies Co Ltd filed Critical Zhejiang Uniview Technologies Co Ltd
Priority to CN201611185468.8A priority Critical patent/CN108205649B/en
Publication of CN108205649A publication Critical patent/CN108205649A/en
Application granted granted Critical
Publication of CN108205649B publication Critical patent/CN108205649B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The application provides a kind of driver and drives to take the state identification method and device of phone, the method includes:The vehicle window of target vehicle in monitoring image is positioned;According to the location information of vehicle window and the license board information of target vehicle, obtain driver and detect candidate region;Utilization orientation histogram of gradients and support vector machines detect candidate region to the driver and are detected, and obtain the head and shoulder region of the driver;The head and shoulder region is sequentially input into first layer CNN networks and second layer CNN networks, the first layer CNN networks tentatively screen the head and shoulder region, obtain the head and shoulder region of the doubtful driver for taking phone, the second layer CNN networks further screen the output result of the first layer CNN networks, obtain the state that the driver takes phone.The application can exclude it is non-genuine under complex scene take accidentally picking up for phone, improve accuracy of identification, adaptability, the robustness of scene are preferable.

Description

Driver drives to take the state identification method and device of phone
Technical field
This application involves field of video monitoring more particularly to a kind of driver drive to take phone state identification method and Device.
Background technology
If driver takes phone in driving process, traffic accident rate can be greatly increased, thus is needed to driving The state that the person's of sailing truck takes phone effectively identified, the important evidence whether broken rules and regulations as it.
With the continuous development of image processing techniques, computer vision technique, depth learning technology and embedded technology, How a research in current intelligent transportation is had become to vehicle (containing occupant) progress automatic discrimination violating the regulations and evidence obtaining Hot spot.
The prior art provide it is a kind of identify that driver phones with mobile telephone the method (CN105868690A) of behavior, acquisition first is driven The video flowing in cabin is sailed, then by face component model orientation human face region, then carries out face correction, uses nonlinear discriminant Relationship trains two groups of parameters to be made whether to make a phone call to differentiate, wherein train classification models include ear region training set, phone with mobile telephone Training set and training set of not phoning with mobile telephone.This method uses DPM (Deformable Part Model, deformable part model) portion Part detection algorithm carry out human face region positioning, detection very take, on block or the Detection accuracy of face ambiguity influence It is larger;This method is made whether to make a phone call to differentiate using Nonlinear Classification method, by training ear region, region of phoning with mobile telephone, Region of not phoning with mobile telephone is identified, and accuracy rate is relatively low.
The prior art also provides the automatic monitoring method that mobile phone behavior is taken in a kind of driver drives vehicle way (CN103366506A), by image acquiring device, driver head and near zone are obtained first, is obtained using Face Detection It to the position of driver face and hand in the picture, reuses support vector machines and classifies, dock the driver to phone with mobile telephone It gives a warning.This method obtains the position of driver face and hand in the picture using Face Detection, and driver's cabin imaging is multiple Miscellaneous, daylighting and special weather can be larger to Imaging in the vehicle window under complex scene, thus this method missing inspection, flase drop compared with It is more.
Invention content
In view of this, the application provides a kind of driver and drives to take the state identification method and device of phone, to solve The problem of state recognition accuracy rate that driver's driving in the prior art takes phone is relatively low.
Specifically, the application is achieved by the following technical solution:
It drives to take the state identification method of phone, the side in a first aspect, providing a kind of driver according to the application's Method includes:
The vehicle window of target vehicle in monitoring image is positioned;
According to the location information of vehicle window and the license board information of target vehicle, obtain driver and detect candidate region;
Utilization orientation histogram of gradients and support vector machines detect candidate region to the driver and are detected, and obtain institute State the head and shoulder region of driver;
The head and shoulder region is sequentially input into first layer CNN networks and second layer CNN networks, the first layer CNN networks The head and shoulder region is tentatively screened, obtains the head and shoulder region of the doubtful driver for taking phone, the second layer CNN nets Network further screens the output result of the first layer CNN networks, obtains the state that the driver takes phone.
Optionally, it before the head and shoulder region is inputted first layer CNN networks and second layer CNN networks, further includes:
The histograms of oriented gradients feature in the head and shoulder region and local binary feature are extracted, and the direction gradient is straight Square figure feature and local binary feature are combined, to form the feature vector of a multidimensional;
Classified using linear discriminant analysis to described eigenvector, filter out non-head and shoulder region.
Optionally, the method further includes:
It is left area, right area and whole area by the head and shoulder region division of the doubtful driver for taking phone Domain;
The left area, right area and overall region are inputted into second layer CNN networks respectively, obtain the driver The state of phone is taken, the state that the driver takes phone includes:The left side takes phone, the right takes number, does not engage electricity Words and can not penalty.
Optionally, the method further includes:
If present frame monitoring image recognition result is in for driver and takes telephone state, to present frame monitoring image Next frame monitoring image continues to identify, if the recognition result of the next frame monitoring image is in for driver takes phone shape State is then alerted;Otherwise, it abandons alerting.
Optionally, the location information acquisition process of the vehicle window includes:
According to the license board information of the target vehicle, vehicle window upper right comer region is obtained;
Utilize the location information of positioning filter positioning vehicle window upper right angle point.
According to the second aspect of the application, a kind of driver is provided and drives to take the status identification means of phone, the dress Put including:
Locating module positions the vehicle window of target vehicle in monitoring image;
Region acquisition module according to the location information of vehicle window and the license board information of target vehicle, obtains driver's detection Candidate region;
Module of target detection, utilization orientation histogram of gradients and support vector machines to the driver detect candidate region into Row detection obtains the head and shoulder region of the driver;
The head and shoulder region is sequentially input first layer CNN networks and second layer CNN networks by identification module, and described first Layer CNN networks tentatively screen the head and shoulder region, the head and shoulder region of the doubtful driver for taking phone of acquisition, and described the Two layers of CNN networks further screen the output result of the first layer CNN networks, obtain the shape that the driver takes phone State.
Optionally, the module of target detection further includes:
Feature extraction submodule extracts the histograms of oriented gradients feature in the head and shoulder region and local binary feature, and The histograms of oriented gradients feature and local binary feature are combined, to form the feature vector of a multidimensional;
Filter submodule classifies to described eigenvector using linear discriminant analysis, filters out non-head and shoulder region.
Optionally, the identification module further includes:
Submodule is divided, is left area, the right area by the head and shoulder region division of the doubtful driver for taking phone Domain and overall region;
The left area, right area and overall region are inputted second layer CNN nets by fusion recognition submodule respectively Network, obtains the state that the driver takes phone, and the state that the driver takes phone includes:The left side takes phone, the right side While it takes phone, make a phone call and can not penalty.
Optionally, described device further includes:
Multiframe authentication module, if present frame monitoring image recognition result is in for driver takes telephone state, to working as The next frame monitoring image of previous frame monitoring image continues to identify, if the recognition result of the next frame monitoring image is at driver In taking telephone state, then alerted;Otherwise, it abandons alerting.
Optionally, the locating module includes:
According to the license board information of the target vehicle, vehicle window upper right comer region is obtained;
Utilize the location information of positioning filter positioning vehicle window upper right angle point.
The advantageous effect of the application:By vehicle glazing zone location, driver's target detection (head and shoulder region detection) with And the step of cascade these pure video detections of CNN networks, the time-consuming consumption that artificial detection driving takes telephone state can be removed from Power, and can exclude under complex scene it is non-genuine take accidentally picking up for phone, improve accuracy of identification, be compared to tradition side Method, penalty accuracy rate higher, adaptability, robustness to scene are more preferably.
It should be understood that above general description and following detailed description are only exemplary and explanatory, not The application can be limited.
Description of the drawings
Attached drawing herein is incorporated into specification and forms the part of this specification, shows the implementation for meeting the application Example, and for explaining the principle of the application together with specification.
Fig. 1 is that a kind of driver provided by the embodiments of the present application drives to take the state identification method flow chart of phone;
Fig. 2 is a kind of flow chart of target detection provided by the embodiments of the present application;
Fig. 3 is a kind of flow chart of target identification provided by the embodiments of the present application;
Fig. 4 is that a kind of driver provided by the embodiments of the present application drives to take the status identification means structure diagram of phone;
Fig. 5 is the structure diagram of module of target detection provided by the embodiments of the present application;
Fig. 6 is the structure diagram of identification module provided by the embodiments of the present application;
Fig. 7 is that another driver provided by the embodiments of the present application drives to take the status identification means structural frames of phone Figure.
Specific embodiment
Here exemplary embodiment will be illustrated in detail, example is illustrated in the accompanying drawings.Following description is related to During attached drawing, unless otherwise indicated, the same numbers in different attached drawings represent the same or similar element.Following exemplary embodiment Described in embodiment do not represent all embodiments consistent with the application.On the contrary, they be only with it is such as appended The example of the consistent device and method of some aspects be described in detail in claims, the application.In addition, do not conflicting In the case of, the feature in following embodiment and embodiment can be combined with each other.
It is only merely for the purpose of description specific embodiment in term used in this application, and is not intended to be limiting the application. It is also intended in the application and " one kind " of singulative used in the attached claims, " described " and "the" including majority Form, unless context clearly shows that other meanings.It is also understood that term "and/or" used herein refers to and wraps Containing one or more associated list items purposes, any or all may be combined.
It will be appreciated that though various information, but this may be described using term first, second, third, etc. in the application A little information should not necessarily be limited by these terms.These terms are only used for same type of information being distinguished from each other out.For example, not departing from In the case of the application range, the first information can also be referred to as the second information, and similarly, the second information can also be referred to as One information.Depending on linguistic context, word as used in this " if " can be construed to " ... when " or " when ... When " or " in response to determining ".
Referring to Fig. 1, a kind of driver provided in this embodiment drives to take the state identification method of phone, and the method can To include:
S101:The vehicle window of target vehicle in monitoring image is positioned.
Wherein, monitoring scene may be selected that vehicle flowrate is more or accident easily sends out road scene, for example, common bayonet road, institute Monitoring image is stated captured by bayonet camera.
In one embodiment, for collected monitoring image fsrc(x, y), x, y are respectively the horizontal seat put on monitoring image Mark and ordinate, the width of monitoring image is Width, a height of Height, it is assumed that has obtained the car plate letter of target vehicle in image Breath, the license board information can include car plate color LpColor and license board information Lp:(x, y, w, h), wherein, Lp (x, y) is respectively It is the abscissa of car plate and ordinate (abscissa and ordinate in the car plate upper left corner or the abscissa or ordinate at car plate center Deng), Lp (w, h) is the width and height of car plate respectively.
Wherein, vehicle window positioning is not limited to following scheme, it is, for example, possible to use Hough straight-line detections position vehicle window, Huo Zhetong It crosses a kind of Adaboost (iterative algorithm) vehicle window detection positioning vehicle window or determines vehicle window area using vehicle window upper right Corner character Domain.
In the present embodiment, vehicle window region is determined using vehicle window upper right Corner character.
Specifically, the location information acquisition process of the vehicle window includes:
According to the license board information of the target vehicle, vehicle window upper right comer region is obtained;
Utilize the location information of positioning filter positioning vehicle window upper right angle point.
In the present embodiment, vehicle window upper right comer region is estimated by license board information first, reuses positioning filter Position the location information of upper right angle point.
For a candidate vehicle window upper right corner image fun(x, y) utilizes positioning filterTo candidate image fun (x, y) carry out convolution calculation formula be:
In formula (1), x, y are respectively the abscissa and ordinate at candidate image midpoint;
Represent convolution operation, g (x, y) is candidate image fun(x, y) and positioning filterConvolution results.
Calculated by formula (1), find out g (x, y) peak point be vehicle window upper right corner dot position information RgtUp (x, y)。
In the present embodiment, vehicle window upper right Corner character wave filterIt is trained by the image demarcated in batches It arrives:
In formula (2), hi(x, y) is the corresponding positioning filter of i-th vehicle window upper right corner picture,
a′iFor the filter weight coefficient after normalized,
In the present embodiment,
In formula (3), xi、yiThe location information of vehicle window upper right angle point for calibration;
δ is empirical coefficient;
fun(x, y) is known vehicle window upper right corner image, and anti-h can be released according to formula (1) and (3)i(x, y), by criticizing Amount has marked image, and positioning filter is calculated
S102:According to the location information of vehicle window and the license board information of target vehicle, obtain driver and detect candidate region.
Wherein, the shape for rule can be set as needed in the shape of driver's detection candidate region.In the present embodiment, drive The person of sailing detects the shape selection rectangle of candidate region.
The vehicle window upper right corner dot position information RgtUp (x, y) of target vehicle has been obtained according to step S101, has been believed with reference to car plate Cease Lp:(x, y, w, h) is calculated and is obtained driver detection candidate region Driver (x, y, w, h):
Driver (x)=min (RgtUp (x)-α * Lp (w), 0) (2)
Driver (y)=min (RgtUp (y)-β * Lp (w), 0) (3)
Driver (w)=min (γ * Lp (w), Width-Driver (x)) (4)
Driver (h)=min (ε * Lp (w), Height-Driver (y)) (5)
In formula (2)-(5), Driver (x, y) be respectively driver detect candidate region vertex (such as can be a left side Upper angular vertex, lower-left angular vertex, upper right angular vertex and bottom right angular vertex) abscissa and ordinate;
Driver (w, h) is the width and height that driver detects candidate region respectively;
α, β, ε and γ are preset empirical coefficient, wherein, α ∈ [1,2], β ∈ [0.5,1], ε ∈ [2.5,3.5], γ ∈[1.5,2.5]。
In the present embodiment, α=1.5, β=0.5, γ=2.0, ε=3.0.
According to above-mentioned formula (2)-(5), it can calculate and obtain driver detection candidate region Driver (x, y, w, h).
S103:Utilization orientation histogram of gradients and support vector machines detect candidate region to the driver and are detected, Obtain the head and shoulder region of the driver.
Referring to Fig. 2, in this step, using histograms of oriented gradients (HOG, Histogram of Oriented Gradient) combination supporting vector machine (SVM, Support Vector Machine) classification obtains driver region coarse positioning Obtain the head and shoulder region of the driver.Extraction driver detect candidate region HOG features after, by HOG features input SVM into Row training, so as to carry out coarse positioning to driver region.
In the present embodiment, after candidate region is detected to driver and carries out histograms of oriented gradients feature extraction, pass through cunning The mode of window detection obtains the histograms of oriented gradients feature of corresponding sliding window size, and input SVM classifies respectively, is driven The person's of sailing head and shoulder detects.And the head and shoulder detection of use direction histogram of gradients feature is compared to colour of skin Face datection, DPM or The advantage of Adaboost Face datections is, for face block (sunshading board blocks face or vehicle window partial occlusion face etc.) or Under the reflective ambiguity of person's face, driver's detection has higher success rate (3% to 5% range).
The classification results in step S103 detection driver's head and shoulders region can be divided into head and shoulder target and non-head and shoulder target, but still There are more flase drops, are affected to result, it is therefore desirable to which histograms of oriented gradients combination supporting vector machine is extracted Flase drop further drops in head and shoulder region.
Again referring to Fig. 2, in the present embodiment, the state identification method that the driver drives to take phone can also include:
Extract the gradient orientation histogram feature drawn game in the head and shoulder region (the head and shoulder region obtained by step S103) Portion's binary feature (LBP, Local Binary Pattern), and by the gradient orientation histogram feature and local binary feature It is combined, to form the feature vector of a multidimensional;
Described eigenvector is carried out using linear discriminant analysis (LDA, Linear Discriminant Analysis) Classification, filters out non-head and shoulder region.
Wherein, LDA belongs to a kind of linear learning method, also referred to as Fisher discriminant analyses (i.e. fischer discriminant analysis), should Method projects to given sample set on straight line, can the separation of the subpoint maximum of different classes of sample is generic Sample as close possible to.
In the present embodiment, using LDA, the head and shoulder window that step S103 is detected is filtered, is divided into someone and nobody Two classes filter out the region of non-head and shoulder, to reduce false drop rate, improve accuracy of detection.
In one embodiment, the process of candidate region progress HOG feature extractions is detected to driver to be included:
Image normalization is 40*40, shares 16 block, wherein, a block (i.e. section) is (i.e. thin by 4 cell Born of the same parents' unit) it forms, the pixel set that a cell is 8*8 is scanned, each cell has 9 bin using 8 pixels as step-length (i.e. 9 parts), therefore HOG intrinsic dimensionalities are tieed up for 16*4*9=576.
And the process that candidate region progress LBP feature extractions are detected to driver can include:
Using uniform LBP patterns, 59 dimensions (59 dimensions are uniform LBP patterns), image normalization 48* are dropped to from 256 dimensions 48, and it is divided into 3*3 blocks, it is 16*16 per block size, every piece has 59 dimensional features, therefore LBP intrinsic dimensionalities are tieed up for 3*3*59=531.
After HOG features and LBP features, HOG and LBP is combined, obtains the feature vector, X of a n (n is natural number) dimensions (x1, x2, x3..., xn).Optionally, n=1009.
Training is iterated to training sample (feature vector of i.e. above-mentioned n dimensions) using LDA linear discriminant analysis, obtains one The optimal training parameter W (w of group1, w2, w3..., wn)。
H=w1*x1+w2*x2+w3*x3+…+wn*xn+b1 (6)
In formula (6), h is the result of linear discriminant analysis;
x1, x2, x3..., xnFor the characteristic value in feature vector X;
w1, w2, w3..., wn、b1For training parameter.
During test, by the feature vector, X (x of extraction1, x2, x3..., xn) substitute into formula (6), h is acquired, if h is more than etc. In 0, then it is assumed that the head and shoulder region of step S103 outputs is final head and shoulder region;If h is less than 0, then it is assumed that step S103 is exported Head and shoulder region for non-head and shoulder region, can directly filter out the non-head and shoulder region.
I.e. according to h, whether the head and shoulder region more than or equal to 0 to be exported to step S103 is further screened, final to obtain more Accurate head and shoulder region, reduces the interference in non-head and shoulder region, so as to improve accuracy of detection.
S104:The head and shoulder region is sequentially input into first layer CNN (Convolutional Neural Network, volume Product neural network) network and second layer CNN networks, the first layer CNN networks tentatively screen the head and shoulder region, obtained The head and shoulder region of the doubtful driver for taking phone is obtained, the second layer CNN networks are to the output knot of the first layer CNN networks Fruit (i.e. the head and shoulder region of the doubtful driver for taking phone of first layer CNN networks output) is further screened, and is driven described in acquisition The person of sailing takes the state of phone.
In the present embodiment, the number of plies of the first layer CNN networks is less than the number of plies of the second layer CNN networks, and described The convolution kernel number of first layer CNN networks is less than the convolution kernel number of the second layer CNN networks.
Wherein, CNN networks include input layer, Nc convolutional layer, Np down-sampling layer, Nf full articulamentums.
Specifically, every layer of convolutional layer includes Nc_Ck convolution kernel, and convolution kernel size is Ckm*Ckm, step-length 1, under every layer The core size of sample level is Pkm*Pkm, step-length Pkm, the neuron that last layer of the full articulamentum full articulamentum exports Quantity is needs number of classifying.
Referring to Fig. 3, in the present embodiment, first layer CNN networks export two classes, take phone and do not take phone the two are driven The person of sailing takes telephone state, i.e. first layer CNN networks output is 2.
Second layer CNN networks are sophisticated category, and it is 4 to export, i.e. 4 drivers take telephone state and (make a phone call, is right in the left side While make a phone call, do not make a phone call and can not penalty).
Wherein, the Nc ∈ [2,10], Np ∈ [2,10], Nf ∈ [1,3];
Nc_Ck∈[Nc_Ckmin,Nc_Ckmax], Nc_Ckmin∈[6,16];
Ckm ∈ [3,7], Pkm ∈ [2,4].
In step S104, head and shoulder region input first layer CNN networks will be obtained in step s 103 first, to be connect It makes a phone call and does not take phone the two states, realize quick, rough identification.
Wherein, the structure of first layer CNN networks is relatively simple, using less the network number of plies and convolution kernel number.Purpose It is fast filtering, it is as much as possible to retain the monitoring picture made a phone call, while exclude the non-monitoring picture made a phone call.
The output result of first layer CNN networks is two classes, that is, makes a phone call and do not make a phone call, a large amount of so as to filter out The non-monitoring image made a phone call, it is not necessary to carry out next layer (i.e. second layer CNN networks) sophisticated category again and (reduce and be input to second The quantity of the monitoring image of layer CNN networks), taking for sophisticated category was both reduced, while driving can be reduced take and make a phone call False Rate.
Compared to first layer CNN networks, second layer CNN network structures are complex, so as to fulfill fine identification.
Again referring to Fig. 3, the state identification method that the driver drives to take phone further includes:
By the head and shoulder region of the doubtful driver for taking phone, (i.e. described first layer CNN networks output is doubtful to be taken The head and shoulder region of the driver of phone) it is divided into left area, right area and overall region;
The left area, right area and overall region are inputted into second layer CNN networks respectively, obtain the driver The state of phone is taken, the state that the driver takes phone includes:The left side takes phone, the right takes number, does not engage electricity Words and can not penalty.
Assuming that the head and shoulder region of driver is Call (x, y, w, h), wherein, Call (x, y) is the abscissa in head and shoulder region And ordinate, Call (w, h) are the width and height in head and shoulder region.
In the present embodiment, head and shoulder region is normalized to w*h, i.e. w=150, h=100, then head and shoulder region to be identified point It is following three pieces:Left area be Call (x, y, α * w, h), right area be Call (x+ (1- α) * w, y, α * w, h), overall area Domain is Call (x, y, w, h).Wherein, α is empirical coefficient.Optionally, α=2/3.
After head and shoulder region to be identified is divided into three pieces, need the left area of division, right area and overall area Domain input second layer CNN networks carry out multiple features fusion differentiation.It can be seen that the present embodiment carries out driver using head and shoulder detection Target coarse positioning carries out driver's flase drop elimination, so as to reduce false drop rate in conjunction with multiple features fusion discriminant analysis.
In the present embodiment, in order to more accurately identify that driver's driving takes the state of phone, the essence of identification is improved Exactness, the state identification method that the driver drives to take phone further include:
If present frame monitoring image recognition result is in for driver and takes telephone state, to present frame monitoring image Next frame monitoring image continues to identify, if the recognition result of the next frame monitoring image is in for driver takes phone shape State is then alerted;Otherwise, it abandons alerting.
Certainly, the frame number of the monitoring image identified needed for can also selecting as needed, for example, at least two frames continuously monitor Image.
As shown in figure 4, the structure diagram of the status identification means of phone is taken for driver's driving that the application provides, with The state identification method that above-mentioned driver drives to take phone is corresponding, can refer to above-mentioned driver and drives to take the state of phone The embodiment of recognition methods come understand or explain the driver drive take phone status identification means content.
Referring to Fig. 4, a kind of driver provided in this embodiment drives to take the status identification means of phone, and described device can To include locating module 100, region acquisition module 200, module of target detection 300 and identification module 400.
Wherein, locating module 100 position for the vehicle window to target vehicle in monitoring image;
Region acquisition module 200 according to the location information of vehicle window and the license board information of target vehicle, obtains driver's inspection Astronomical observation favored area;
Module of target detection 300, utilization orientation histogram of gradients and support vector machines detect candidate regions to the driver Domain is detected, and obtains the head and shoulder region of the driver;
The head and shoulder region is sequentially input first layer CNN networks and second layer CNN networks by identification module 400, and described One layer of CNN network tentatively screens the head and shoulder region, obtains the head and shoulder region of the doubtful driver for taking phone, described Second layer CNN networks further screen the output result of the first layer CNN networks, obtain the driver and take phone State.
Wherein, the number of plies of the first layer CNN networks is less than the number of plies of the second layer CNN networks, and the first layer The convolution kernel number of CNN networks is less than the convolution kernel number of the second layer CNN networks.
Further, the locating module 100 can include:
According to the license board information of the target vehicle, vehicle window upper right comer region is obtained;
Utilize the location information of positioning filter positioning vehicle window upper right angle point.
Further, referring to Fig. 5, the module of target detection 300 can also include feature extraction submodule 301 and filtering Submodule 302.
Wherein, feature extraction submodule 301, for extracting the histograms of oriented gradients feature in the head and shoulder region and part Binary feature, and the histograms of oriented gradients feature and local binary feature are combined, to form the spy of a multidimensional Sign vector;
Filter submodule 302 classifies to described eigenvector using linear discriminant analysis, filters out non-head and shoulder area Domain.
Further, referring to Fig. 6, the identification module 400 can also include dividing submodule 401 and fusion recognition submodule Block 402.
Wherein, submodule 401 is divided, by head and shoulder region (the first layer CNN of the doubtful driver for taking phone Network exports the head and shoulder region of the doubtful driver for taking phone) it is divided into left area, right area and overall region;
The left area, right area and overall region are inputted second layer CNN by fusion recognition submodule 402 respectively Network, obtains the state that the driver takes phone, and the state that the driver takes phone includes:The left side take phone, The right takes phone, does not make a phone call and can not penalty.
Referring to Fig. 7, the status identification means that the driver drives to take phone may also include:
Multiframe authentication module 500, it is right if present frame monitoring image recognition result is in for driver takes telephone state The next frame monitoring image of present frame monitoring image continues to identify, if the recognition result of the next frame monitoring image is driver In telephone state is taken, then alerted;Otherwise, it abandons alerting.
In conclusion the driver of the application drives to take the state identification method of phone and device passes through vehicle glazing area It the step of domain positioning, driver's target detection (head and shoulder region detection) and cascade these pure video detections of CNN networks, can Remove artificial detection driving from and take taking time and effort for telephone state, and can exclude under complex scene non-genuine takes phone Accidentally pick up, improve accuracy of identification, be compared to conventional method, penalty accuracy rate higher, adaptability, robustness to scene are more It is good.
The foregoing is merely the preferred embodiment of the application, not limiting the application, all essences in the application God and any modification, equivalent substitution, improvement and etc. within principle, done, should be included within the scope of the application protection.

Claims (10)

1. a kind of driver drives to take the state identification method of phone, which is characterized in that the method includes:
The vehicle window of target vehicle in monitoring image is positioned;
According to the location information of vehicle window and the license board information of target vehicle, obtain driver and detect candidate region;
Utilization orientation histogram of gradients and support vector machines detect candidate region to the driver and are detected, and are driven described in acquisition The head and shoulder region for the person of sailing;
The head and shoulder region is sequentially input into first layer CNN networks and second layer CNN networks, the first layer CNN networks are to institute It states head and shoulder region tentatively to be screened, obtains the head and shoulder region of the doubtful driver for taking phone, the second layer CNN networks pair The output result of the first layer CNN networks is further screened, and obtains the state that the driver takes phone.
2. driver as described in claim 1 drives to take the state identification method of phone, which is characterized in that by the head Shoulder region is inputted before first layer CNN networks and second layer CNN networks, is further included:
Extract the histograms of oriented gradients feature in the head and shoulder region and local binary feature, and by the histograms of oriented gradients Feature and local binary feature are combined, to form the feature vector of a multidimensional;
Classified using linear discriminant analysis to described eigenvector, filter out non-head and shoulder region.
3. driver as described in claim 1 drives to take the state identification method of phone, which is characterized in that
The method further includes:
It is left area, right area and overall region by the head and shoulder region division of the doubtful driver for taking phone;
The left area, right area and overall region are inputted into second layer CNN networks respectively, the driver is obtained and takes The state of phone, the state that the driver takes phone include:The left side takes phone, the right takes phone, do not make a phone call with It and can not penalty.
4. driver as described in claim 1 drives to take the state identification method of phone, which is characterized in that the method is also Including:
If present frame monitoring image recognition result is in for driver and takes telephone state, to the next of present frame monitoring image Frame monitoring image continues to identify, if the recognition result of the next frame monitoring image is in for driver takes telephone state, It is alerted;Otherwise, it abandons alerting.
5. driver as described in claim 1 drives to take the state identification method of phone, which is characterized in that the vehicle window Location information acquisition process includes:
According to the license board information of the target vehicle, vehicle window upper right comer region is obtained;
Utilize the location information of positioning filter positioning vehicle window upper right angle point.
6. a kind of driver drives to take the status identification means of phone, which is characterized in that described device includes:
Locating module positions the vehicle window of target vehicle in monitoring image;
Region acquisition module according to the location information of vehicle window and the license board information of target vehicle, obtains driver and detects candidate Region;
Module of target detection, utilization orientation histogram of gradients and support vector machines detect candidate region to the driver and examine It surveys, obtains the head and shoulder region of the driver;
The head and shoulder region is sequentially input first layer CNN networks and second layer CNN networks, the first layer CNN by identification module Network tentatively screens the head and shoulder region, obtains the head and shoulder region of the doubtful driver for taking phone, the second layer CNN networks further screen the output result of the first layer CNN networks, obtain the state that the driver takes phone.
7. driver as claimed in claim 6 drives to take the status identification means of phone, which is characterized in that the target inspection Module is surveyed to further include:
Feature extraction submodule, extracts the histograms of oriented gradients feature in the head and shoulder region and local binary feature, and by institute It states histograms of oriented gradients feature and local binary feature is combined, to form the feature vector of a multidimensional;
Filter submodule classifies to described eigenvector using linear discriminant analysis, filters out non-head and shoulder region.
8. driver as claimed in claim 6 drives to take the status identification means of phone, which is characterized in that the identification mould Block further includes:
Divide submodule, by the head and shoulder region division of the doubtful driver for taking phone for left area, right area and Overall region;
The left area, right area and overall region are inputted second layer CNN networks, obtained by fusion recognition submodule respectively The state that the driver takes phone is obtained, the state that the driver takes phone includes:The left side takes phone, the right takes Phone is not made a phone call and can not penalty.
9. driver as claimed in claim 6 drives to take the status identification means of phone, which is characterized in that described device is also Including:
Multiframe authentication module, if present frame monitoring image recognition result is in for driver takes telephone state, to present frame The next frame monitoring image of monitoring image continues to identify, be connect if the recognition result of the next frame monitoring image is in for driver Telephoning state is then alerted;Otherwise, it abandons alerting.
10. driver as claimed in claim 6 drives to take the status identification means of phone, which is characterized in that the positioning Module includes:
According to the license board information of the target vehicle, vehicle window upper right comer region is obtained;
Utilize the location information of positioning filter positioning vehicle window upper right angle point.
CN201611185468.8A 2016-12-20 2016-12-20 Method and device for recognizing state of driver for calling and answering Active CN108205649B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611185468.8A CN108205649B (en) 2016-12-20 2016-12-20 Method and device for recognizing state of driver for calling and answering

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611185468.8A CN108205649B (en) 2016-12-20 2016-12-20 Method and device for recognizing state of driver for calling and answering

Publications (2)

Publication Number Publication Date
CN108205649A true CN108205649A (en) 2018-06-26
CN108205649B CN108205649B (en) 2021-08-31

Family

ID=62603495

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611185468.8A Active CN108205649B (en) 2016-12-20 2016-12-20 Method and device for recognizing state of driver for calling and answering

Country Status (1)

Country Link
CN (1) CN108205649B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109165607A (en) * 2018-08-29 2019-01-08 浙江工业大学 A kind of hand-held phone detection method of the driver based on deep learning
CN109214289A (en) * 2018-08-02 2019-01-15 厦门瑞为信息技术有限公司 A kind of Activity recognition method of making a phone call from entirety to local two stages
CN109376634A (en) * 2018-10-15 2019-02-22 北京航天控制仪器研究所 A kind of Bus driver unlawful practice detection system neural network based
CN110110631A (en) * 2019-04-25 2019-08-09 深兰科技(上海)有限公司 It is a kind of to identify the method and apparatus made a phone call
CN111310751A (en) * 2018-12-12 2020-06-19 北京嘀嘀无限科技发展有限公司 License plate recognition method and device, electronic equipment and storage medium
CN111723602A (en) * 2019-03-19 2020-09-29 杭州海康威视数字技术股份有限公司 Driver behavior recognition method, device, equipment and storage medium
CN112307821A (en) * 2019-07-29 2021-02-02 顺丰科技有限公司 Video stream processing method, device, equipment and storage medium
CN112966563A (en) * 2021-02-04 2021-06-15 同济大学 Behavior identification method based on human skeleton detection and tracking algorithm

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120250983A1 (en) * 2011-03-30 2012-10-04 Sony Corporation Object detecting apparatus and method
CN103971135A (en) * 2014-05-05 2014-08-06 中国民航大学 Human body target detection method based on head and shoulder depth information features
CN104156717A (en) * 2014-08-31 2014-11-19 王好贤 Method for recognizing rule breaking of phoning of driver during driving based on image processing technology
CN104715238A (en) * 2015-03-11 2015-06-17 南京邮电大学 Pedestrian detection method based on multi-feature fusion
CN105512683A (en) * 2015-12-08 2016-04-20 浙江宇视科技有限公司 Target positioning method and device based on convolution neural network
CN106056071A (en) * 2016-05-30 2016-10-26 北京智芯原动科技有限公司 Method and device for detection of driver' behavior of making call

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120250983A1 (en) * 2011-03-30 2012-10-04 Sony Corporation Object detecting apparatus and method
CN103971135A (en) * 2014-05-05 2014-08-06 中国民航大学 Human body target detection method based on head and shoulder depth information features
CN104156717A (en) * 2014-08-31 2014-11-19 王好贤 Method for recognizing rule breaking of phoning of driver during driving based on image processing technology
CN104715238A (en) * 2015-03-11 2015-06-17 南京邮电大学 Pedestrian detection method based on multi-feature fusion
CN105512683A (en) * 2015-12-08 2016-04-20 浙江宇视科技有限公司 Target positioning method and device based on convolution neural network
CN106056071A (en) * 2016-05-30 2016-10-26 北京智芯原动科技有限公司 Method and device for detection of driver' behavior of making call

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LÁSZLÓ TÓTH: ""Phone recognition with deep sparse rectifier neural networks"", 《2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING》 *
王丹: ""基于机器视觉的驾驶员打电话行为检测"", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109214289B (en) * 2018-08-02 2021-10-22 厦门瑞为信息技术有限公司 Method for recognizing two-stage calling behavior from whole to local
CN109214289A (en) * 2018-08-02 2019-01-15 厦门瑞为信息技术有限公司 A kind of Activity recognition method of making a phone call from entirety to local two stages
CN109165607A (en) * 2018-08-29 2019-01-08 浙江工业大学 A kind of hand-held phone detection method of the driver based on deep learning
CN109165607B (en) * 2018-08-29 2021-12-14 浙江工业大学 Driver handheld phone detection method based on deep learning
CN109376634A (en) * 2018-10-15 2019-02-22 北京航天控制仪器研究所 A kind of Bus driver unlawful practice detection system neural network based
CN111310751A (en) * 2018-12-12 2020-06-19 北京嘀嘀无限科技发展有限公司 License plate recognition method and device, electronic equipment and storage medium
CN111310751B (en) * 2018-12-12 2023-08-29 北京嘀嘀无限科技发展有限公司 License plate recognition method, license plate recognition device, electronic equipment and storage medium
CN111723602A (en) * 2019-03-19 2020-09-29 杭州海康威视数字技术股份有限公司 Driver behavior recognition method, device, equipment and storage medium
CN111723602B (en) * 2019-03-19 2023-08-08 杭州海康威视数字技术股份有限公司 Method, device, equipment and storage medium for identifying driver behavior
CN110110631A (en) * 2019-04-25 2019-08-09 深兰科技(上海)有限公司 It is a kind of to identify the method and apparatus made a phone call
CN110110631B (en) * 2019-04-25 2021-06-29 深兰科技(上海)有限公司 Method and equipment for recognizing and making call
CN112307821A (en) * 2019-07-29 2021-02-02 顺丰科技有限公司 Video stream processing method, device, equipment and storage medium
CN112966563B (en) * 2021-02-04 2022-09-20 同济大学 Behavior identification method based on human skeleton detection and tracking algorithm
CN112966563A (en) * 2021-02-04 2021-06-15 同济大学 Behavior identification method based on human skeleton detection and tracking algorithm

Also Published As

Publication number Publication date
CN108205649B (en) 2021-08-31

Similar Documents

Publication Publication Date Title
CN108205649A (en) Driver drives to take the state identification method and device of phone
CN106682601B (en) A kind of driver's violation call detection method based on multidimensional information Fusion Features
CN105844257B (en) The early warning system and method for road sign are missed based on machine vision travelling in fog day
CN110119676A (en) A kind of Driver Fatigue Detection neural network based
CN104378582B (en) A kind of intelligent video analysis system and method cruised based on Pan/Tilt/Zoom camera
CN105809138B (en) A kind of road warning markers detection and recognition methods based on piecemeal identification
CN105488453B (en) A kind of driver based on image procossing does not fasten the safety belt detection recognition method
CN106469309A (en) The method and apparatus of vehicle monitoring, processor, image capture device
CN109635656A (en) Vehicle attribute recognition methods, device, equipment and medium neural network based
CN102867188B (en) Method for detecting seat state in meeting place based on cascade structure
CN108309311A (en) A kind of real-time doze of train driver sleeps detection device and detection algorithm
CN104217217B (en) A kind of vehicle mark object detecting method and system based on two layers of classified
CN106781282A (en) A kind of intelligent travelling crane driver fatigue early warning system
CN109711264A (en) A kind of bus zone road occupying detection method and device
CN110245663A (en) One kind knowing method for distinguishing for coil of strip information
CN107330373A (en) A kind of parking offense monitoring system based on video
CN106529494A (en) Human face recognition method based on multi-camera model
CN109670515A (en) Method and system for detecting building change in unmanned aerial vehicle image
CN103824081A (en) Method for detecting rapid robustness traffic signs on outdoor bad illumination condition
CN101127076A (en) Human eye state detection method based on cascade classification and hough circle transform
CN107704853A (en) A kind of recognition methods of the traffic lights based on multi-categorizer
Shopa et al. Traffic sign detection and recognition using OpenCV
CN103021179B (en) Based on the Safe belt detection method in real-time monitor video
CN107292933A (en) A kind of vehicle color identification method based on BP neural network
CN106022242B (en) Method for identifying call receiving and making of driver in intelligent traffic system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant