CN104573659B - A kind of driver based on svm takes phone-monitoring method - Google Patents
A kind of driver based on svm takes phone-monitoring method Download PDFInfo
- Publication number
- CN104573659B CN104573659B CN201510013139.4A CN201510013139A CN104573659B CN 104573659 B CN104573659 B CN 104573659B CN 201510013139 A CN201510013139 A CN 201510013139A CN 104573659 B CN104573659 B CN 104573659B
- Authority
- CN
- China
- Prior art keywords
- msub
- mrow
- mtd
- mtr
- hand
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/59—Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
- G06V20/597—Recognising the driver's state or behaviour, e.g. attention or drowsiness
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/107—Static hand or arm
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
- Nitrogen And Oxygen Or Sulfur-Condensed Heterocyclic Ring Systems (AREA)
Abstract
The step of the present invention relates to a kind of driver based on svm to take phone-monitoring method, and the monitoring method includes following order:Establish Face datection grader and hand images detect grader as the hand of training positive sample during taking phone;The driving condition image of collection driver in real time;Suitable image region is selected to detect and take the effective detection region that phone detects as hand exercise;According to effective detection area image, real-time judge driver whether there is the warming-up exercise for taking phone;According to effective detection area image, the time that the hand of driver is rested on beside ear is monitored in detail, and judge whether driver takes phone according to the length of residence time;The real-time video that driver is taken to phone is sent to remote server, and receives the order of remote server transmission.
Description
Technical field
The present invention relates to safe driving technical field, and in particular to a kind of driver based on svm takes phone-monitoring side
Method.
Background technology
With the rapid growth of car ownership, people enjoy the facility of traffic and it is quick while, also along with
All kinds of traffic accidents take place frequently, and cause huge personnel and economic loss.Cause the factor of traffic accident a lot, driver drives
It is one of important inducement that phone is taken in way.Due to that can not obtain the driving behavior video of driver in real time, some passenger traffics and
The supervision department of Shipping enterprises can only can not carry out monitoring in advance and in advance using foundation of the deduction as division responsibility afterwards
It is anti-.Therefore, monitor driver in real time takes phone behavior, feeds back to supervision department of transport enterprise in time and is prevented, right
In avoiding major traffic accidents, play a part of can not be substituted.
SVMs (svm) is a kind of supervised learning algorithm, and it is solving small sample, the knowledge of non-linear and high dimensional pattern
Many distinctive advantages are shown in not, the VC dimensions that this method is built upon Statistical Learning Theory are theoretical former with Structural risk minization
On the basis of reason, according to limited sample information in the complexity (the study precision i.e. to specific training sample) of model and study
Ability (i.e. without error identify arbitrary sample ability) between seek best compromise, to obtain best Generalization Ability (or
Claim generalization ability).
At present, the monitoring of phone behavior is taken for driver, conventional technical method has following several:
(1) be monitored based on mobile phone signal, such method by driver's cabin place a mobile phone signal detector,
According to the different degrees of of signal fluctuation, judge whether to take phone behavior.Such method can reach on goods stock
Monitoring takes the effect of phone.But on passenger stock, more interference be present, such as the mobile phone signal interference of passenger on car
Deng serious missing inspection and flase drop being present, can not realize that comprehensive monitoring driver takes phone behavior in real time.
(2) it is monitored based on video image, such method by monitoring whether driver's both hands are placed on steering wheel in real time
On, once there is certain hand departure direction disk, that is, it is considered as to take phone.Serious flase drop be present in such method, because
There is the custom of a hand steered steering wheel in many drivers, therefore, this method is in actual environment using having the problem of larger.
The content of the invention
It is an object of the invention to provide a kind of driver based on svm to take phone-monitoring method, and the monitoring method is
Based on video image processing technology, judge to take phone behavior by monitoring the state of driver's ear region hand in real time,
With the features such as monitoring degree of accuracy height, missing inspection flase drop is less, and speed is fast, and cost is low.
The technical scheme is that:
The step of a kind of driver based on svm takes phone-monitoring method, and the monitoring method includes following order:
(1) establish Face datection grader and hand images detect classification as the hand of training positive sample during taking phone
Device;
(2) the driving condition image of driver is gathered in real time;
(3) suitable image region is selected to detect and take the effective detection region that phone detects as hand exercise;
(4) warming-up exercise for taking phone whether there is according to effective detection area image, real-time judge driver;If
It is then to perform step (5);Step (2) is performed if it is not, then returning;
(5) time rested on according to effective detection area image, the detailed hand for monitoring driver beside ear, and according to
The time length that the hand of driver is rested on beside ear judges whether driver takes phone;If so, then perform step
(6);Step (2) is performed if it is not, then returning;
(6) real-time video that driver is taken to phone is sent to remote server, and receives remote server transmission
Order.
In step (3), the described suitable image region of selection detects and taken phone detection as hand exercise
Effective detection region, the step of specifically including following order:
(31) harr features and adaboost sorting algorithms are used, detects face location;
(32) according to five, the three front yard layout rule of face, the position that coarse positioning is left and right two;
(33) it is accurately positioned out the position of eyes;
(34) suitable effective image subregion is selected.
Described according to effective detection area image in step (4), real-time judge driver, which whether there is, takes phone
Warming-up exercise, the step of specifically including following order:
(41) uniform point sampling is carried out respectively in left and right effective detection region;
(42) accurate tracking of sampled point is carried out;
(43) movable information of correct tracking sampling point is obtained;
(44) statistical nature in effective detection region is obtained, the statistical nature of described detection zone includes left effective detection
The mean motion intensity avem in regionl, right effective detection region mean motion intensity avemr, left effective detection region motion
Range RlWith the motion range R in right effective detection regionr;
(45) judge whether to lift action of the hand close to ear.
It is described according to effective detection area image in step (5), monitor that the hand of driver is rested on beside ear in detail
Time, and the time length rested on according to the hand of driver beside ear judges whether driver takes phone;Tool
Body includes the step of following order:
(51) using effective detection area image as present image, the Gradient Features of present image are calculated, and to gradient side
It is corrected to angle, the scope of gradient direction is limited to [0 π];
(52) one a width of W of construction, a height of H hough transform window, W and H value are respectively equal to the instruction of hand trainer
Practice the wide high level of sample, and subwindow division is carried out to hough transform window;
(53) hough transform window is slided with a fixed step size in effective detection region, judges that hough transform window covers
Region whether there is hand images;
(54) judge whether hough transform window is slided at the ending of present image;If so, then perform step (55);If
No, then hough transform window sliding to next position, performs step (53) again;
(55) a width of W of present image is setimage, a height of Himage, change of scale is carried out to present image, carries out yardstick change
A width of σ W of image after changingimage, a height of σ Himage, the image after progress change of scale is set to present image;Wherein, 0 < σ
< 1;
(56) judge whether the size of present image is less than minimum threshold;If the size of present image is less than minimum threshold,
Then perform step (57);If the size of present image is more than minimum threshold, returns and perform step (51);
(57) number of the candidate target in the list of candidate region and position, the position of hand in comprehensive descision present frame
Put;
(58) judge to accumulate whether frame number reaches unit frame number;If so, then perform step (59);Performed if it is not, then returning
Step (51);
(59) in statistical unit frame number, the frame number proportion of hand images be present;And according to the frame number institute of hand images
Accounting example judges whether driver is in and takes telephone state.
In step (34), the suitable effective image subregion of described selection, specifically realized using below equation:
Wherein, rectleft、rectrightSub-rectangular areas near the left and right ear of selection is represented respectively, and rect represents inspection
The face location rectangular area measured, pointl、pointrThe left hand edge point of left eye and the right hand edge point of right eye are represented respectively.
In step (43), the movable information of the described correct tracking sampling point of acquisition, specifically realized using below equation:
Wherein, M (i) represents the motion amplitude of ith sample point, and θ (i) represents the direction of motion of ith sample point,
xpointiRepresent coordinate of the ith sample point on present frame, ypointiCoordinate of the ith sample point on previous frame is represented,
Dx represents amount of exercise of some sampled point in x directions, and Dy represents amount of exercise of some sampled point in y directions.
In step (44), the statistical nature in described acquisition effective detection region, specifically realized using below equation:
Wherein, sumlRepresent that there is the motion amplitude of the sampled point substantially moved, N in left detection zonemlRepresent left detection
There is the number of the sampled point substantially moved, N in regionlRepresent that left detection zone is interior with obvious motion and close to ear
The number of sampled point;
sumrRepresent that there is the motion amplitude of the sampled point substantially moved, N in right detection zonemrRepresent in right detection zone
Number with the sampled point substantially moved, NrRepresent there is obvious motion and to the close sampled point of ear in right detection zone
Number;
Ml(i) motion amplitude of i-th of correct tracking sampling point in left detection zone, θ are representedl(i) left detection zone is represented
The direction of motion of interior i-th of correct tracking sampling point;
Mr(i) motion amplitude of i-th of correct tracking sampling point in right detection zone, θ are representedr(i) right detection zone is represented
The direction of motion of interior i-th of correct tracking sampling point.
In step (45), described judges whether to lift action of the hand close to ear, is specifically realized using below equation:
Wherein, exist=1 represents action of the lift hand close to ear be present, and exist=0 represents to be not present lift hand close to ear
Piece action, slRepresent the action with the presence or absence of lift hand close to ear in the effective detection region of left side;srRepresent effective on right side
Action in detection zone with the presence or absence of lift hand close to ear;TmlWhen representing to lift hand close to ear in the effective detection region of left side
Motion strength threshold;TθlMotion range threshold value when representing to lift hand close to ear in the effective detection region of left side;TmrRepresent
Motion strength threshold when lifting hand close to ear in the effective detection region of right side;TθrRepresent that hand is lifted in the effective detection region of right side to be leaned on
Motion range threshold value during nearly ear;TmlbRepresent in the effective detection region of left side, being averaged when the lift hand of standard is close to ear
Exercise intensity;TθlbRepresent in the effective detection region of left side, the mean motion range when lift hand of standard is close to ear;TmrbTable
Show in the effective detection region of right side, the mean motion intensity when lift hand of standard is close to ear;TθrbExpression is effectively examined on right side
Survey in region, the mean motion range when lift hand of standard is close to ear.
In step (51), the Gradient Features of described calculating present image, and gradient direction angle is corrected, specifically
Realized using below equation:
Wherein, M (x, y), θ (x, y) represent the gradient magnitude and gradient direction at pixel (x, y) place respectively, and f (x, y) is represented
The gray value at pixel (x, y) place.Gx represents the partial derivative in pixel (x, y) place x directions, and Gy represents the inclined of pixel (x, y) place y directions
Derivative.
It is described to slide hough transform window in effective detection region with a fixed step size in step (53), judge square
The region of shape detection window covering whether there is hand images;Specifically realized using the step of following order:
(531) histogram of gradients of each subwindow overlay area is counted;
(532) characteristic vector of hough transform window overlay area is obtained;
(533) characteristic vector of hough transform window overlay area is normalized;
(534) characteristic vector of hough transform window overlay area is sent into hand svm graders, predicts its affiliated class
Not;If hand svm graders predict that this feature vector belongs to hand images, hough transform window position is added into candidate
In zone list.
Phone-monitoring method is taken compared to other drivers, the present invention uses video image processing technology, is supervised by triggering
Control pattern monitors the hand existence of driver's ear region in real time, and judgement takes phone behavior, has monitoring degree of accuracy height,
The features such as missing inspection flase drop is less, and speed is fast, and cost is low.
Brief description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is trigger module workflow diagram;
Fig. 3 is monitoring module workflow diagram;
Fig. 4 is training positive sample master drawing;
Fig. 5 is the effective detection zone schematic diagram of positioning;
Fig. 6 is point sampling design sketch;
Fig. 7 is the subwindow division schematic diagram of hough transform window;
Fig. 8 is hough transform window sliding schematic diagram;
Fig. 9 is hand candidate test position design sketch, wherein, rectangle frame 1 is effective detection region, and rectangle frame 2 is candidate
Detection zone;
Figure 10 is the final test position design sketch of hand, wherein, rectangle frame 1 is effective detection region, and rectangle frame 2 is final
Test position.
Embodiment
The present invention is further illustrated below in conjunction with the accompanying drawings.
As shown in figure 1, in embodiment, the driver of the present invention based on svm takes phone-monitoring system and included initially
Change module, acquisition module, locating module, trigger module, monitoring module and communication module.The driver based on svm takes electricity
The specific implementation step for talking about monitoring system is as follows:
S1, perform initialization module.
The function of initial module is loading and the necessary grader file of training system, to comprise the following steps that:
S11, the existing Face datection grader file of loading.
S12, as shown in figure 4, collecting hand images when taking phone is used as positive sample, examined based on the theoretical training hands of svm
Survey grader file.
S2, perform acquisition module.
The function of acquisition module is to gather the head image of the driving condition image, mainly driver of driver in real time.
S3, perform locating module.
The function of locating module is to select suitable image region to be detected as hand exercise and take what phone detected
Effective coverage, mainly select the regional area near left and right ear.The module can greatly lift detection speed, remove
More interference regions, as shown in Figure 5.The module comprises the following steps that:
S31, using harr features and adaboost sorting algorithms, detect face location.
S32, five, the three front yard layout rule based on face, the position that coarse positioning is left and right two.
S33, the position for being accurately positioned out eyes.
S34, suitable effective image subregion selected using formula (1) and formula (2).
Wherein, rectleft、rectrightSub-rectangular areas near the left and right ear of selection is represented respectively, and rect represents inspection
The face location rectangular area measured, pointl、pointrThe left hand edge point of left eye and the right hand edge point of right eye are represented respectively.
S4, judge trigger module open and-shut mode, if trigger module is in opening, step S5 is performed, if trigger mode
Block is closed, then performs step S7.
S5, perform trigger module.
The function of trigger module is that real-time judge driver whether there is the warming-up exercise for taking phone, in particular to drive
Action of the person of sailing with the presence or absence of lift hand close to ear.If it does, explanation driver is possible to preparation and takes phone, now exit
Trigger module, return and open monitoring module signal;If there is no the action, system may proceed to carry out real-time judge, under wait
Once take the appearance of phone warming-up exercise.As shown in Fig. 2 the module comprises the following steps that:
S51, carry out uniform point sampling, effect such as Fig. 6 respectively in left and right effective detection region.
S52, the accurate tracking for carrying out sampled point.Specific track algorithm is referring to document:Forward-Backward Error:
Automatic Detection of Tracking Failures, Zdenek Kalal, Krystian Mikolajczyk,
Jiri Matas, Pattern Recognition (ICPR), 2010 20th International Conference on.
S53, the movable information for obtaining using formula (3) and formula (4) correct tracking sampling point;
Wherein, M (i) represents the motion amplitude of ith sample point, and θ (i) represents the direction of motion of ith sample point,
xpointiRepresent coordinate of the ith sample point on present frame, ypointiCoordinate of the ith sample point on previous frame is represented,
Dx represents amount of exercise of some sampled point in x directions, and Dy represents amount of exercise of some sampled point in y directions.
S54, the statistical nature for obtaining using formula (5) and formula (6) effective detection region.Described effective detection region
Statistical nature include the mean motion intensity avem in left side effective detection regionl, the right effective detection region mean motion
Intensity avemr, left side effective detection region motion range RlWith the motion range R in the right effective detection regionr。
Wherein, sumlRepresent that there is the motion amplitude of the sampled point substantially moved, N in left detection zonemlRepresent left detection
There is the number of the sampled point substantially moved, N in regionlRepresent that left detection zone is interior with obvious motion and close to ear
The number of sampled point;sumrRepresent that there is the motion amplitude of the sampled point substantially moved, N in right detection zonemrRepresent right detection
There is the number of the sampled point substantially moved, N in regionrRepresent that right detection zone is interior with obvious motion and close to ear
The number of sampled point;Ml(i) motion amplitude of i-th of correct tracking sampling point in left detection zone, θ are representedl(i) left inspection is represented
Survey the direction of motion of i-th of correct tracking sampling point in region;Mr(i) i-th of correct tracking sampling in right detection zone is represented
The motion amplitude of point, θr(i) direction of motion of i-th of correct tracking sampling point in right detection zone is represented.
S55, using formula (7)-(10), judge whether to lift action of the hand close to ear.If in the presence of exiting triggering
Module;If it is not, then continue executing with trigger module.
Wherein, exist=1 represents action of the lift hand close to ear be present, and exist=0 represents to be not present lift hand close to ear
Piece action.slRepresent the action with the presence or absence of lift hand close to ear in the effective detection region of left side;srRepresent effective on right side
Action in detection zone with the presence or absence of lift hand close to ear;TmlWhen representing to lift hand close to ear in the effective detection region of left side
Motion strength threshold;TθlMotion range threshold value when representing to lift hand close to ear in the effective detection region of left side;TmrRepresent
Motion strength threshold when lifting hand close to ear in the effective detection region of right side;TθrRepresent that hand is lifted in the effective detection region of right side to be leaned on
Motion range threshold value during nearly ear;TmlbRepresent in the effective detection region of left side, being averaged when the lift hand of standard is close to ear
Exercise intensity;TθlbRepresent in the effective detection region of left side, the mean motion range when lift hand of standard is close to ear;TmrbTable
Show in the effective detection region of right side, the mean motion intensity when lift hand of standard is close to ear;TθrbExpression is effectively examined on right side
Survey in region, the mean motion range when lift hand of standard is close to ear.
S6, judge whether to trigger monitoring module.If in the presence of the warming-up exercise for taking phone, monitoring module will be triggered out
Open, carry out in-depth monitoring, simultaneously close off trigger module.If in the absence of the warming-up exercise for taking phone, it is returned directly to gather mould
Block, the triggering carried out next time judge.
S7, perform monitoring module.
The function of monitoring module is to monitor the time that the hand of driver is rested on beside ear in detail.If the hand of driver
The time long enough rested on beside ear, illustrate that driver takes phone, then return to open communication module by signal.If
It is no, then illustrate that this triggering unlatching belongs to erroneous judgement.As shown in figure 3, the module comprises the following steps that:
S71, using effective detection area image as present image, using formula (11) and formula (12), calculate current
The Gradient Features of image, and gradient direction angle is corrected using formula (13), the scope of gradient direction is limited to [0
π]。
Wherein, M (x, y), θ (x, y) represent the gradient magnitude and gradient direction at pixel (x, y) place, and f (x, y) represents pixel
The gray value at (x, y) place.
The one wide high respectively W and H hough transform window of S72, construction, W and H value are equal to hand svm graders
The wide high level of training sample.As shown in fig. 7, subwindow division is carried out to hough transform window.Partitioning standards are:Driver does not connect
During phone, due to the interference beside ear without hand, now the contouring head in effective detection region is generally vertical, because
This, the narrow rectangular area in selection longitudinal direction is rational as subwindow.
S73, as shown in figure 8, hough transform window is slided with a fixed step size in effective detection region, and judge rectangle
The region of detection window covering whether there is hand images, comprise the following steps that:
S731, the histogram of gradients hist [bin] for counting using formula (14) each subwindow overlay area.
Wherein, floor (x) represents to choose the maximum integer no more than x, and bin span is 1~10.
S732, the histogram of gradients of all subwindow overlay areas of series connection, obtain the feature of detection window overlay area to
Amount.
S733, the characteristic vector normalization by hough transform window overlay area;
S734, judge whether detection window overlay area belongs to hand images, effect such as Fig. 9.In Fig. 9, square frame 1 represents left
Right effective detection region, square frame 2 represent that hough transform window overlay area belongs to hand images.Specific method is by detection window
The characteristic vector of overlay area is sent into hand svm graders, predicts its generic.If belonging to hand images, rectangle is examined
Window position is surveyed to be put into the list of candidate region.
S74, judge whether detection window is slided at the ending of present image.If not reaching image ending, rectangle inspection
Window sliding is surveyed to next position, and performs step S73 again;If having reached image ending, step S75 is performed.
S75, present image size is set as Wimage、Himage, change of scale is carried out to present image, then schemed after change of scale
As size is changed into σ Wimage、σHimage, and the image after change of scale is designated as present image.Wherein 0 < σ < 1.
S76, judge whether present image size is less than minimum threshold.If being less than thresholding, illustrate that present image is too small,
Through that there can not possibly be hand, step S77 is performed;If being more than thresholding, illustrate in present image it is possible to hand be present, perform
Step S71 continues detection process.
S77, the number of candidate target in the list of candidate region and position, the position of hand in comprehensive descision present frame
Put, effect such as Figure 10.In Figure 10, square frame 1 represents left and right effective detection region, and square frame 2 represents hough transform window overlay area
Belong to hand images.
S78, judge to accumulate whether frame number reaches unit frame number.If reaching, step S79 is performed;If it is not, then perform step
S71 continues detection process.
In S79, statistical unit frame number, the frame number proportion of hand images be present.If the ratio is more than certain threshold value,
Illustrate that current driver's are in and take telephone state.
S710, trigger module is opened, close monitoring module.
S8, perform communication module.
The function of communication module is that, when driver, which is in, takes telephone state, the module sends to remote server and driven
The person of sailing takes the real-time video of phone, and now supervision department of transport enterprise can timely be handled by the video;If
Need to converse with driver, remote command can also be received by the module.
Embodiment described above is only that the preferred embodiment of the present invention is described, not to the model of the present invention
Enclose and be defined, on the premise of design spirit of the present invention is not departed from, technical side of the those of ordinary skill in the art to the present invention
The various modifications and improvement that case is made, it all should fall into the protection domain of claims of the present invention determination.
Claims (6)
1. a kind of driver based on svm takes phone-monitoring method, it is characterised in that:The monitoring method includes following order
Step:
(1) establish Face datection grader and hand images detect grader as the hand of training positive sample during taking phone;
(2) the driving condition image of driver is gathered in real time;
(3) suitable image region is selected to detect and take the effective detection region that phone detects as hand exercise;
(4) warming-up exercise for taking phone whether there is according to effective detection area image, real-time judge driver;If so, then
Perform step (5);Step (2) is performed if it is not, then returning;
(5) according to effective detection area image, the time that the hand of driver is rested on beside ear is monitored in detail, and according to driving
The time length that the hand of member is rested on beside ear judges whether driver takes phone;If so, then perform step (6);
Step (2) is performed if it is not, then returning;
(6) real-time video that driver is taken to phone is sent to remote server, and receives the order of remote server transmission;
It is described that the preparation for taking phone whether there is according to effective detection area image, real-time judge driver in step (4)
Action, the step of specifically including following order:
(41) uniform point sampling is carried out respectively in left and right effective detection region;
(42) accurate tracking of sampled point is carried out;
(43) movable information of correct tracking sampling point is obtained;
(44) statistical nature in effective detection region is obtained, the statistical nature of described detection zone includes left effective detection region
Mean motion intensity aveml, right effective detection region mean motion intensity avemr, left effective detection region motion range
RlWith the motion range R in right effective detection regionr;
(45) judge whether to lift action of the hand close to ear;
In step (43), the movable information of the described correct tracking sampling point of acquisition, specifically realized using below equation:
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mi>M</mi>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
<mo>=</mo>
<mi>s</mi>
<mi>q</mi>
<mi>r</mi>
<mi>t</mi>
<mo>(</mo>
<msup>
<msub>
<mi>D</mi>
<mi>x</mi>
</msub>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<msub>
<mi>D</mi>
<mi>y</mi>
</msub>
<mn>2</mn>
</msup>
<mo>)</mo>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>&theta;</mi>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
<mo>=</mo>
<mi>a</mi>
<mi>r</mi>
<mi>c</mi>
<mi>t</mi>
<mi>a</mi>
<mi>n</mi>
<mo>(</mo>
<mfrac>
<msub>
<mi>D</mi>
<mi>y</mi>
</msub>
<msub>
<mi>D</mi>
<mi>x</mi>
</msub>
</mfrac>
<mo>)</mo>
</mtd>
</mtr>
</mtable>
</mfenced>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<msub>
<mi>D</mi>
<mi>x</mi>
</msub>
<mo>=</mo>
<mi>x</mi>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>i</mi>
</msub>
<mo>.</mo>
<mi>x</mi>
<mo>-</mo>
<mi>y</mi>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>i</mi>
</msub>
<mo>.</mo>
<mi>x</mi>
</mtd>
</mtr>
<mtr>
<mtd>
<msub>
<mi>D</mi>
<mi>y</mi>
</msub>
<mo>=</mo>
<mi>x</mi>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>i</mi>
</msub>
<mo>.</mo>
<mi>y</mi>
<mo>-</mo>
<mi>y</mi>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>i</mi>
</msub>
<mo>.</mo>
<mi>y</mi>
</mtd>
</mtr>
</mtable>
</mfenced>
Wherein, M (i) represents the motion amplitude of ith sample point, and θ (i) represents the direction of motion of ith sample point, xpointiTable
Show coordinate of the ith sample point on present frame, ypointiCoordinate of the ith sample point on previous frame is represented, Dx represents certain
For individual sampled point in the amount of exercise in x directions, Dy represents amount of exercise of some sampled point in y directions;
In step (44), the statistical nature in described acquisition effective detection region, specifically realized using below equation:
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mi>a</mi>
<mi>v</mi>
<mi>e</mi>
<msub>
<mi>m</mi>
<mi>l</mi>
</msub>
<mo>=</mo>
<mfrac>
<mrow>
<msub>
<mi>sum</mi>
<mi>l</mi>
</msub>
</mrow>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
</mrow>
</msub>
</mfrac>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>R</mi>
<mi>l</mi>
</msub>
<mo>=</mo>
<mfrac>
<msub>
<mi>N</mi>
<mi>l</mi>
</msub>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
</mrow>
</msub>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>avem</mi>
<mi>r</mi>
</msub>
<mo>=</mo>
<mfrac>
<mrow>
<msub>
<mi>sum</mi>
<mi>r</mi>
</msub>
</mrow>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>R</mi>
<mi>r</mi>
</msub>
<mo>=</mo>
<mfrac>
<msub>
<mi>N</mi>
<mi>r</mi>
</msub>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
</mfrac>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mrow>
<msub>
<mi>sum</mi>
<mi>l</mi>
</msub>
<mo>=</mo>
<msub>
<mi>sum</mi>
<mi>l</mi>
</msub>
<mo>+</mo>
<msub>
<mi>M</mi>
<mi>l</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
<mtd>
<mrow>
<msub>
<mi>M</mi>
<mi>l</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>2</mn>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
</mrow>
</msub>
<mo>=</mo>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mn>1</mn>
</mrow>
</msub>
<mo>+</mo>
<mn>1</mn>
</mrow>
</mtd>
<mtd>
<mrow>
<msub>
<mi>M</mi>
<mi>l</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>2</mn>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>N</mi>
<mi>l</mi>
</msub>
<mo>=</mo>
<msub>
<mi>N</mi>
<mi>l</mi>
</msub>
<mo>+</mo>
<mn>1</mn>
</mrow>
</mtd>
<mtd>
<mrow>
<msub>
<mi>M</mi>
<mi>l</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>2</mn>
<mi>a</mi>
<mi>n</mi>
<mi>d</mi>
<mfrac>
<mi>&pi;</mi>
<mn>4</mn>
</mfrac>
<mo><</mo>
<msub>
<mi>&theta;</mi>
<mi>l</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&le;</mo>
<mfrac>
<mi>&pi;</mi>
<mn>2</mn>
</mfrac>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>sum</mi>
<mi>r</mi>
</msub>
<mo>=</mo>
<msub>
<mi>sum</mi>
<mi>r</mi>
</msub>
<mo>+</mo>
<msub>
<mi>M</mi>
<mi>r</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
<mtd>
<mrow>
<msub>
<mi>M</mi>
<mi>r</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>2</mn>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
<mo>=</mo>
<msub>
<mi>N</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
<mo>+</mo>
<mn>1</mn>
</mrow>
</mtd>
<mtd>
<mrow>
<msub>
<mi>M</mi>
<mi>r</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>2</mn>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<msub>
<mi>N</mi>
<mi>r</mi>
</msub>
<mo>=</mo>
<msub>
<mi>N</mi>
<mi>r</mi>
</msub>
<mo>+</mo>
<mn>1</mn>
</mrow>
</mtd>
<mtd>
<mrow>
<msub>
<mi>M</mi>
<mi>r</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>2</mn>
<mi>a</mi>
<mi>n</mi>
<mi>d</mi>
<mfrac>
<mi>&pi;</mi>
<mn>2</mn>
</mfrac>
<mo><</mo>
<msub>
<mi>&theta;</mi>
<mi>r</mi>
</msub>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
<mo>&le;</mo>
<mfrac>
<mrow>
<mn>3</mn>
<mo>*</mo>
<mi>&pi;</mi>
</mrow>
<mn>4</mn>
</mfrac>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
Wherein, sumlRepresent that there is the motion amplitude of the sampled point substantially moved, N in left detection zonemlRepresent left detection zone
The interior number with the sampled point substantially moved, NlRepresent there is obvious motion and to the close sampling of ear in left detection zone
The number of point;
sumrRepresent that there is the motion amplitude of the sampled point substantially moved, N in right detection zonemrRepresent have in right detection zone
The number of the sampled point substantially moved, NrRepresent there is obvious motion and to the number of the close sampled point of ear in right detection zone
Mesh;
Ml(i) motion amplitude of i-th of correct tracking sampling point in left detection zone, θ are representedl(i) the is represented in left detection zone
The direction of motion of i correct tracking sampling points;
Mr(i) motion amplitude of i-th of correct tracking sampling point in right detection zone, θ are representedr(i) the is represented in right detection zone
The direction of motion of i correct tracking sampling points;
In step (45), described judges whether to lift action of the hand close to ear, is specifically realized using below equation:
<mrow>
<mi>e</mi>
<mi>x</mi>
<mi>i</mi>
<mi>s</mi>
<mi>t</mi>
<mo>=</mo>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mn>1</mn>
</mtd>
<mtd>
<mrow>
<msub>
<mi>s</mi>
<mi>l</mi>
</msub>
<mo>+</mo>
<msub>
<mi>s</mi>
<mi>r</mi>
</msub>
<mo>=</mo>
<mn>1</mn>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mn>0</mn>
</mtd>
<mtd>
<mrow>
<msub>
<mi>s</mi>
<mi>l</mi>
</msub>
<mo>+</mo>
<msub>
<mi>s</mi>
<mi>r</mi>
</msub>
<mo>&NotEqual;</mo>
<mn>1</mn>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
</mrow>
<mrow>
<msub>
<mi>s</mi>
<mi>l</mi>
</msub>
<mo>=</mo>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mn>1</mn>
</mtd>
<mtd>
<mrow>
<msub>
<mi>avem</mi>
<mi>l</mi>
</msub>
<mo>&GreaterEqual;</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
</mrow>
</msub>
<mi>a</mi>
<mi>n</mi>
<mi>d</mi>
<mi> </mi>
<msub>
<mi>R</mi>
<mi>l</mi>
</msub>
<mo>&GreaterEqual;</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>l</mi>
</mrow>
</msub>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mn>0</mn>
</mtd>
<mtd>
<mrow>
<msub>
<mi>avem</mi>
<mi>l</mi>
</msub>
<mo><</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
</mrow>
</msub>
<mi>o</mi>
<mi>r</mi>
<mi> </mi>
<msub>
<mi>R</mi>
<mi>l</mi>
</msub>
<mo><</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>l</mi>
</mrow>
</msub>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
</mrow>
<mrow>
<msub>
<mi>s</mi>
<mi>r</mi>
</msub>
<mo>=</mo>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mn>1</mn>
</mtd>
<mtd>
<mrow>
<msub>
<mi>avem</mi>
<mi>r</mi>
</msub>
<mo>&GreaterEqual;</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
<mi>a</mi>
<mi>n</mi>
<mi>d</mi>
<mi> </mi>
<msub>
<mi>R</mi>
<mi>r</mi>
</msub>
<mo>&GreaterEqual;</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>r</mi>
</mrow>
</msub>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mn>0</mn>
</mtd>
<mtd>
<mrow>
<msub>
<mi>avem</mi>
<mi>r</mi>
</msub>
<mo><</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
<mi>o</mi>
<mi>r</mi>
<mi> </mi>
<msub>
<mi>R</mi>
<mi>r</mi>
</msub>
<mo><</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>r</mi>
</mrow>
</msub>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
</mrow>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
</mrow>
</msub>
<mo>=</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>l</mi>
<mi>b</mi>
</mrow>
</msub>
<mo>*</mo>
<mn>0.7</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>l</mi>
</mrow>
</msub>
<mo>=</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>l</mi>
<mi>b</mi>
</mrow>
</msub>
<mo>*</mo>
<mn>0.7</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
</mrow>
</msub>
<mo>=</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>m</mi>
<mi>r</mi>
<mi>b</mi>
</mrow>
</msub>
<mo>*</mo>
<mn>0.7</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>r</mi>
</mrow>
</msub>
<mo>=</mo>
<msub>
<mi>T</mi>
<mrow>
<mi>&theta;</mi>
<mi>r</mi>
<mi>b</mi>
</mrow>
</msub>
<mo>*</mo>
<mn>0.7</mn>
</mtd>
</mtr>
</mtable>
</mfenced>
Wherein, exist=1 represents action of the lift hand close to ear be present, and exist=0 represents to be not present lift hand close to ear
Action;slRepresent the action with the presence or absence of lift hand close to ear in the effective detection region of left side;srRepresent in right side effective detection
Action in region with the presence or absence of lift hand close to ear;
TmlMotion strength threshold when representing to lift hand close to ear in the effective detection region of left side;TθlRepresent in left side effective detection
Motion range threshold value when lifting hand close to ear in region;TmrRepresent to lift hand in the effective detection region of right side close to ear luck
Fatigue resistance threshold value;TθrMotion range threshold value when representing to lift hand close to ear in the effective detection region of right side;
TmlbRepresent in the effective detection region of left side, the mean motion intensity when lift hand of standard is close to ear;TθlbRepresent
In the effective detection region of left side, the mean motion range when lift hand of standard is close to ear;TmrbRepresent in right side effective detection area
In domain, the mean motion intensity when lift hand of standard is close to ear;TθrbRepresent in the effective detection region of right side, the lift of standard
Mean motion range when hand is close to ear.
2. a kind of driver based on svm according to claim 1 takes phone-monitoring method, it is characterised in that:Step
(3) in, the described suitable image region of selection detects and taken the effective detection region of phone detection as hand exercise,
The step of specifically including following order:
(31) harr features and adaboost sorting algorithms are used, detects face location;
(32) according to five, the three front yard layout rule of face, the position that coarse positioning is left and right two;
(33) it is accurately positioned out the position of eyes;
(34) suitable effective image subregion is selected.
3. a kind of driver based on svm according to claim 1 takes phone-monitoring method, it is characterised in that:Step
(5) it is described according to effective detection area image in, the time that the hand of driver is rested on beside ear, and root are monitored in detail
The time length rested on according to the hand of driver beside ear judges whether driver takes phone;Specifically include following suitable
The step of sequence:
(51) using effective detection area image as present image, the Gradient Features of present image are calculated, and to gradient direction angle
Degree is corrected, and the scope of gradient direction is limited to [0 π];
(52) one a width of W of construction, a height of H hough transform window, W and H value are respectively equal to the training sample of hand trainer
This wide high level, and subwindow division is carried out to hough transform window;
(53) hough transform window is slided with a fixed step size in effective detection region, judges the area of hough transform window covering
Domain whether there is hand images;
(54) judge whether hough transform window is slided at the ending of present image;If so, then perform step (55);If it is not,
Then hough transform window sliding performs step (53) again to next position;
(55) a width of W of present image is setimage, a height of Himage, change of scale is carried out to present image, after carrying out change of scale
A width of σ W of imageimage, a height of σ Himage, the image after progress change of scale is set to present image;Wherein, 0 < σ < 1;
(56) judge whether the size of present image is less than minimum threshold;If the size of present image is less than minimum threshold, hold
Row step (57);If the size of present image is more than minimum threshold, returns and perform step (51);
(57) number of the candidate target in the list of candidate region and position, the position of hand in comprehensive descision present frame;
(58) judge to accumulate whether frame number reaches unit frame number;If so, then perform step (59);Step is performed if it is not, then returning
(51);
(59) in statistical unit frame number, the frame number proportion of hand images be present;And according to the frame number institute accounting of hand images
Example judges whether driver is in and takes telephone state.
4. a kind of driver based on svm according to claim 2 takes phone-monitoring method, it is characterised in that:Step
(34) in, the suitable effective image subregion of described selection, specifically realized using below equation:
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>l</mi>
<mi>e</mi>
<mi>f</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>x</mi>
<mo>=</mo>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>l</mi>
</msub>
<mo>.</mo>
<mi>x</mi>
<mo>-</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>w</mi>
<mi>i</mi>
<mi>d</mi>
<mi>t</mi>
<mi>h</mi>
<mo>*</mo>
<mn>0.6</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>l</mi>
<mi>e</mi>
<mi>f</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>y</mi>
<mo>=</mo>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>l</mi>
</msub>
<mo>.</mo>
<mi>y</mi>
<mo>-</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>h</mi>
<mi>e</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
<mo>*</mo>
<mn>0.16</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>l</mi>
<mi>e</mi>
<mi>f</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>w</mi>
<mi>i</mi>
<mi>d</mi>
<mi>t</mi>
<mi>h</mi>
<mo>=</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>w</mi>
<mi>i</mi>
<mi>d</mi>
<mi>t</mi>
<mi>h</mi>
<mo>*</mo>
<mn>0.6</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>l</mi>
<mi>e</mi>
<mi>f</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>h</mi>
<mi>e</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
<mo>=</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>h</mi>
<mi>e</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
<mo>*</mo>
<mn>1.16</mn>
</mtd>
</mtr>
</mtable>
</mfenced>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>r</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>x</mi>
<mo>=</mo>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>r</mi>
</msub>
<mo>.</mo>
<mi>x</mi>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>r</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>y</mi>
<mo>=</mo>
<mi>p</mi>
<mi>o</mi>
<msub>
<mi>int</mi>
<mi>r</mi>
</msub>
<mo>.</mo>
<mi>y</mi>
<mo>-</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>h</mi>
<mi>e</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
<mo>*</mo>
<mn>0.16</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>r</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>w</mi>
<mi>i</mi>
<mi>d</mi>
<mi>t</mi>
<mi>h</mi>
<mo>=</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>w</mi>
<mi>i</mi>
<mi>d</mi>
<mi>t</mi>
<mi>h</mi>
<mo>*</mo>
<mn>0.6</mn>
</mtd>
</mtr>
<mtr>
<mtd>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<msub>
<mi>t</mi>
<mrow>
<mi>r</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
</mrow>
</msub>
<mo>.</mo>
<mi>h</mi>
<mi>e</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
<mo>=</mo>
<mi>r</mi>
<mi>e</mi>
<mi>c</mi>
<mi>t</mi>
<mo>.</mo>
<mi>h</mi>
<mi>e</mi>
<mi>i</mi>
<mi>g</mi>
<mi>h</mi>
<mi>t</mi>
<mo>*</mo>
<mn>1.16</mn>
</mtd>
</mtr>
</mtable>
</mfenced>
Wherein, rectleft、rectrightSub-rectangular areas near the left and right ear of selection is represented respectively, and rect represents to detect
Face location rectangular area, pointl、pointrThe left hand edge point of left eye and the right hand edge point of right eye are represented respectively.
5. a kind of driver based on svm according to claim 3 takes phone-monitoring method, it is characterised in that:Step
(51) in, the Gradient Features of described calculating present image, and gradient direction angle is corrected, specifically using below equation
Realize:
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mi>M</mi>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
<mo>=</mo>
<msqrt>
<mrow>
<msup>
<msub>
<mi>G</mi>
<mi>x</mi>
</msub>
<mn>2</mn>
</msup>
<mo>+</mo>
<msup>
<msub>
<mi>G</mi>
<mi>y</mi>
</msub>
<mn>2</mn>
</msup>
</mrow>
</msqrt>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mi>&theta;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mi>a</mi>
<mi>r</mi>
<mi>c</mi>
<mi>t</mi>
<mi>a</mi>
<mi>n</mi>
<mrow>
<mo>(</mo>
<mfrac>
<msub>
<mi>G</mi>
<mi>y</mi>
</msub>
<msub>
<mi>G</mi>
<mi>x</mi>
</msub>
</mfrac>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<msub>
<mi>G</mi>
<mi>x</mi>
</msub>
<mo>=</mo>
<mi>f</mi>
<mo>(</mo>
<mi>x</mi>
<mo>-</mo>
<mn>1</mn>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
<mo>+</mo>
<mi>f</mi>
<mo>(</mo>
<mi>x</mi>
<mo>+</mo>
<mn>1</mn>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
<mo>-</mo>
<mn>2</mn>
<mo>*</mo>
<mi>f</mi>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mtd>
</mtr>
<mtr>
<mtd>
<msub>
<mi>G</mi>
<mi>y</mi>
</msub>
<mo>=</mo>
<mi>f</mi>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>-</mo>
<mn>1</mn>
<mo>)</mo>
<mo>+</mo>
<mi>f</mi>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>+</mo>
<mn>1</mn>
<mo>)</mo>
<mo>-</mo>
<mn>2</mn>
<mo>*</mo>
<mi>f</mi>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mtd>
</mtr>
</mtable>
</mfenced>
<mrow>
<mi>&theta;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>=</mo>
<mfenced open = "{" close = "">
<mtable>
<mtr>
<mtd>
<mrow>
<mi>&theta;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>+</mo>
<mi>&pi;</mi>
</mrow>
</mtd>
<mtd>
<mrow>
<mi>&theta;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo><</mo>
<mn>0</mn>
</mrow>
</mtd>
</mtr>
<mtr>
<mtd>
<mrow>
<mi>&theta;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
</mrow>
</mtd>
<mtd>
<mrow>
<mi>&theta;</mi>
<mrow>
<mo>(</mo>
<mi>x</mi>
<mo>,</mo>
<mi>y</mi>
<mo>)</mo>
</mrow>
<mo>&GreaterEqual;</mo>
<mn>0</mn>
</mrow>
</mtd>
</mtr>
</mtable>
</mfenced>
</mrow>
Wherein, M (x, y), θ (x, y) represent the gradient magnitude and gradient direction at pixel (x, y) place respectively, and f (x, y) represents pixel
The gray value at (x, y) place;Gx represents the partial derivative in pixel (x, y) place x directions, and Gy represents the local derviation in pixel (x, y) place y directions
Number.
6. a kind of driver based on svm according to claim 3 takes phone-monitoring method, it is characterised in that:Step
(53) it is described to slide hough transform window in effective detection region with a fixed step size in, judge that hough transform window covers
The region of lid whether there is hand images;Specifically realized using the step of following order:
(531) histogram of gradients of each subwindow overlay area is counted;
(532) characteristic vector of hough transform window overlay area is obtained;
(533) characteristic vector of hough transform window overlay area is normalized;
(534) characteristic vector of hough transform window overlay area is sent into hand svm graders, predicts its generic;If
Hand svm graders predict that this feature vector belongs to hand images, then hough transform window position are added into candidate region
In list.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510013139.4A CN104573659B (en) | 2015-01-09 | 2015-01-09 | A kind of driver based on svm takes phone-monitoring method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510013139.4A CN104573659B (en) | 2015-01-09 | 2015-01-09 | A kind of driver based on svm takes phone-monitoring method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104573659A CN104573659A (en) | 2015-04-29 |
CN104573659B true CN104573659B (en) | 2018-01-09 |
Family
ID=53089681
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510013139.4A Active CN104573659B (en) | 2015-01-09 | 2015-01-09 | A kind of driver based on svm takes phone-monitoring method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104573659B (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104966059B (en) * | 2015-06-15 | 2018-04-27 | 安徽创世科技股份有限公司 | Driver based on intelligent monitor system drives detection method of making a phone call |
CN105469073A (en) * | 2015-12-16 | 2016-04-06 | 安徽创世科技有限公司 | Kinect-based call making and answering monitoring method of driver |
CN105868690A (en) * | 2016-03-11 | 2016-08-17 | 博康智能信息技术有限公司 | Method and apparatus for identifying mobile phone use behavior of driver |
CN106022242B (en) * | 2016-05-13 | 2019-05-03 | 哈尔滨工业大学(威海) | Method for identifying call receiving and making of driver in intelligent traffic system |
CN106530730A (en) * | 2016-11-02 | 2017-03-22 | 重庆中科云丛科技有限公司 | Traffic violation detection method and system |
CN108345819B (en) * | 2017-01-23 | 2020-09-15 | 杭州海康威视数字技术股份有限公司 | Method and device for sending alarm message |
CN108509902B (en) * | 2018-03-30 | 2020-07-03 | 湖北文理学院 | Method for detecting call behavior of handheld phone in driving process of driver |
CN110956060A (en) * | 2018-09-27 | 2020-04-03 | 北京市商汤科技开发有限公司 | Motion recognition method, driving motion analysis method, device and electronic equipment |
CN110309764B (en) * | 2019-06-27 | 2021-06-01 | 浙江工业大学 | Multi-stage driver call-making behavior detection method based on deep learning |
CN111523380B (en) * | 2020-03-11 | 2023-06-30 | 浙江工业大学 | Mask wearing condition monitoring method based on face and gesture recognition |
CN111553217A (en) * | 2020-04-20 | 2020-08-18 | 哈尔滨工程大学 | Driver call monitoring method and system |
CN112487990A (en) * | 2020-12-02 | 2021-03-12 | 重庆邮电大学 | DSP-based driver call-making behavior detection method and system |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102567743A (en) * | 2011-12-20 | 2012-07-11 | 东南大学 | Automatic identification method of driver gestures based on video images |
CN102592143A (en) * | 2012-01-09 | 2012-07-18 | 清华大学 | Method for detecting phone holding violation of driver in driving |
CN103129492A (en) * | 2013-03-01 | 2013-06-05 | 公安部第三研究所 | Vehicular intelligent multimedia terminal device |
CN103279750A (en) * | 2013-06-14 | 2013-09-04 | 清华大学 | Detecting method of mobile telephone holding behavior of driver based on skin color range |
CN103366506A (en) * | 2013-06-27 | 2013-10-23 | 北京理工大学 | Device and method for automatically monitoring telephone call behavior of driver when driving |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8315617B2 (en) * | 2009-10-31 | 2012-11-20 | Btpatent Llc | Controlling mobile device functions |
-
2015
- 2015-01-09 CN CN201510013139.4A patent/CN104573659B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102567743A (en) * | 2011-12-20 | 2012-07-11 | 东南大学 | Automatic identification method of driver gestures based on video images |
CN102592143A (en) * | 2012-01-09 | 2012-07-18 | 清华大学 | Method for detecting phone holding violation of driver in driving |
CN103129492A (en) * | 2013-03-01 | 2013-06-05 | 公安部第三研究所 | Vehicular intelligent multimedia terminal device |
CN103279750A (en) * | 2013-06-14 | 2013-09-04 | 清华大学 | Detecting method of mobile telephone holding behavior of driver based on skin color range |
CN103366506A (en) * | 2013-06-27 | 2013-10-23 | 北京理工大学 | Device and method for automatically monitoring telephone call behavior of driver when driving |
Non-Patent Citations (1)
Title |
---|
"面向智能驾驶行为的机器学习";陈雪梅 等;《道路交通与安全》;20141231;第14卷(第6期);第60-64页 * |
Also Published As
Publication number | Publication date |
---|---|
CN104573659A (en) | 2015-04-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104573659B (en) | A kind of driver based on svm takes phone-monitoring method | |
CN109359536B (en) | Passenger behavior monitoring method based on machine vision | |
CN106780612B (en) | Object detecting method and device in a kind of image | |
CN104637058B (en) | A kind of volume of the flow of passengers identify statistical methods based on image information | |
CN106874894A (en) | A kind of human body target detection method based on the full convolutional neural networks in region | |
CN108388871B (en) | Vehicle detection method based on vehicle body regression | |
CN110298265A (en) | Specific objective detection method in a kind of elevator based on YOLO neural network | |
CN103324937A (en) | Method and device for labeling targets | |
CN103390164A (en) | Object detection method based on depth image and implementing device thereof | |
CN107273852A (en) | Escalator floor plates object and passenger behavior detection algorithm based on machine vision | |
CN108364466A (en) | A kind of statistical method of traffic flow based on unmanned plane traffic video | |
CN103632427B (en) | A kind of gate cracking protection method and gate control system | |
CN110197166A (en) | A kind of car body loading condition identification device and method based on image recognition | |
CN110728225A (en) | High-speed face searching method for attendance checking | |
CN107909027A (en) | It is a kind of that there is the quick human body target detection method for blocking processing | |
CN107491715A (en) | A kind of subway carriage passenger flow statistical method, apparatus and system based on video analysis | |
CN111460924B (en) | Gate ticket-evading behavior detection method based on target detection | |
CN108038424A (en) | A kind of vision automated detection method suitable for working at height | |
CN111292432A (en) | Vehicle charging type distinguishing method and device based on vehicle type recognition and wheel axle detection | |
CN109829382A (en) | The abnormal object early warning tracing system and method for Behavior-based control feature intelligent analysis | |
CN112488042B (en) | Pedestrian traffic bottleneck discrimination method and system based on video analysis | |
CN111523415A (en) | Image-based two-passenger one-dangerous vehicle detection method and device | |
CN109543617A (en) | The detection method of intelligent vehicle movement traffic information based on YOLO target detection technique | |
CN106682600A (en) | Method and terminal for detecting targets | |
CN109598200A (en) | A kind of digital image recognition system and method for hot-metal bottle tank number |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |