CN104573659B - A kind of driver based on svm takes phone-monitoring method - Google Patents

A kind of driver based on svm takes phone-monitoring method Download PDF

Info

Publication number
CN104573659B
CN104573659B CN201510013139.4A CN201510013139A CN104573659B CN 104573659 B CN104573659 B CN 104573659B CN 201510013139 A CN201510013139 A CN 201510013139A CN 104573659 B CN104573659 B CN 104573659B
Authority
CN
China
Prior art keywords
msub
mrow
mtd
mtr
hand
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510013139.4A
Other languages
Chinese (zh)
Other versions
CN104573659A (en
Inventor
张卡
何佳
曹昌龙
尼秀明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ANHUI QINGXIN INTERNET INFORMATION TECHNOLOGY Co Ltd
Original Assignee
ANHUI QINGXIN INTERNET INFORMATION TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ANHUI QINGXIN INTERNET INFORMATION TECHNOLOGY Co Ltd filed Critical ANHUI QINGXIN INTERNET INFORMATION TECHNOLOGY Co Ltd
Priority to CN201510013139.4A priority Critical patent/CN104573659B/en
Publication of CN104573659A publication Critical patent/CN104573659A/en
Application granted granted Critical
Publication of CN104573659B publication Critical patent/CN104573659B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • Nitrogen And Oxygen Or Sulfur-Condensed Heterocyclic Ring Systems (AREA)

Abstract

The step of the present invention relates to a kind of driver based on svm to take phone-monitoring method, and the monitoring method includes following order:Establish Face datection grader and hand images detect grader as the hand of training positive sample during taking phone;The driving condition image of collection driver in real time;Suitable image region is selected to detect and take the effective detection region that phone detects as hand exercise;According to effective detection area image, real-time judge driver whether there is the warming-up exercise for taking phone;According to effective detection area image, the time that the hand of driver is rested on beside ear is monitored in detail, and judge whether driver takes phone according to the length of residence time;The real-time video that driver is taken to phone is sent to remote server, and receives the order of remote server transmission.

Description

A kind of driver based on svm takes phone-monitoring method
Technical field
The present invention relates to safe driving technical field, and in particular to a kind of driver based on svm takes phone-monitoring side Method.
Background technology
With the rapid growth of car ownership, people enjoy the facility of traffic and it is quick while, also along with All kinds of traffic accidents take place frequently, and cause huge personnel and economic loss.Cause the factor of traffic accident a lot, driver drives It is one of important inducement that phone is taken in way.Due to that can not obtain the driving behavior video of driver in real time, some passenger traffics and The supervision department of Shipping enterprises can only can not carry out monitoring in advance and in advance using foundation of the deduction as division responsibility afterwards It is anti-.Therefore, monitor driver in real time takes phone behavior, feeds back to supervision department of transport enterprise in time and is prevented, right In avoiding major traffic accidents, play a part of can not be substituted.
SVMs (svm) is a kind of supervised learning algorithm, and it is solving small sample, the knowledge of non-linear and high dimensional pattern Many distinctive advantages are shown in not, the VC dimensions that this method is built upon Statistical Learning Theory are theoretical former with Structural risk minization On the basis of reason, according to limited sample information in the complexity (the study precision i.e. to specific training sample) of model and study Ability (i.e. without error identify arbitrary sample ability) between seek best compromise, to obtain best Generalization Ability (or Claim generalization ability).
At present, the monitoring of phone behavior is taken for driver, conventional technical method has following several:
(1) be monitored based on mobile phone signal, such method by driver's cabin place a mobile phone signal detector, According to the different degrees of of signal fluctuation, judge whether to take phone behavior.Such method can reach on goods stock Monitoring takes the effect of phone.But on passenger stock, more interference be present, such as the mobile phone signal interference of passenger on car Deng serious missing inspection and flase drop being present, can not realize that comprehensive monitoring driver takes phone behavior in real time.
(2) it is monitored based on video image, such method by monitoring whether driver's both hands are placed on steering wheel in real time On, once there is certain hand departure direction disk, that is, it is considered as to take phone.Serious flase drop be present in such method, because There is the custom of a hand steered steering wheel in many drivers, therefore, this method is in actual environment using having the problem of larger.
The content of the invention
It is an object of the invention to provide a kind of driver based on svm to take phone-monitoring method, and the monitoring method is Based on video image processing technology, judge to take phone behavior by monitoring the state of driver's ear region hand in real time, With the features such as monitoring degree of accuracy height, missing inspection flase drop is less, and speed is fast, and cost is low.
The technical scheme is that:
The step of a kind of driver based on svm takes phone-monitoring method, and the monitoring method includes following order:
(1) establish Face datection grader and hand images detect classification as the hand of training positive sample during taking phone Device;
(2) the driving condition image of driver is gathered in real time;
(3) suitable image region is selected to detect and take the effective detection region that phone detects as hand exercise;
(4) warming-up exercise for taking phone whether there is according to effective detection area image, real-time judge driver;If It is then to perform step (5);Step (2) is performed if it is not, then returning;
(5) time rested on according to effective detection area image, the detailed hand for monitoring driver beside ear, and according to The time length that the hand of driver is rested on beside ear judges whether driver takes phone;If so, then perform step (6);Step (2) is performed if it is not, then returning;
(6) real-time video that driver is taken to phone is sent to remote server, and receives remote server transmission Order.
In step (3), the described suitable image region of selection detects and taken phone detection as hand exercise Effective detection region, the step of specifically including following order:
(31) harr features and adaboost sorting algorithms are used, detects face location;
(32) according to five, the three front yard layout rule of face, the position that coarse positioning is left and right two;
(33) it is accurately positioned out the position of eyes;
(34) suitable effective image subregion is selected.
Described according to effective detection area image in step (4), real-time judge driver, which whether there is, takes phone Warming-up exercise, the step of specifically including following order:
(41) uniform point sampling is carried out respectively in left and right effective detection region;
(42) accurate tracking of sampled point is carried out;
(43) movable information of correct tracking sampling point is obtained;
(44) statistical nature in effective detection region is obtained, the statistical nature of described detection zone includes left effective detection The mean motion intensity avem in regionl, right effective detection region mean motion intensity avemr, left effective detection region motion Range RlWith the motion range R in right effective detection regionr
(45) judge whether to lift action of the hand close to ear.
It is described according to effective detection area image in step (5), monitor that the hand of driver is rested on beside ear in detail Time, and the time length rested on according to the hand of driver beside ear judges whether driver takes phone;Tool Body includes the step of following order:
(51) using effective detection area image as present image, the Gradient Features of present image are calculated, and to gradient side It is corrected to angle, the scope of gradient direction is limited to [0 π];
(52) one a width of W of construction, a height of H hough transform window, W and H value are respectively equal to the instruction of hand trainer Practice the wide high level of sample, and subwindow division is carried out to hough transform window;
(53) hough transform window is slided with a fixed step size in effective detection region, judges that hough transform window covers Region whether there is hand images;
(54) judge whether hough transform window is slided at the ending of present image;If so, then perform step (55);If No, then hough transform window sliding to next position, performs step (53) again;
(55) a width of W of present image is setimage, a height of Himage, change of scale is carried out to present image, carries out yardstick change A width of σ W of image after changingimage, a height of σ Himage, the image after progress change of scale is set to present image;Wherein, 0 < σ < 1;
(56) judge whether the size of present image is less than minimum threshold;If the size of present image is less than minimum threshold, Then perform step (57);If the size of present image is more than minimum threshold, returns and perform step (51);
(57) number of the candidate target in the list of candidate region and position, the position of hand in comprehensive descision present frame Put;
(58) judge to accumulate whether frame number reaches unit frame number;If so, then perform step (59);Performed if it is not, then returning Step (51);
(59) in statistical unit frame number, the frame number proportion of hand images be present;And according to the frame number institute of hand images Accounting example judges whether driver is in and takes telephone state.
In step (34), the suitable effective image subregion of described selection, specifically realized using below equation:
Wherein, rectleft、rectrightSub-rectangular areas near the left and right ear of selection is represented respectively, and rect represents inspection The face location rectangular area measured, pointl、pointrThe left hand edge point of left eye and the right hand edge point of right eye are represented respectively.
In step (43), the movable information of the described correct tracking sampling point of acquisition, specifically realized using below equation:
Wherein, M (i) represents the motion amplitude of ith sample point, and θ (i) represents the direction of motion of ith sample point, xpointiRepresent coordinate of the ith sample point on present frame, ypointiCoordinate of the ith sample point on previous frame is represented, Dx represents amount of exercise of some sampled point in x directions, and Dy represents amount of exercise of some sampled point in y directions.
In step (44), the statistical nature in described acquisition effective detection region, specifically realized using below equation:
Wherein, sumlRepresent that there is the motion amplitude of the sampled point substantially moved, N in left detection zonemlRepresent left detection There is the number of the sampled point substantially moved, N in regionlRepresent that left detection zone is interior with obvious motion and close to ear The number of sampled point;
sumrRepresent that there is the motion amplitude of the sampled point substantially moved, N in right detection zonemrRepresent in right detection zone Number with the sampled point substantially moved, NrRepresent there is obvious motion and to the close sampled point of ear in right detection zone Number;
Ml(i) motion amplitude of i-th of correct tracking sampling point in left detection zone, θ are representedl(i) left detection zone is represented The direction of motion of interior i-th of correct tracking sampling point;
Mr(i) motion amplitude of i-th of correct tracking sampling point in right detection zone, θ are representedr(i) right detection zone is represented The direction of motion of interior i-th of correct tracking sampling point.
In step (45), described judges whether to lift action of the hand close to ear, is specifically realized using below equation:
Wherein, exist=1 represents action of the lift hand close to ear be present, and exist=0 represents to be not present lift hand close to ear Piece action, slRepresent the action with the presence or absence of lift hand close to ear in the effective detection region of left side;srRepresent effective on right side Action in detection zone with the presence or absence of lift hand close to ear;TmlWhen representing to lift hand close to ear in the effective detection region of left side Motion strength threshold;TθlMotion range threshold value when representing to lift hand close to ear in the effective detection region of left side;TmrRepresent Motion strength threshold when lifting hand close to ear in the effective detection region of right side;TθrRepresent that hand is lifted in the effective detection region of right side to be leaned on Motion range threshold value during nearly ear;TmlbRepresent in the effective detection region of left side, being averaged when the lift hand of standard is close to ear Exercise intensity;TθlbRepresent in the effective detection region of left side, the mean motion range when lift hand of standard is close to ear;TmrbTable Show in the effective detection region of right side, the mean motion intensity when lift hand of standard is close to ear;TθrbExpression is effectively examined on right side Survey in region, the mean motion range when lift hand of standard is close to ear.
In step (51), the Gradient Features of described calculating present image, and gradient direction angle is corrected, specifically Realized using below equation:
Wherein, M (x, y), θ (x, y) represent the gradient magnitude and gradient direction at pixel (x, y) place respectively, and f (x, y) is represented The gray value at pixel (x, y) place.Gx represents the partial derivative in pixel (x, y) place x directions, and Gy represents the inclined of pixel (x, y) place y directions Derivative.
It is described to slide hough transform window in effective detection region with a fixed step size in step (53), judge square The region of shape detection window covering whether there is hand images;Specifically realized using the step of following order:
(531) histogram of gradients of each subwindow overlay area is counted;
(532) characteristic vector of hough transform window overlay area is obtained;
(533) characteristic vector of hough transform window overlay area is normalized;
(534) characteristic vector of hough transform window overlay area is sent into hand svm graders, predicts its affiliated class Not;If hand svm graders predict that this feature vector belongs to hand images, hough transform window position is added into candidate In zone list.
Phone-monitoring method is taken compared to other drivers, the present invention uses video image processing technology, is supervised by triggering Control pattern monitors the hand existence of driver's ear region in real time, and judgement takes phone behavior, has monitoring degree of accuracy height, The features such as missing inspection flase drop is less, and speed is fast, and cost is low.
Brief description of the drawings
Fig. 1 is flow chart of the method for the present invention;
Fig. 2 is trigger module workflow diagram;
Fig. 3 is monitoring module workflow diagram;
Fig. 4 is training positive sample master drawing;
Fig. 5 is the effective detection zone schematic diagram of positioning;
Fig. 6 is point sampling design sketch;
Fig. 7 is the subwindow division schematic diagram of hough transform window;
Fig. 8 is hough transform window sliding schematic diagram;
Fig. 9 is hand candidate test position design sketch, wherein, rectangle frame 1 is effective detection region, and rectangle frame 2 is candidate Detection zone;
Figure 10 is the final test position design sketch of hand, wherein, rectangle frame 1 is effective detection region, and rectangle frame 2 is final Test position.
Embodiment
The present invention is further illustrated below in conjunction with the accompanying drawings.
As shown in figure 1, in embodiment, the driver of the present invention based on svm takes phone-monitoring system and included initially Change module, acquisition module, locating module, trigger module, monitoring module and communication module.The driver based on svm takes electricity The specific implementation step for talking about monitoring system is as follows:
S1, perform initialization module.
The function of initial module is loading and the necessary grader file of training system, to comprise the following steps that:
S11, the existing Face datection grader file of loading.
S12, as shown in figure 4, collecting hand images when taking phone is used as positive sample, examined based on the theoretical training hands of svm Survey grader file.
S2, perform acquisition module.
The function of acquisition module is to gather the head image of the driving condition image, mainly driver of driver in real time.
S3, perform locating module.
The function of locating module is to select suitable image region to be detected as hand exercise and take what phone detected Effective coverage, mainly select the regional area near left and right ear.The module can greatly lift detection speed, remove More interference regions, as shown in Figure 5.The module comprises the following steps that:
S31, using harr features and adaboost sorting algorithms, detect face location.
S32, five, the three front yard layout rule based on face, the position that coarse positioning is left and right two.
S33, the position for being accurately positioned out eyes.
S34, suitable effective image subregion selected using formula (1) and formula (2).
Wherein, rectleft、rectrightSub-rectangular areas near the left and right ear of selection is represented respectively, and rect represents inspection The face location rectangular area measured, pointl、pointrThe left hand edge point of left eye and the right hand edge point of right eye are represented respectively.
S4, judge trigger module open and-shut mode, if trigger module is in opening, step S5 is performed, if trigger mode Block is closed, then performs step S7.
S5, perform trigger module.
The function of trigger module is that real-time judge driver whether there is the warming-up exercise for taking phone, in particular to drive Action of the person of sailing with the presence or absence of lift hand close to ear.If it does, explanation driver is possible to preparation and takes phone, now exit Trigger module, return and open monitoring module signal;If there is no the action, system may proceed to carry out real-time judge, under wait Once take the appearance of phone warming-up exercise.As shown in Fig. 2 the module comprises the following steps that:
S51, carry out uniform point sampling, effect such as Fig. 6 respectively in left and right effective detection region.
S52, the accurate tracking for carrying out sampled point.Specific track algorithm is referring to document:Forward-Backward Error: Automatic Detection of Tracking Failures, Zdenek Kalal, Krystian Mikolajczyk, Jiri Matas, Pattern Recognition (ICPR), 2010 20th International Conference on.
S53, the movable information for obtaining using formula (3) and formula (4) correct tracking sampling point;
Wherein, M (i) represents the motion amplitude of ith sample point, and θ (i) represents the direction of motion of ith sample point, xpointiRepresent coordinate of the ith sample point on present frame, ypointiCoordinate of the ith sample point on previous frame is represented, Dx represents amount of exercise of some sampled point in x directions, and Dy represents amount of exercise of some sampled point in y directions.
S54, the statistical nature for obtaining using formula (5) and formula (6) effective detection region.Described effective detection region Statistical nature include the mean motion intensity avem in left side effective detection regionl, the right effective detection region mean motion Intensity avemr, left side effective detection region motion range RlWith the motion range R in the right effective detection regionr
Wherein, sumlRepresent that there is the motion amplitude of the sampled point substantially moved, N in left detection zonemlRepresent left detection There is the number of the sampled point substantially moved, N in regionlRepresent that left detection zone is interior with obvious motion and close to ear The number of sampled point;sumrRepresent that there is the motion amplitude of the sampled point substantially moved, N in right detection zonemrRepresent right detection There is the number of the sampled point substantially moved, N in regionrRepresent that right detection zone is interior with obvious motion and close to ear The number of sampled point;Ml(i) motion amplitude of i-th of correct tracking sampling point in left detection zone, θ are representedl(i) left inspection is represented Survey the direction of motion of i-th of correct tracking sampling point in region;Mr(i) i-th of correct tracking sampling in right detection zone is represented The motion amplitude of point, θr(i) direction of motion of i-th of correct tracking sampling point in right detection zone is represented.
S55, using formula (7)-(10), judge whether to lift action of the hand close to ear.If in the presence of exiting triggering Module;If it is not, then continue executing with trigger module.
Wherein, exist=1 represents action of the lift hand close to ear be present, and exist=0 represents to be not present lift hand close to ear Piece action.slRepresent the action with the presence or absence of lift hand close to ear in the effective detection region of left side;srRepresent effective on right side Action in detection zone with the presence or absence of lift hand close to ear;TmlWhen representing to lift hand close to ear in the effective detection region of left side Motion strength threshold;TθlMotion range threshold value when representing to lift hand close to ear in the effective detection region of left side;TmrRepresent Motion strength threshold when lifting hand close to ear in the effective detection region of right side;TθrRepresent that hand is lifted in the effective detection region of right side to be leaned on Motion range threshold value during nearly ear;TmlbRepresent in the effective detection region of left side, being averaged when the lift hand of standard is close to ear Exercise intensity;TθlbRepresent in the effective detection region of left side, the mean motion range when lift hand of standard is close to ear;TmrbTable Show in the effective detection region of right side, the mean motion intensity when lift hand of standard is close to ear;TθrbExpression is effectively examined on right side Survey in region, the mean motion range when lift hand of standard is close to ear.
S6, judge whether to trigger monitoring module.If in the presence of the warming-up exercise for taking phone, monitoring module will be triggered out Open, carry out in-depth monitoring, simultaneously close off trigger module.If in the absence of the warming-up exercise for taking phone, it is returned directly to gather mould Block, the triggering carried out next time judge.
S7, perform monitoring module.
The function of monitoring module is to monitor the time that the hand of driver is rested on beside ear in detail.If the hand of driver The time long enough rested on beside ear, illustrate that driver takes phone, then return to open communication module by signal.If It is no, then illustrate that this triggering unlatching belongs to erroneous judgement.As shown in figure 3, the module comprises the following steps that:
S71, using effective detection area image as present image, using formula (11) and formula (12), calculate current The Gradient Features of image, and gradient direction angle is corrected using formula (13), the scope of gradient direction is limited to [0 π]。
Wherein, M (x, y), θ (x, y) represent the gradient magnitude and gradient direction at pixel (x, y) place, and f (x, y) represents pixel The gray value at (x, y) place.
The one wide high respectively W and H hough transform window of S72, construction, W and H value are equal to hand svm graders The wide high level of training sample.As shown in fig. 7, subwindow division is carried out to hough transform window.Partitioning standards are:Driver does not connect During phone, due to the interference beside ear without hand, now the contouring head in effective detection region is generally vertical, because This, the narrow rectangular area in selection longitudinal direction is rational as subwindow.
S73, as shown in figure 8, hough transform window is slided with a fixed step size in effective detection region, and judge rectangle The region of detection window covering whether there is hand images, comprise the following steps that:
S731, the histogram of gradients hist [bin] for counting using formula (14) each subwindow overlay area.
Wherein, floor (x) represents to choose the maximum integer no more than x, and bin span is 1~10.
S732, the histogram of gradients of all subwindow overlay areas of series connection, obtain the feature of detection window overlay area to Amount.
S733, the characteristic vector normalization by hough transform window overlay area;
S734, judge whether detection window overlay area belongs to hand images, effect such as Fig. 9.In Fig. 9, square frame 1 represents left Right effective detection region, square frame 2 represent that hough transform window overlay area belongs to hand images.Specific method is by detection window The characteristic vector of overlay area is sent into hand svm graders, predicts its generic.If belonging to hand images, rectangle is examined Window position is surveyed to be put into the list of candidate region.
S74, judge whether detection window is slided at the ending of present image.If not reaching image ending, rectangle inspection Window sliding is surveyed to next position, and performs step S73 again;If having reached image ending, step S75 is performed.
S75, present image size is set as Wimage、Himage, change of scale is carried out to present image, then schemed after change of scale As size is changed into σ Wimage、σHimage, and the image after change of scale is designated as present image.Wherein 0 < σ < 1.
S76, judge whether present image size is less than minimum threshold.If being less than thresholding, illustrate that present image is too small, Through that there can not possibly be hand, step S77 is performed;If being more than thresholding, illustrate in present image it is possible to hand be present, perform Step S71 continues detection process.
S77, the number of candidate target in the list of candidate region and position, the position of hand in comprehensive descision present frame Put, effect such as Figure 10.In Figure 10, square frame 1 represents left and right effective detection region, and square frame 2 represents hough transform window overlay area Belong to hand images.
S78, judge to accumulate whether frame number reaches unit frame number.If reaching, step S79 is performed;If it is not, then perform step S71 continues detection process.
In S79, statistical unit frame number, the frame number proportion of hand images be present.If the ratio is more than certain threshold value, Illustrate that current driver's are in and take telephone state.
S710, trigger module is opened, close monitoring module.
S8, perform communication module.
The function of communication module is that, when driver, which is in, takes telephone state, the module sends to remote server and driven The person of sailing takes the real-time video of phone, and now supervision department of transport enterprise can timely be handled by the video;If Need to converse with driver, remote command can also be received by the module.
Embodiment described above is only that the preferred embodiment of the present invention is described, not to the model of the present invention Enclose and be defined, on the premise of design spirit of the present invention is not departed from, technical side of the those of ordinary skill in the art to the present invention The various modifications and improvement that case is made, it all should fall into the protection domain of claims of the present invention determination.

Claims (6)

1. a kind of driver based on svm takes phone-monitoring method, it is characterised in that:The monitoring method includes following order Step:
(1) establish Face datection grader and hand images detect grader as the hand of training positive sample during taking phone;
(2) the driving condition image of driver is gathered in real time;
(3) suitable image region is selected to detect and take the effective detection region that phone detects as hand exercise;
(4) warming-up exercise for taking phone whether there is according to effective detection area image, real-time judge driver;If so, then Perform step (5);Step (2) is performed if it is not, then returning;
(5) according to effective detection area image, the time that the hand of driver is rested on beside ear is monitored in detail, and according to driving The time length that the hand of member is rested on beside ear judges whether driver takes phone;If so, then perform step (6); Step (2) is performed if it is not, then returning;
(6) real-time video that driver is taken to phone is sent to remote server, and receives the order of remote server transmission;
It is described that the preparation for taking phone whether there is according to effective detection area image, real-time judge driver in step (4) Action, the step of specifically including following order:
(41) uniform point sampling is carried out respectively in left and right effective detection region;
(42) accurate tracking of sampled point is carried out;
(43) movable information of correct tracking sampling point is obtained;
(44) statistical nature in effective detection region is obtained, the statistical nature of described detection zone includes left effective detection region Mean motion intensity aveml, right effective detection region mean motion intensity avemr, left effective detection region motion range RlWith the motion range R in right effective detection regionr
(45) judge whether to lift action of the hand close to ear;
In step (43), the movable information of the described correct tracking sampling point of acquisition, specifically realized using below equation:
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mi>M</mi> <mo>(</mo> <mi>i</mi> <mo>)</mo> <mo>=</mo> <mi>s</mi> <mi>q</mi> <mi>r</mi> <mi>t</mi> <mo>(</mo> <msup> <msub> <mi>D</mi> <mi>x</mi> </msub> <mn>2</mn> </msup> <mo>+</mo> <msup> <msub> <mi>D</mi> <mi>y</mi> </msub> <mn>2</mn> </msup> <mo>)</mo> </mtd> </mtr> <mtr> <mtd> <mi>&amp;theta;</mi> <mo>(</mo> <mi>i</mi> <mo>)</mo> <mo>=</mo> <mi>a</mi> <mi>r</mi> <mi>c</mi> <mi>t</mi> <mi>a</mi> <mi>n</mi> <mo>(</mo> <mfrac> <msub> <mi>D</mi> <mi>y</mi> </msub> <msub> <mi>D</mi> <mi>x</mi> </msub> </mfrac> <mo>)</mo> </mtd> </mtr> </mtable> </mfenced>
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <msub> <mi>D</mi> <mi>x</mi> </msub> <mo>=</mo> <mi>x</mi> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>i</mi> </msub> <mo>.</mo> <mi>x</mi> <mo>-</mo> <mi>y</mi> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>i</mi> </msub> <mo>.</mo> <mi>x</mi> </mtd> </mtr> <mtr> <mtd> <msub> <mi>D</mi> <mi>y</mi> </msub> <mo>=</mo> <mi>x</mi> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>i</mi> </msub> <mo>.</mo> <mi>y</mi> <mo>-</mo> <mi>y</mi> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>i</mi> </msub> <mo>.</mo> <mi>y</mi> </mtd> </mtr> </mtable> </mfenced>
Wherein, M (i) represents the motion amplitude of ith sample point, and θ (i) represents the direction of motion of ith sample point, xpointiTable Show coordinate of the ith sample point on present frame, ypointiCoordinate of the ith sample point on previous frame is represented, Dx represents certain For individual sampled point in the amount of exercise in x directions, Dy represents amount of exercise of some sampled point in y directions;
In step (44), the statistical nature in described acquisition effective detection region, specifically realized using below equation:
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mi>a</mi> <mi>v</mi> <mi>e</mi> <msub> <mi>m</mi> <mi>l</mi> </msub> <mo>=</mo> <mfrac> <mrow> <msub> <mi>sum</mi> <mi>l</mi> </msub> </mrow> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>l</mi> </mrow> </msub> </mfrac> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>R</mi> <mi>l</mi> </msub> <mo>=</mo> <mfrac> <msub> <mi>N</mi> <mi>l</mi> </msub> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>l</mi> </mrow> </msub> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>avem</mi> <mi>r</mi> </msub> <mo>=</mo> <mfrac> <mrow> <msub> <mi>sum</mi> <mi>r</mi> </msub> </mrow> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>R</mi> <mi>r</mi> </msub> <mo>=</mo> <mfrac> <msub> <mi>N</mi> <mi>r</mi> </msub> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> </mfrac> </mrow> </mtd> </mtr> </mtable> </mfenced>
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <msub> <mi>sum</mi> <mi>l</mi> </msub> <mo>=</mo> <msub> <mi>sum</mi> <mi>l</mi> </msub> <mo>+</mo> <msub> <mi>M</mi> <mi>l</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> </mrow> </mtd> <mtd> <mrow> <msub> <mi>M</mi> <mi>l</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>2</mn> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mn>1</mn> </mrow> </msub> <mo>+</mo> <mn>1</mn> </mrow> </mtd> <mtd> <mrow> <msub> <mi>M</mi> <mi>l</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>2</mn> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>N</mi> <mi>l</mi> </msub> <mo>=</mo> <msub> <mi>N</mi> <mi>l</mi> </msub> <mo>+</mo> <mn>1</mn> </mrow> </mtd> <mtd> <mrow> <msub> <mi>M</mi> <mi>l</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>2</mn> <mi>a</mi> <mi>n</mi> <mi>d</mi> <mfrac> <mi>&amp;pi;</mi> <mn>4</mn> </mfrac> <mo>&lt;</mo> <msub> <mi>&amp;theta;</mi> <mi>l</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;le;</mo> <mfrac> <mi>&amp;pi;</mi> <mn>2</mn> </mfrac> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>sum</mi> <mi>r</mi> </msub> <mo>=</mo> <msub> <mi>sum</mi> <mi>r</mi> </msub> <mo>+</mo> <msub> <mi>M</mi> <mi>r</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> </mrow> </mtd> <mtd> <mrow> <msub> <mi>M</mi> <mi>r</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>2</mn> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <msub> <mi>N</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> <mo>+</mo> <mn>1</mn> </mrow> </mtd> <mtd> <mrow> <msub> <mi>M</mi> <mi>r</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>2</mn> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <msub> <mi>N</mi> <mi>r</mi> </msub> <mo>=</mo> <msub> <mi>N</mi> <mi>r</mi> </msub> <mo>+</mo> <mn>1</mn> </mrow> </mtd> <mtd> <mrow> <msub> <mi>M</mi> <mi>r</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>2</mn> <mi>a</mi> <mi>n</mi> <mi>d</mi> <mfrac> <mi>&amp;pi;</mi> <mn>2</mn> </mfrac> <mo>&lt;</mo> <msub> <mi>&amp;theta;</mi> <mi>r</mi> </msub> <mrow> <mo>(</mo> <mi>i</mi> <mo>)</mo> </mrow> <mo>&amp;le;</mo> <mfrac> <mrow> <mn>3</mn> <mo>*</mo> <mi>&amp;pi;</mi> </mrow> <mn>4</mn> </mfrac> </mrow> </mtd> </mtr> </mtable> </mfenced>
Wherein, sumlRepresent that there is the motion amplitude of the sampled point substantially moved, N in left detection zonemlRepresent left detection zone The interior number with the sampled point substantially moved, NlRepresent there is obvious motion and to the close sampling of ear in left detection zone The number of point;
sumrRepresent that there is the motion amplitude of the sampled point substantially moved, N in right detection zonemrRepresent have in right detection zone The number of the sampled point substantially moved, NrRepresent there is obvious motion and to the number of the close sampled point of ear in right detection zone Mesh;
Ml(i) motion amplitude of i-th of correct tracking sampling point in left detection zone, θ are representedl(i) the is represented in left detection zone The direction of motion of i correct tracking sampling points;
Mr(i) motion amplitude of i-th of correct tracking sampling point in right detection zone, θ are representedr(i) the is represented in right detection zone The direction of motion of i correct tracking sampling points;
In step (45), described judges whether to lift action of the hand close to ear, is specifically realized using below equation:
<mrow> <mi>e</mi> <mi>x</mi> <mi>i</mi> <mi>s</mi> <mi>t</mi> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mn>1</mn> </mtd> <mtd> <mrow> <msub> <mi>s</mi> <mi>l</mi> </msub> <mo>+</mo> <msub> <mi>s</mi> <mi>r</mi> </msub> <mo>=</mo> <mn>1</mn> </mrow> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mrow> <msub> <mi>s</mi> <mi>l</mi> </msub> <mo>+</mo> <msub> <mi>s</mi> <mi>r</mi> </msub> <mo>&amp;NotEqual;</mo> <mn>1</mn> </mrow> </mtd> </mtr> </mtable> </mfenced> </mrow>
<mrow> <msub> <mi>s</mi> <mi>l</mi> </msub> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mn>1</mn> </mtd> <mtd> <mrow> <msub> <mi>avem</mi> <mi>l</mi> </msub> <mo>&amp;GreaterEqual;</mo> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>l</mi> </mrow> </msub> <mi>a</mi> <mi>n</mi> <mi>d</mi> <mi> </mi> <msub> <mi>R</mi> <mi>l</mi> </msub> <mo>&amp;GreaterEqual;</mo> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>l</mi> </mrow> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mrow> <msub> <mi>avem</mi> <mi>l</mi> </msub> <mo>&lt;</mo> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>l</mi> </mrow> </msub> <mi>o</mi> <mi>r</mi> <mi> </mi> <msub> <mi>R</mi> <mi>l</mi> </msub> <mo>&lt;</mo> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>l</mi> </mrow> </msub> </mrow> </mtd> </mtr> </mtable> </mfenced> </mrow>
<mrow> <msub> <mi>s</mi> <mi>r</mi> </msub> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mn>1</mn> </mtd> <mtd> <mrow> <msub> <mi>avem</mi> <mi>r</mi> </msub> <mo>&amp;GreaterEqual;</mo> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> <mi>a</mi> <mi>n</mi> <mi>d</mi> <mi> </mi> <msub> <mi>R</mi> <mi>r</mi> </msub> <mo>&amp;GreaterEqual;</mo> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>r</mi> </mrow> </msub> </mrow> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mrow> <msub> <mi>avem</mi> <mi>r</mi> </msub> <mo>&lt;</mo> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> <mi>o</mi> <mi>r</mi> <mi> </mi> <msub> <mi>R</mi> <mi>r</mi> </msub> <mo>&lt;</mo> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>r</mi> </mrow> </msub> </mrow> </mtd> </mtr> </mtable> </mfenced> </mrow>
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>l</mi> <mi>b</mi> </mrow> </msub> <mo>*</mo> <mn>0.7</mn> </mtd> </mtr> <mtr> <mtd> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>l</mi> </mrow> </msub> <mo>=</mo> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>l</mi> <mi>b</mi> </mrow> </msub> <mo>*</mo> <mn>0.7</mn> </mtd> </mtr> <mtr> <mtd> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <msub> <mi>T</mi> <mrow> <mi>m</mi> <mi>r</mi> <mi>b</mi> </mrow> </msub> <mo>*</mo> <mn>0.7</mn> </mtd> </mtr> <mtr> <mtd> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>r</mi> </mrow> </msub> <mo>=</mo> <msub> <mi>T</mi> <mrow> <mi>&amp;theta;</mi> <mi>r</mi> <mi>b</mi> </mrow> </msub> <mo>*</mo> <mn>0.7</mn> </mtd> </mtr> </mtable> </mfenced>
Wherein, exist=1 represents action of the lift hand close to ear be present, and exist=0 represents to be not present lift hand close to ear Action;slRepresent the action with the presence or absence of lift hand close to ear in the effective detection region of left side;srRepresent in right side effective detection Action in region with the presence or absence of lift hand close to ear;
TmlMotion strength threshold when representing to lift hand close to ear in the effective detection region of left side;TθlRepresent in left side effective detection Motion range threshold value when lifting hand close to ear in region;TmrRepresent to lift hand in the effective detection region of right side close to ear luck Fatigue resistance threshold value;TθrMotion range threshold value when representing to lift hand close to ear in the effective detection region of right side;
TmlbRepresent in the effective detection region of left side, the mean motion intensity when lift hand of standard is close to ear;TθlbRepresent In the effective detection region of left side, the mean motion range when lift hand of standard is close to ear;TmrbRepresent in right side effective detection area In domain, the mean motion intensity when lift hand of standard is close to ear;TθrbRepresent in the effective detection region of right side, the lift of standard Mean motion range when hand is close to ear.
2. a kind of driver based on svm according to claim 1 takes phone-monitoring method, it is characterised in that:Step (3) in, the described suitable image region of selection detects and taken the effective detection region of phone detection as hand exercise, The step of specifically including following order:
(31) harr features and adaboost sorting algorithms are used, detects face location;
(32) according to five, the three front yard layout rule of face, the position that coarse positioning is left and right two;
(33) it is accurately positioned out the position of eyes;
(34) suitable effective image subregion is selected.
3. a kind of driver based on svm according to claim 1 takes phone-monitoring method, it is characterised in that:Step (5) it is described according to effective detection area image in, the time that the hand of driver is rested on beside ear, and root are monitored in detail The time length rested on according to the hand of driver beside ear judges whether driver takes phone;Specifically include following suitable The step of sequence:
(51) using effective detection area image as present image, the Gradient Features of present image are calculated, and to gradient direction angle Degree is corrected, and the scope of gradient direction is limited to [0 π];
(52) one a width of W of construction, a height of H hough transform window, W and H value are respectively equal to the training sample of hand trainer This wide high level, and subwindow division is carried out to hough transform window;
(53) hough transform window is slided with a fixed step size in effective detection region, judges the area of hough transform window covering Domain whether there is hand images;
(54) judge whether hough transform window is slided at the ending of present image;If so, then perform step (55);If it is not, Then hough transform window sliding performs step (53) again to next position;
(55) a width of W of present image is setimage, a height of Himage, change of scale is carried out to present image, after carrying out change of scale A width of σ W of imageimage, a height of σ Himage, the image after progress change of scale is set to present image;Wherein, 0 < σ < 1;
(56) judge whether the size of present image is less than minimum threshold;If the size of present image is less than minimum threshold, hold Row step (57);If the size of present image is more than minimum threshold, returns and perform step (51);
(57) number of the candidate target in the list of candidate region and position, the position of hand in comprehensive descision present frame;
(58) judge to accumulate whether frame number reaches unit frame number;If so, then perform step (59);Step is performed if it is not, then returning (51);
(59) in statistical unit frame number, the frame number proportion of hand images be present;And according to the frame number institute accounting of hand images Example judges whether driver is in and takes telephone state.
4. a kind of driver based on svm according to claim 2 takes phone-monitoring method, it is characterised in that:Step (34) in, the suitable effective image subregion of described selection, specifically realized using below equation:
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>l</mi> <mi>e</mi> <mi>f</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>x</mi> <mo>=</mo> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>l</mi> </msub> <mo>.</mo> <mi>x</mi> <mo>-</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>w</mi> <mi>i</mi> <mi>d</mi> <mi>t</mi> <mi>h</mi> <mo>*</mo> <mn>0.6</mn> </mtd> </mtr> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>l</mi> <mi>e</mi> <mi>f</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>y</mi> <mo>=</mo> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>l</mi> </msub> <mo>.</mo> <mi>y</mi> <mo>-</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>h</mi> <mi>e</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> <mo>*</mo> <mn>0.16</mn> </mtd> </mtr> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>l</mi> <mi>e</mi> <mi>f</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>w</mi> <mi>i</mi> <mi>d</mi> <mi>t</mi> <mi>h</mi> <mo>=</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>w</mi> <mi>i</mi> <mi>d</mi> <mi>t</mi> <mi>h</mi> <mo>*</mo> <mn>0.6</mn> </mtd> </mtr> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>l</mi> <mi>e</mi> <mi>f</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>h</mi> <mi>e</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> <mo>=</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>h</mi> <mi>e</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> <mo>*</mo> <mn>1.16</mn> </mtd> </mtr> </mtable> </mfenced>
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>r</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>x</mi> <mo>=</mo> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>r</mi> </msub> <mo>.</mo> <mi>x</mi> </mtd> </mtr> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>r</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>y</mi> <mo>=</mo> <mi>p</mi> <mi>o</mi> <msub> <mi>int</mi> <mi>r</mi> </msub> <mo>.</mo> <mi>y</mi> <mo>-</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>h</mi> <mi>e</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> <mo>*</mo> <mn>0.16</mn> </mtd> </mtr> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>r</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>w</mi> <mi>i</mi> <mi>d</mi> <mi>t</mi> <mi>h</mi> <mo>=</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>w</mi> <mi>i</mi> <mi>d</mi> <mi>t</mi> <mi>h</mi> <mo>*</mo> <mn>0.6</mn> </mtd> </mtr> <mtr> <mtd> <mi>r</mi> <mi>e</mi> <mi>c</mi> <msub> <mi>t</mi> <mrow> <mi>r</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> </mrow> </msub> <mo>.</mo> <mi>h</mi> <mi>e</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> <mo>=</mo> <mi>r</mi> <mi>e</mi> <mi>c</mi> <mi>t</mi> <mo>.</mo> <mi>h</mi> <mi>e</mi> <mi>i</mi> <mi>g</mi> <mi>h</mi> <mi>t</mi> <mo>*</mo> <mn>1.16</mn> </mtd> </mtr> </mtable> </mfenced>
Wherein, rectleft、rectrightSub-rectangular areas near the left and right ear of selection is represented respectively, and rect represents to detect Face location rectangular area, pointl、pointrThe left hand edge point of left eye and the right hand edge point of right eye are represented respectively.
5. a kind of driver based on svm according to claim 3 takes phone-monitoring method, it is characterised in that:Step (51) in, the Gradient Features of described calculating present image, and gradient direction angle is corrected, specifically using below equation Realize:
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mi>M</mi> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> <mo>=</mo> <msqrt> <mrow> <msup> <msub> <mi>G</mi> <mi>x</mi> </msub> <mn>2</mn> </msup> <mo>+</mo> <msup> <msub> <mi>G</mi> <mi>y</mi> </msub> <mn>2</mn> </msup> </mrow> </msqrt> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;theta;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>a</mi> <mi>r</mi> <mi>c</mi> <mi>t</mi> <mi>a</mi> <mi>n</mi> <mrow> <mo>(</mo> <mfrac> <msub> <mi>G</mi> <mi>y</mi> </msub> <msub> <mi>G</mi> <mi>x</mi> </msub> </mfrac> <mo>)</mo> </mrow> </mrow> </mtd> </mtr> </mtable> </mfenced>
<mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <msub> <mi>G</mi> <mi>x</mi> </msub> <mo>=</mo> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>-</mo> <mn>1</mn> <mo>,</mo> <mi>y</mi> <mo>)</mo> <mo>+</mo> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>+</mo> <mn>1</mn> <mo>,</mo> <mi>y</mi> <mo>)</mo> <mo>-</mo> <mn>2</mn> <mo>*</mo> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mtd> </mtr> <mtr> <mtd> <msub> <mi>G</mi> <mi>y</mi> </msub> <mo>=</mo> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>-</mo> <mn>1</mn> <mo>)</mo> <mo>+</mo> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>+</mo> <mn>1</mn> <mo>)</mo> <mo>-</mo> <mn>2</mn> <mo>*</mo> <mi>f</mi> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mtd> </mtr> </mtable> </mfenced>
<mrow> <mi>&amp;theta;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfenced open = "{" close = ""> <mtable> <mtr> <mtd> <mrow> <mi>&amp;theta;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>+</mo> <mi>&amp;pi;</mi> </mrow> </mtd> <mtd> <mrow> <mi>&amp;theta;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&lt;</mo> <mn>0</mn> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mi>&amp;theta;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow> </mtd> <mtd> <mrow> <mi>&amp;theta;</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>&amp;GreaterEqual;</mo> <mn>0</mn> </mrow> </mtd> </mtr> </mtable> </mfenced> </mrow>
Wherein, M (x, y), θ (x, y) represent the gradient magnitude and gradient direction at pixel (x, y) place respectively, and f (x, y) represents pixel The gray value at (x, y) place;Gx represents the partial derivative in pixel (x, y) place x directions, and Gy represents the local derviation in pixel (x, y) place y directions Number.
6. a kind of driver based on svm according to claim 3 takes phone-monitoring method, it is characterised in that:Step (53) it is described to slide hough transform window in effective detection region with a fixed step size in, judge that hough transform window covers The region of lid whether there is hand images;Specifically realized using the step of following order:
(531) histogram of gradients of each subwindow overlay area is counted;
(532) characteristic vector of hough transform window overlay area is obtained;
(533) characteristic vector of hough transform window overlay area is normalized;
(534) characteristic vector of hough transform window overlay area is sent into hand svm graders, predicts its generic;If Hand svm graders predict that this feature vector belongs to hand images, then hough transform window position are added into candidate region In list.
CN201510013139.4A 2015-01-09 2015-01-09 A kind of driver based on svm takes phone-monitoring method Active CN104573659B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510013139.4A CN104573659B (en) 2015-01-09 2015-01-09 A kind of driver based on svm takes phone-monitoring method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510013139.4A CN104573659B (en) 2015-01-09 2015-01-09 A kind of driver based on svm takes phone-monitoring method

Publications (2)

Publication Number Publication Date
CN104573659A CN104573659A (en) 2015-04-29
CN104573659B true CN104573659B (en) 2018-01-09

Family

ID=53089681

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510013139.4A Active CN104573659B (en) 2015-01-09 2015-01-09 A kind of driver based on svm takes phone-monitoring method

Country Status (1)

Country Link
CN (1) CN104573659B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104966059B (en) * 2015-06-15 2018-04-27 安徽创世科技股份有限公司 Driver based on intelligent monitor system drives detection method of making a phone call
CN105469073A (en) * 2015-12-16 2016-04-06 安徽创世科技有限公司 Kinect-based call making and answering monitoring method of driver
CN105868690A (en) * 2016-03-11 2016-08-17 博康智能信息技术有限公司 Method and apparatus for identifying mobile phone use behavior of driver
CN106022242B (en) * 2016-05-13 2019-05-03 哈尔滨工业大学(威海) Method for identifying call receiving and making of driver in intelligent traffic system
CN106530730A (en) * 2016-11-02 2017-03-22 重庆中科云丛科技有限公司 Traffic violation detection method and system
CN108345819B (en) * 2017-01-23 2020-09-15 杭州海康威视数字技术股份有限公司 Method and device for sending alarm message
CN108509902B (en) * 2018-03-30 2020-07-03 湖北文理学院 Method for detecting call behavior of handheld phone in driving process of driver
CN110956060A (en) * 2018-09-27 2020-04-03 北京市商汤科技开发有限公司 Motion recognition method, driving motion analysis method, device and electronic equipment
CN110309764B (en) * 2019-06-27 2021-06-01 浙江工业大学 Multi-stage driver call-making behavior detection method based on deep learning
CN111523380B (en) * 2020-03-11 2023-06-30 浙江工业大学 Mask wearing condition monitoring method based on face and gesture recognition
CN111553217A (en) * 2020-04-20 2020-08-18 哈尔滨工程大学 Driver call monitoring method and system
CN112487990A (en) * 2020-12-02 2021-03-12 重庆邮电大学 DSP-based driver call-making behavior detection method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102567743A (en) * 2011-12-20 2012-07-11 东南大学 Automatic identification method of driver gestures based on video images
CN102592143A (en) * 2012-01-09 2012-07-18 清华大学 Method for detecting phone holding violation of driver in driving
CN103129492A (en) * 2013-03-01 2013-06-05 公安部第三研究所 Vehicular intelligent multimedia terminal device
CN103279750A (en) * 2013-06-14 2013-09-04 清华大学 Detecting method of mobile telephone holding behavior of driver based on skin color range
CN103366506A (en) * 2013-06-27 2013-10-23 北京理工大学 Device and method for automatically monitoring telephone call behavior of driver when driving

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8315617B2 (en) * 2009-10-31 2012-11-20 Btpatent Llc Controlling mobile device functions

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102567743A (en) * 2011-12-20 2012-07-11 东南大学 Automatic identification method of driver gestures based on video images
CN102592143A (en) * 2012-01-09 2012-07-18 清华大学 Method for detecting phone holding violation of driver in driving
CN103129492A (en) * 2013-03-01 2013-06-05 公安部第三研究所 Vehicular intelligent multimedia terminal device
CN103279750A (en) * 2013-06-14 2013-09-04 清华大学 Detecting method of mobile telephone holding behavior of driver based on skin color range
CN103366506A (en) * 2013-06-27 2013-10-23 北京理工大学 Device and method for automatically monitoring telephone call behavior of driver when driving

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"面向智能驾驶行为的机器学习";陈雪梅 等;《道路交通与安全》;20141231;第14卷(第6期);第60-64页 *

Also Published As

Publication number Publication date
CN104573659A (en) 2015-04-29

Similar Documents

Publication Publication Date Title
CN104573659B (en) A kind of driver based on svm takes phone-monitoring method
CN109359536B (en) Passenger behavior monitoring method based on machine vision
CN106780612B (en) Object detecting method and device in a kind of image
CN104637058B (en) A kind of volume of the flow of passengers identify statistical methods based on image information
CN106874894A (en) A kind of human body target detection method based on the full convolutional neural networks in region
CN108388871B (en) Vehicle detection method based on vehicle body regression
CN110298265A (en) Specific objective detection method in a kind of elevator based on YOLO neural network
CN103324937A (en) Method and device for labeling targets
CN103390164A (en) Object detection method based on depth image and implementing device thereof
CN107273852A (en) Escalator floor plates object and passenger behavior detection algorithm based on machine vision
CN108364466A (en) A kind of statistical method of traffic flow based on unmanned plane traffic video
CN103632427B (en) A kind of gate cracking protection method and gate control system
CN110197166A (en) A kind of car body loading condition identification device and method based on image recognition
CN110728225A (en) High-speed face searching method for attendance checking
CN107909027A (en) It is a kind of that there is the quick human body target detection method for blocking processing
CN107491715A (en) A kind of subway carriage passenger flow statistical method, apparatus and system based on video analysis
CN111460924B (en) Gate ticket-evading behavior detection method based on target detection
CN108038424A (en) A kind of vision automated detection method suitable for working at height
CN111292432A (en) Vehicle charging type distinguishing method and device based on vehicle type recognition and wheel axle detection
CN109829382A (en) The abnormal object early warning tracing system and method for Behavior-based control feature intelligent analysis
CN112488042B (en) Pedestrian traffic bottleneck discrimination method and system based on video analysis
CN111523415A (en) Image-based two-passenger one-dangerous vehicle detection method and device
CN109543617A (en) The detection method of intelligent vehicle movement traffic information based on YOLO target detection technique
CN106682600A (en) Method and terminal for detecting targets
CN109598200A (en) A kind of digital image recognition system and method for hot-metal bottle tank number

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant