CN106650661A - Terminal usage state detection method and apparatus - Google Patents
Terminal usage state detection method and apparatus Download PDFInfo
- Publication number
- CN106650661A CN106650661A CN201611189557.XA CN201611189557A CN106650661A CN 106650661 A CN106650661 A CN 106650661A CN 201611189557 A CN201611189557 A CN 201611189557A CN 106650661 A CN106650661 A CN 106650661A
- Authority
- CN
- China
- Prior art keywords
- terminal
- user
- terminal use
- face
- angle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/30—Monitoring
- G06F11/34—Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
- G06F11/3438—Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment monitoring of user actions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/162—Detection; Localisation; Normalisation using pixel segmentation or colour matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/18—Status alarms
- G08B21/24—Reminder alarms, e.g. anti-loss alarms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
Abstract
The invention discloses a terminal usage state detection method and solves the technical problem that the bending degree of the cervical vertebra of a user cannot be accurately calculated in the prior art. The terminal usage state detection method comprises the steps of acquiring state information of a terminal and state information of a terminal user, and calculating a usage state parameter of the terminal user; judging whether the usage state parameter of the terminal user exceeds a threshold or not; and if the usage state parameter of the terminal user exceeds the threshold, sending out a feedback signal by the terminal, wherein the state information of the terminal comprises an included angle between a plane on which the terminal is located and a feature plane, the state information of the terminal user comprises an included angle between a plane on which the face of the terminal user is located and the plane on which the terminal is located, and the usage state parameter of the terminal use refers to a complementary angle of an included angle between the terminal user and the feature plane.
Description
Technical field
The present invention relates to field of terminal technology, more particularly to a kind of detection method and device of using terminal.
Background technology
With the popularization and application of terminal, such as mobile phone, notebook computer, panel computer mobile device can not in life
Or lack, using terminal equipment becomes the custom of people.Terminal unit has brought great convenience, while with terminal work(
Can it is perfect, the time of the daily using terminal of people is increasingly longer.For example, some users see novel or video using mobile phone,
General continuous use time is dozens of minutes, and time length can reach several hours.
At present, because long-time using terminal gives people also to bring worry.When long-time is using mobile phone, head will be from
But it is so more and more lower, it is increasingly closer to mobile phone.The state that long-time cervical vertebra is anteflexion causes compressing to spinal column.Data display,
Head constantly turns forward what the compressing caused to spinal column almost increased in multiple, and head turns forward 15 degree, will give ridge
Post increases by 12.24 kilograms of weight.Seriously may result in cervical curvature straighten, the disease such as prolapse of cervical intervertebral disc.It is such
Injury is difficult to recover completely, and needing drug administration, physiotherapy even to perform the operation could alleviate or treat.
Following problem is primarily present in prior art:
(1) mobile phone and desktop angle are only judged, it is impossible to accurately calculate user's cervical vertebra degree of crook;
(2) when user is in faulty posture using terminal, prompting is not issued the user with;
(3) user's life-time service mobile phone custom cannot be recorded, it is difficult to issue the user with the accurate, prompting of hommization.
The content of the invention
One of purpose of the present invention is the detection method and device for proposing a kind of using terminal, solves prior art
Presence cannot accurately judge the technical problem of user's rachiocamposiss angle when with faulty posture using terminal.What the present invention was provided
Many technique effects that optimal technical scheme in many technical schemes can be generated are (when the bad using terminal of user's certain time length
When, issue the user with prompting etc.) elaboration as detailed below.
For achieving the above object, the invention provides technical scheme below:
A kind of detection method of terminal use state, by obtaining terminal use and terminal relative position relation, judges eventually
End use state, it is characterised in that including step:Acquisition terminal status information and terminal use's status information, meter
Calculate terminal use's use state parameter;Judge whether terminal use's use state parameter exceedes threshold value;And if the end
End subscriber use state parameter exceedes threshold value, and the terminal sends feedback signal;Wherein, the terminal status packet contains
The angle of plane residing for terminal and characteristic face, the residing plane of terminal use's status packet face containing terminal use with
The angle of plane residing for terminal;Terminal use's use state parameter refers to the complementary angle of terminal use and the angle of characteristic face.
Preferably, including step:Acquisition terminal status information and terminal use's status information, computing terminal is used
Family use state parameter;Judge whether terminal use's use state parameter exceedes threshold value;And if terminal use uses shape
When state parameter exceedes threshold value, judge the terminal use and use the terminal, the terminal to send feedback signal with faulty posture;
Wherein, angle t of the terminal status packet containing plane residing for terminal with characteristic face1, shape residing for the terminal use
The angle t of the residing plane of state packet face containing terminal use and plane residing for terminal2;Terminal use's use state parameter
Finger terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
Preferably, the terminal prestores terminal use's standard faces image information.
Preferably, after the terminal prestores terminal use's standard faces image information, the terminal use standard people that prestores is gathered
The distance between the first position parameter of face image information characteristics point, including user's specification human face image information characteristic point.
Preferably, the method for the acquisition terminal user status information includes:Enable the terminal front-facing camera
Terminal use's human face image information in collection use, recognizes described use from terminal use's human face image information in the use
Middle terminal use's face feature point, gathers the second position parameter of terminal use's face feature point in the use, according to described
First position parameter and the second position parameter calculate terminal use's use state parameter.
Preferably, the second position parameter at least includes:Terminal use's face feature point shows with terminal in the use
Horizontal range L between display screen curtain characteristic point1, terminal use's face feature point and terminal display characteristic point in the use
Between vertical dimension L2。
Preferably, it is characterised in that described according to the first position parameter and the second position parameter are calculated
The method of terminal use's use state parameter includes:It is preposition in the terminal according to terminal use's face feature point in the use
Camera imaging relation and second position gain of parameter user's first facial characteristic point and the terminal front-facing camera institute
In the angle t of straight line and terminal place plane3;User's first facial characteristic point is obtained with the preposition shooting of the terminal
Head is apart from MO and the face feature point of the user second with the terminal front-facing camera apart from CO;User's first facial is special
Levy a little and between the face feature point of the user second apart from CM using the terminal use's standard header that prestores as image first is special
Levy a little and between second feature point apart from C ' M ';Wherein, the residing plane of terminal use's face and the face of user first in the use
The angle of portion's characteristic point and the terminal front-facing camera place straight line Calculate institute
State the angle t of the residing plane of terminal use's face and plane residing for terminal in use2=t5-t3, terminal use's use state
Parameter t=90 °-(t1+t2)=90 °-(t1+t5-t3)。
Present invention also offers a kind of detection means of terminal use state, by obtaining terminal use position relative with terminal
Relation is put, terminal use state is judged, it is characterised in that include:Acquisition module, for acquisition terminal status information and end
End subscriber status information;Detection module, for computing terminal user's use state parameter;Judge module, for judging end
Whether end subscriber is with faulty posture using terminal;Wherein, the judge module is specifically for judging that the terminal use uses shape
Whether state parameter exceedes predetermined threshold value;Angle t of the terminal status packet containing plane residing for terminal with characteristic face1,
The angle t of the residing plane of terminal use's status packet face containing terminal use and plane residing for terminal2;The end
End subscriber use state parameter refers to terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
Preferably, also including memory module, for the terminal use's standard faces image information that prestores;Wherein, it is described
Terminal use's standard faces image information specially gathers prestored user standard header as the first position parameter of image characteristic point, bag
Include the distance between user's specification head portrait characteristic point.
Based on above-mentioned technical proposal, the embodiment of the present invention can at least produce following technique effect:Accurately judge that user makes
Degree of crook forward is bent with cervical vertebra during terminal.
Description of the drawings
Accompanying drawing described herein is used for providing a further understanding of the present invention, constitutes the part of the application, this
Bright schematic description and description does not constitute inappropriate limitation of the present invention for explaining the present invention.In the accompanying drawings:
A kind of detection method flow process of terminal use state that Fig. 1 is provided by detection method first embodiment
Figure;
Fig. 2 is illustrated by user's human face image information positioning feature point that detection method first embodiment is provided
Figure;
The position of the terminal gravity sensor that Fig. 3 is provided by detection method first embodiment and function are illustrated
Figure;
The acquisition of the monocular camera machine vision method facial positions that Fig. 4 is provided by detection method first embodiment is former
Reason figure;
The position of the terminal gravity sensor that Fig. 5 is provided by detection method second embodiment and function are illustrated
Figure;
The acquisition of the monocular camera machine vision method facial positions that Fig. 6 is provided by detection method second embodiment is former
Reason figure;
A kind of structure of the detection means of terminal use state that Fig. 7 is provided by detection method 3rd embodiment
Figure.
Specific embodiment
Be referred to accompanying drawing Fig. 1~Fig. 7 below and word content understand present disclosure and the present invention with it is existing
Distinctive points between technology.Hereafter accompanying drawing and enumerate the present invention some alternative embodiments by way of, to the present invention
Technical scheme (including optimal technical scheme) is described in further detail.It should be noted that:Any skill in the present embodiment
Art feature, any technical scheme are one or more in the technical characteristic or optional technical scheme of plurality of optional, in order to
Description is succinct need in presents cannot the exhaustive present invention all alternative technical characteristic and alternative technical side
Case, the embodiment for being also not easy to each technical characteristic emphasizes that it is one of optional numerous embodiments, so this area
Technical staff should know:Any two for arbitrary technological means that the present invention is provided being replaced or providing the present invention
Individual or more technological means or technical characteristic are combined mutually and obtain new technical scheme.Any skill in the present embodiment
Art feature and any technical scheme are not limited the scope of the invention, and protection scope of the present invention should include this area
Technical staff does not pay the thinkable any alternate embodiments of creative work institute and those skilled in the art by the present invention
Any two of offer or more technological means or technical characteristic technical scheme new obtained from being combined mutually.
First embodiment of the invention provides a kind of detection method of the terminal use state when user's handheld terminal.
A kind of detection method of the terminal use state provided the present invention with reference to Fig. 1 to Fig. 4 is carried out specifically
Elaboration, the detection method of the terminal use state comprises the following steps:
Step 101, prestored user standard faces image information.
User is obtained in the standard faces image information with standard gestures using terminal by terminal, terminal storage is stored in
In device.Prestored user standard faces image information function is integrated in terminal system setting, and when being configured by user, terminal is drawn
Leading user carries out shooting acquisition.
After acquiring user's specification human face image information, terminal use's standard faces image information feature point that prestores is obtained
Horizontal range between first position parameter, including user's specification human face image information characteristic point.Obtain the first position ginseng
Several methods include:
First, user's specification head portrait positioning feature point.Eyes, the face of user are for example chosen as user's specification face figure
As information characteristics point.Fig. 2 is user's human face image information positioning feature point schematic diagram.
Eye locating method includes:Thick division is carried out according to the basic law of face distribution.Such as every face sending out from him
Border to eyebrow, eyebrow to nose bottom, nose bottom to chin bottom distance substantially 1/3rd of face.From face
Front sees that the width of face can be approximated to be the width of five eyes, and wherein the other three eye widths are arrived for two ectocanthion
Width between the width of two ears, and two.Eye areas are carried out to need to be accurately positioned after thick division, in such as Fig. 2
E1And E2.Most-often used is Gray Projection method, and this method refers to the product for carrying out to face's gray level image both horizontally and vertically
Point projection, count respectively gray value in both direction and, specific change point is found out, then according to statistical law by difference
Change point position on direction combines, that is, obtain the position C ' at human eye center.
Mouth localization method includes:Because the gray scale of mouth region will be less than its neighboring area, so being accurately positioned face
In the case of portion region, it is possible to use the characteristic difference of mouth and surrounding skin intensity profile is being accurately positioned mouth.Mouth is determined
Position adopts and eye location identical method.Can obtain three under face/part is mouth area-of-interest, to the region
Dynamic threshold segmentation is carried out, vertically and horizontally integral projection is then carried out, that is, obtains mouth center M '.
Then, according to position location, using monocular detection method, passed by the distance in the photographic head and terminal of terminal
Sensor, obtains human eye center C ' and mouth position M ' apart from C ' M '.
Whether step 102, detection terminal is in whether behaviour in service and terminal use are the terminal use standard people that prestores
The user of face image information.
Whether whether detection terminal display screen in dispaly state and in substantially motionless state.If display screen close or
Terminal is kept in motion, and step 102 is re-executed after intervals.If display screen is normally used and kept substantially not
Whether dynamic state, detection terminal user is the user of terminal use's standard faces image information of prestoring.If terminal use is to prestore
The user of terminal use's standard faces image information, continues executing with following each steps.If terminal use is not the terminal use that prestores
The user of standard faces image information, warning terminal user arranges terminal use's standard faces image letter that prestores by terminal system
Breath, after terminal use is by prompting storage standard faces image information, re-executes step 102 after intervals.
Step 103, detection terminal status information.
Terminal status information includes the angle t of plane residing for terminal and characteristic face1, generally using horizontal plane as feature
Face.When user's handheld terminal, t1>0。
The gravity sensor of terminal is called, the angle t of plane residing for terminal and horizontal plane is obtained1。
If Fig. 3 is position and the functional schematic of terminal gravity sensor.The hardware of terminal 1 including at least one towards with
The display screen 3 at family and a gravity sensor 2 that plane residing for terminal and the angle of horizontal plane can be detected.The gravity sensor
2 install on the terminal device, have fixed position relationship with display screen 3.The data exported by gravity sensor 2, can be with
Obtain the angle t of plane residing for terminal and horizontal plane1。
Step 104, detection terminal user's status information.
The angle t of the residing plane of terminal use's status packet face containing terminal use and plane residing for terminal2, also
Including environmental information during user's using terminal, such as in high light or low light environment, uitraviolet intensity, temperature information, air matter
Amount information etc..
When display screen 3 is in dispaly state, opens front-facing camera collection at interval of a period of time (such as 5 minutes) and use
Middle terminal use's human face image information, the prefixed time interval is set by the user.
In collection use after terminal use's human face image information, know from terminal use's human face image information in the use
Not other terminal use's face feature point in use, obtains the second position parameter of terminal use's face feature point in use, and described the
Two location parameters include horizontal range L in use between terminal use's face feature point and terminal display characteristic point1, make
With vertical dimension L between middle terminal use's face feature point and terminal display characteristic point2。
According to first position parameter and second position parameter computing terminal user's use state parameter, the method includes:
First, user's head portrait positioning feature point.Eyes, the face of user are for example chosen as user's head portrait characteristic point.It is special
Levy independent positioning method and obtain the characteristic point positioning method in the first position parameter of user's specification human face image information characteristic point
It is identical.
Then, according to position location, the second position parameter of characteristic point is obtained.
Fig. 4 is the acquisition schematic diagram of monocular camera machine vision method facial positions.Using the monocular detection method, there is one
Individual photographic head, takes mouth position M, the oculocentric position C of people special respectively as user's first facial characteristic point, the face of user second
Levy a little, O points are the front-facing camera center position on terminal front panel.AB places plane is front-facing camera sensor place
Plane, 2 points of A, B is respectively M, C` picture on the image sensor.Because the geometrical relationship of imageing sensor and camera lens is consolidated
It is fixed, therefore angle t can be calculated by A, B position on a sensor and first position parameter3, i.e. the angle of MA and AB.It is logical
The method for crossing monocular camera machine vision, such as geometric similarity method, geometrical optics approach, can obtain the length of CO, MO.And CM can be using pre-
User's specification head portrait C ' M ' values are deposited, so angle t can be calculated with the cosine law again5, computing formula is
.Being derived by simple plane geometry can obtain t2=t5-t3, i.e., the residing plane of terminal use's face in use
With the angle of plane residing for terminal.
Finally, computing terminal user use state parameter, i.e.,
T=90 °-(t1+t2)=90 °-(t1+t5-t3),
T may be used as middle terminal use's cervical vertebra forward leaning angle.
Have preferably experience to make user, terminal also detection terminal user using terminal when monochrome information and/or ultraviolet
Line information and/or temperature information and/or air quality information.Monochrome information passes through light sensor sensors sense environmental luminosity
Value simultaneously calculates terminal use's head portrait image intensity value collection during use.Ultraviolet information, temperature information, air quality information
By networking detection collection.
Whether step 105, judge user with faulty posture using terminal.
Generally, vertebra has about 5 ° of angle of bend.Therefore, can arrange when terminal use's cervical vertebra leans forward in use
Angle judges user with faulty posture using terminal when being more than 5 °.
When terminal judges user is with faulty posture using terminal, the persistent period of faulty posture is recorded.Faulty posture
When persistent period exceeds the scheduled time, display reminding information.
Preferably, when monochrome information, ultraviolet information, temperature information, air quality information one exceed predetermined threshold value,
Display reminding information.When using mobile phone time more than 10 minutes under high light such as user, " you have used mobile phone under high light for prompting
More than 10 minutes, it is unfavorable for your use eye health, it is proposed that take a quick nap or use under ordinary ray intensity mobile phone ".
Terminal can also record the information of terminal use's using terminal, such as user with faulty posture using terminal it is lasting when
Between, user is per twice with the ring when interval time of faulty posture using terminal, user's cervical vertebra forward leaning angle, user's using terminal
Environment information, when at least one terminal use's using terminal information exceedes threshold value, display reminding information.Separately, terminal is used according to terminal
The information of family using terminal is counted, and by statistical result suggestion of the user health using mobile phone is provided.For example, user one day
It is interior with faulty posture using mobile phone total length of time more than 1 hour when, " you use mobile phone today to screen display with faulty posture
Time surpassed 1 hour, bring grievous injury to your vertebra health, it is proposed that adjustment occupation mode ".
Second embodiment of the invention provides a kind of detection of the terminal use state when terminal is placed in desktop by user
Method.
A kind of detection method of the terminal use state provided the present invention with reference to Fig. 1, Fig. 5, Fig. 6 carries out more detailed
Thin elaboration, the detection method of the terminal use state is comprised the following steps:
Step 101, prestored user standard faces image information.
User is obtained in the standard faces image information with standard gestures using terminal by terminal, terminal storage is stored in
In device.Prestored user standard faces image information function is integrated in terminal system setting, and when being configured by user, terminal is drawn
Leading user carries out shooting acquisition.
After acquiring user's specification human face image information, terminal use's standard faces image information feature point that prestores is obtained
Horizontal range between first position parameter, including user's specification human face image information characteristic point.Obtain first position parameter
Method is identical with the method that first position parameter is obtained in first embodiment.
Whether step 102, detection terminal is in whether behaviour in service and terminal use are the terminal use's standard header that prestores
The user of picture.
Whether whether detection terminal display screen in dispaly state and in substantially motionless state.If display screen close or
Terminal is kept in motion, and step 102 is re-executed after intervals.If display screen is normally used and kept substantially not
Whether dynamic state, detection terminal user is the user of terminal use's standard faces image information of prestoring.If terminal use is to prestore
The user of terminal use's standard faces image information, continues executing with following each steps.If terminal use is not the terminal use that prestores
The user of standard faces image information, warning terminal user arranges terminal use's standard faces image letter that prestores by terminal system
Breath, after terminal use is by prompting storage standard faces image information, re-executes step 102 after intervals.
Step 103, detection terminal status information.
Terminal status information includes the angle t of plane residing for terminal and characteristic face1, generally using horizontal plane as feature
Face.When terminal is placed in desktop by user, t1=0.
The gravity sensor of terminal is called, the angle t of plane residing for terminal and horizontal plane is obtained1。
If Fig. 5 is position and the functional schematic of terminal gravity sensor.The hardware of terminal 1 including at least one towards with
The display screen 3 at family and a gravity sensor 2 that plane residing for terminal and the angle of horizontal plane can be detected.The gravity sensor
2 install on the terminal device, have fixed position relationship with display screen 3.The data exported by gravity sensor 2, can be with
Obtaining plane residing for terminal can draw t with the angle of horizontal plane1=0.
Step 104, detection terminal user's status information.
The angle t of the residing plane of terminal use's status packet face containing terminal use and plane residing for terminal2, also
Including environmental information during user's using terminal, such as whether strong light environment, low light environment are in, uitraviolet intensity, temperature information,
Air quality information etc..
When display screen 3 is in dispaly state, front-facing camera is opened at interval of a period of time (such as 5 minutes), this is preset
Time interval is set by the user.
In collection use after terminal use's human face image information, know from terminal use's human face image information in the use
Not other terminal use's face feature point in use, obtains the second position parameter of terminal use's face feature point in use, and described the
Two location parameters include horizontal range L in use between terminal use's face feature point and terminal display characteristic point1‘、
Vertical dimension L in use between terminal use's face feature point and terminal display characteristic point2‘。
According to first position parameter and second position parameter computing terminal user's use state parameter, the method includes:
First, user's head portrait positioning feature point.Eyes, the face of user are for example chosen as user's head portrait characteristic point.It is special
Levy independent positioning method and obtain the characteristic point positioning method in the first position parameter of user's specification human face image information characteristic point
It is identical.Then, according to position location, the second position parameter of characteristic point is obtained.
Fig. 6 is the acquisition schematic diagram of monocular camera machine vision method facial positions.Using the monocular detection method, there is one
Individual photographic head, takes mouth position M, the oculocentric position C of people as user's first facial characteristic point, the facial characteristics of user second
Point, O points are the front-facing camera center position on terminal front panel.A ' B ' places plane is front-facing camera sensor place
Plane, A ', 2 points of B ' are respectively M, C pictures on the image sensor.Because the geometrical relationship of imageing sensor and camera lens is consolidated
It is fixed, therefore angle t can be calculated by A ', B ' position on a sensor and first position parameter3', i.e. the folder of MA ' and A ' B '
Angle.By the method for monocular camera machine vision, such as geometric similarity method, geometrical optics approach can obtain the length of CO, MO.And CM can
Using prestored user standard header as C ' M ' are worth, so angle t can be calculated with the cosine law again5', computing formula is
Being derived by simple plane geometry can obtain t2'=t5‘-t3', i.e., terminal use's face is residing flat in use
Face and the angle of plane residing for terminal.
Finally, computing terminal user use state parameter, because terminal is placed in desktop, then t '=90 °-t2'=90 °-(t5
‘-t3'),
T may be used as middle terminal use's cervical vertebra forward leaning angle.
Have preferably experience to make user, terminal also detection terminal user using terminal when monochrome information and/or ultraviolet
Line information and/or temperature information and/or air quality information.Monochrome information passes through light sensor sensors sense environmental luminosity
Value simultaneously calculates terminal use's head portrait image intensity value collection during use.Ultraviolet information, temperature information, air quality information
By networking detection collection.Whether step 105, judge user with faulty posture using terminal.
Generally, vertebra has about 5 ° of angle of bend.Therefore, can arrange when user's cervical vertebra forward leaning angle in use
During more than 5 ° user is judged with faulty posture using terminal.
When terminal judges user is with faulty posture using terminal, the persistent period of faulty posture is recorded.Faulty posture
When persistent period exceeds the scheduled time, display reminding information.
Preferably, when monochrome information, ultraviolet information, temperature information, air quality information one exceed predetermined threshold value,
Display reminding information.When using mobile phone time more than 10 minutes under high light such as user, " you have used mobile phone under high light for prompting
More than 10 minutes, it is unfavorable for your use eye health, it is proposed that take a quick nap or use under ordinary ray intensity mobile phone ".
Terminal can also record the information of terminal use's using terminal, such as user with faulty posture using terminal it is lasting when
Between, user is per twice with the ring when interval time of faulty posture using terminal, user's cervical vertebra forward leaning angle, user's using terminal
Environment information, when at least one terminal use's using terminal information exceedes threshold value, display reminding information.Separately, terminal is used according to terminal
The information of family using terminal is counted, and by statistical result suggestion of the user health using mobile phone is provided.For example, user one day
It is interior with faulty posture using mobile phone total length of time more than 1 hour when, " you use mobile phone today to screen display with faulty posture
Time surpassed 1 hour, bring grievous injury to your vertebra health, it is proposed that adjustment occupation mode ".
Third embodiment of the invention provides a kind of detection means of terminal use state.
A kind of detection means of the terminal use state provided the present invention with reference to Fig. 7 is explained in more detail
State.Detection means a00 includes acquisition module a10, detection module a20, judge module a30.
Acquisition module a10, for acquisition terminal status information and user's status information;Detection module a20, uses
In calculating user's use state parameter;Judge module a30, for judging user whether with faulty posture using terminal.
The judge module a30 is specifically for judging whether terminal use's use state parameter exceedes predetermined threshold value;
Angle t of the terminal status packet containing plane residing for terminal with characteristic face1, residing for the terminal use
Angle t of the status information comprising the residing plane of terminal use's face with plane residing for terminal2;
Terminal use's use state parameter refers to terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
Detection means a00 also includes memory module a40, for the terminal use's standard faces image information that prestores, in advance
Deposit user's specification head portrait image function to be integrated in terminal system setting, when being configured by user, terminal guiding user carried out
Shoot and obtain;Wherein, terminal use's standard faces image information specially gathers prestored user standard header as characteristics of image
The distance between the first position parameter of point, including user's specification head portrait characteristic point.
After acquiring user's specification head portrait image, user's specification human face image information feature is obtained by acquisition module a10
The distance between the first position parameter of point, including user's specification head portrait characteristic point.First user standard header as characteristic point
In the preservation of one location parameter and memory module a40.Obtaining the method for first position parameter includes:
First, user's specification head portrait positioning feature point.Eyes, the face of user are for example chosen as user's specification face figure
As information characteristics point.Fig. 2 is user's human face image information positioning feature point schematic diagram.
Eye locating method includes:Thick division is carried out according to the basic law of face distribution.Such as every face sending out from him
Border to eyebrow, eyebrow to nose bottom, nose bottom to chin bottom distance substantially 1/3rd of face.From face
Front sees that the width of face can be approximated to be the width of five eyes, and wherein the other three eye widths are arrived for two ectocanthion
Width between the width of two ears, and two.Eye areas are carried out to need to be accurately positioned after thick division, in such as Fig. 2
E1And E2.Most-often used is Gray Projection method, and this method refers to the product for carrying out to face's gray level image both horizontally and vertically
Point projection, count respectively gray value in both direction and, specific change point is found out, then according to statistical law by difference
Change point position on direction combines, that is, obtain the position C ' at human eye center.
Mouth localization method includes:Because the gray scale of mouth region will be less than its neighboring area, so being accurately positioned face
In the case of portion region, it is possible to use the characteristic difference of mouth and surrounding skin intensity profile is being accurately positioned mouth.Mouth is determined
Position adopts and eye location identical method.Can obtain three under face/part is mouth area-of-interest, to the region
Dynamic threshold segmentation is carried out, vertically and horizontally integral projection is then carried out, that is, obtains mouth center M '.
Then, according to position location, using monocular detection method, passed by the distance in the photographic head and terminal of terminal
Sensor, obtains human eye center C ' and mouth position M ' apart from C ' M '.
Detection module a20 is additionally operable in acquisition module a10 acquisition terminal status information and user's status information
Whether front detection terminal is in behaviour in service and terminal status information, and whether terminal use is the terminal use's mark that prestores
The user of accuracy picture.If detect terminal use be not prestore terminal use's standard faces image information user when, point out eventually
End subscriber arranges default terminal use's standard faces image by terminal system.
The gravity sensor of detection module a20 is called, the angle t of plane residing for terminal and horizontal plane is obtained1。
If Fig. 3 is position and the functional schematic of terminal gravity sensor.The hardware of terminal 1 including at least one towards with
The display screen 3 at family and a gravity sensor 2 that plane residing for terminal and the angle of horizontal plane can be detected.The gravity sensor
2 install on the terminal device, have fixed position relationship with display screen 3.The data exported by gravity sensor 2, can be with
Obtain the angle t of plane residing for terminal and horizontal plane1。
Detection module a20 is additionally operable to detect user's status information.
Angle t of user's status packet containing plane residing for user's face with plane residing for terminal2, also including user
Environmental information during using terminal, such as in high light or low light environment, uitraviolet intensity, temperature situation, air quality etc..
When display screen 3 is in dispaly state, opens front-facing camera collection at interval of a period of time (such as 5 minutes) and use
Middle terminal use's human face image information, the prefixed time interval is set by the user.
In acquisition module a10 collection uses after terminal use's human face image information, from terminal use's face in the use
Terminal use's face feature point in use is recognized in image information, the second position of terminal use's face feature point in use is obtained
Parameter, the second position parameter includes the water in use between terminal use's face feature point and terminal display characteristic point
Put down apart from L1, vertical dimension L in use between terminal use's face feature point and terminal display characteristic point2。
According to first position parameter and second position parameter computing terminal user's use state parameter, the method includes:
First, user's head portrait positioning feature point.Eyes, the face of user are for example chosen as user's head portrait characteristic point.It is special
Levy independent positioning method and obtain the characteristic point positioning method in the first position parameter of user's specification human face image information characteristic point
It is identical.
Then, according to position location, the second position parameter of characteristic point is obtained.
Fig. 4 is the acquisition schematic diagram of monocular camera machine vision method facial positions.Using the monocular detection method, there is one
Individual photographic head, takes mouth position M, the oculocentric position C of people special respectively as user's first facial characteristic point, the face of user second
Levy a little, O points are the front-facing camera center position on terminal front panel.AB places plane is front-facing camera sensor place
Plane, 2 points of A, B is respectively M, C` picture on the image sensor.Because the geometrical relationship of imageing sensor and camera lens is consolidated
It is fixed, therefore angle t can be calculated by A, B position on a sensor and first position parameter3, i.e. the angle of MA and AB.It is logical
The method for crossing monocular camera machine vision, such as geometric similarity method, geometrical optics approach, can obtain the length of CO, MO.And CM can be using pre-
User's specification head portrait C ' M ' values are deposited, so angle t can be calculated with the cosine law again5, computing formula is
.Being derived by simple plane geometry can obtain t2=t5-t3, i.e., the residing plane of terminal use's face in use
With the angle of plane residing for terminal.
Finally, computing terminal user use state parameter, i.e.,
T=90 °-(t1+t2)=90 °-(t1+t5-t3),
T may be used as middle terminal use's cervical vertebra forward leaning angle.
Whether judge module a30 is used to judge user with faulty posture using terminal.
Generally, vertebra is about in 5 ° of angle of bend.Therefore, can arrange when user's cervical vertebra forward leaning angle is more than 5 °
When judge user with faulty posture using terminal.
When terminal judges user is with faulty posture using terminal, the persistent period of faulty posture is recorded and by storing mould
Block a40 is stored.When the persistent period of faulty posture exceeds the scheduled time, display reminding information.
Additionally, terminal can also be by user's using terminal information record in memory module a40, such as user is made with faulty posture
With the persistent period of terminal, user per twice with the interval time of faulty posture using terminal, user's cervical vertebra forward leaning angle, user
Environmental information during using terminal, is counted according to these information, is provided user health by statistical result and is used mobile phone
Suggestion.For example, in user one day with faulty posture using the total length of time of mobile phone more than 1 hour when, screen display " your today
1 hour is surpassed using the time of mobile phone with faulty posture, to your vertebra health grievous injury has been brought, it is proposed that adjustment user
Formula ".
The embodiment of the present invention additionally provides a kind of detection methods of terminal use state of A1., by obtain terminal use with
Terminal relative position relation, judges terminal use state, it is characterised in that including step:
Acquisition terminal status information and terminal use's status information, computing terminal user's use state parameter;
Judge whether terminal use's use state parameter exceedes threshold value;And
If terminal use's use state parameter exceedes threshold value, the terminal sends feedback signal;
Wherein, angle of the terminal status packet containing plane residing for terminal with characteristic face, the terminal use
The angle of the residing plane of status packet face containing terminal use and plane residing for terminal;
Terminal use's use state parameter refers to the complementary angle of terminal use and the angle of characteristic face.
A2. the method as described in A1, it is characterised in that including step:
Acquisition terminal status information and terminal use's status information, computing terminal user's use state parameter;
Judge whether terminal use's use state parameter exceedes threshold value;And
If terminal use's use state parameter exceedes threshold value, the terminal use is judged with faulty posture and use the end
End, the terminal sends feedback signal;
Wherein, angle t of the terminal status packet containing plane residing for terminal with characteristic face1, the terminal use
The angle t of the residing plane of family status packet face containing terminal use and plane residing for terminal2;
Terminal use's use state parameter refers to terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
A3. the method as described in A1, it is characterised in that the terminal prestores terminal use's standard faces image letter
Breath.
A4. the method as described in A3, it is characterised in that the terminal prestores after terminal use's standard faces image information,
Gather the first position parameter of terminal use's standard faces image information feature point that prestores, including user's specification human face image information
The distance between characteristic point.
A5. the method as described in A1, it is characterised in that perform " residing for acquisition terminal status information and terminal use
Before status information, computing terminal user's use state parameter ", whether whether detection terminal display screen be in dispaly state and locate
In substantially motionless state.
A6. the method as described in A1, it is characterised in that perform " residing for acquisition terminal status information and terminal use
Before status information, computing terminal user's use state parameter ", detection terminal user be whether prestore terminal use's standard header as
User.
A7. the method as described in A6, it is characterised in that if the terminal use is not the terminal use standard people that prestores
During the user of face image information, point out the terminal use that the terminal use's standard picture that prestores is set by terminal system.
A8. the method as described in A1, it is characterised in that the method for the acquisition terminal user status information includes:
Terminal use's human face image information in the terminal front-facing camera collection use is enabled, from terminal use's face in the use
Terminal use's face feature point in the use is recognized in image information, terminal use's face feature point in the use is gathered
Second position parameter, according to the first position parameter and the second position parameter terminal use's use state ginseng is calculated
Number.
A9. the method as described in A8, it is characterised in that the second position parameter at least includes:Terminal in the use
Horizontal range L between user's face characteristic point and terminal display characteristic point1, terminal use's facial characteristics in the use
Vertical dimension L between point and terminal display characteristic point2。
A10. the method as described in A8 or A9, it is characterised in that described according to the first position parameter and described second
Location parameter calculates the method for terminal use's use state parameter to be included:
According to terminal use's face feature point in the use in the terminal front-facing camera imaging relations and described
It is flat with the terminal front-facing camera place straight line and terminal place that two location parameters obtain user's first facial characteristic point
The angle t in face3;
User's first facial characteristic point is obtained with the terminal front-facing camera apart from MO and the face of the user second
Portion's characteristic point is with the terminal front-facing camera apart from CO;
Being prestored using described apart from CM between user's first facial characteristic point and the face feature point of the user second
Terminal use's standard header as between image fisrt feature point and second feature point apart from C ' M ';
Wherein, the residing plane of terminal use's face is preposition with user's first facial characteristic point and the terminal in the use
The angle of photographic head place straight line
Calculate the angle t of the residing plane of terminal use's face and plane residing for terminal in the use2=t5-t3, the end
End subscriber use state parameter t=90 °-(t1+t2)=90 °-(t1+t5-t3)。
A11. the method as described in A8, it is characterised in that in the front-facing camera collection use for enabling the terminal
Terminal use's human face image information is spaced default time interval.
A12. the method as described in A1, it is characterised in that terminal use's status information also includes the terminal
Environmental information during user's using terminal, at least including monochrome information, ultraviolet information, temperature information, air quality information.
A13. the method as described in A12, it is characterised in that the monochrome information acquisition method includes:Light sensor is passed
Sensor detects ambient light angle value and calculates terminal use's head portrait image intensity value during use.
A14. the method as described in A12, it is characterised in that the ultraviolet information, temperature information, air quality information are adopted
Diversity method includes:Networking detection is obtained.
A15. the method as described in A2, it is characterised in that when the judgement user is with faulty posture using terminal, records institute
State terminal use's using terminal information.
A16. the method as described in A15, it is characterised in that terminal use's using terminal information at least includes:User
With the persistent period of faulty posture using terminal, user is per twice with the interval time of faulty posture using terminal, user's cervical vertebra
Forward leaning angle, environmental information during user's using terminal.
A17. the method as described in A15, it is characterised in that at least one described terminal use's using terminal information exceeds threshold
During value, display reminding information.
A18. the method as described in A17, it is characterised in that the information according to record user's using terminal information,
Count to user and propose healthy using terminal suggestion.
Present invention also offers:B19. a kind of detection means of terminal use state, by obtaining terminal use and terminal
Relative position relation, judges terminal use state, it is characterised in that include:
Acquisition module, for acquisition terminal status information and terminal use's status information;
Detection module, for computing terminal user's use state parameter;
Judge module, for judging terminal use whether with faulty posture using terminal;
Wherein, the judge module is specifically for judging whether terminal use's use state parameter exceedes default threshold
Value;
Angle t of the terminal status packet containing plane residing for terminal with characteristic face1, residing for the terminal use
Angle t of the status information comprising the residing plane of terminal use's face with plane residing for terminal2;
Terminal use's use state parameter refers to terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
B20. the detection means as described in B19, it is characterised in that also including memory module, for prestoring, the terminal is used
Family standard faces image information;
Wherein, terminal use's standard faces image information specially gathers prestored user standard header as image characteristic point
First position parameter, including the distance between user's specification head portrait characteristic point.
B21. the detection means as described in B19, it is characterised in that the detection module is additionally operable to the acquisition module collection
Before terminal status information and terminal use's status information, whether detection terminal display screen is in dispaly state and is
It is no in substantially motionless state.
B22. the detection means as described in B19, it is characterised in that the detection module is additionally operable to the acquisition module collection
Before terminal status information and terminal use's status information, whether detection terminal user is the terminal use's standard header that prestores
The user of picture.
B23. the detection means as described in B22, it is characterised in that if detection terminal user is not the terminal use that prestores
During the user of standard faces image information, point out the terminal use that default terminal use's standard faces are set by terminal system
Image.
B24. the detection means as described in B19, it is characterised in that the method for the acquisition terminal user status information
Including:Terminal use's human face image information in the terminal front-facing camera collection use is enabled, is used from terminal in the use
Terminal use's face feature point in the use is recognized in the human face image information of family, terminal use's face in the use is gathered special
Second position parameter a little is levied, the terminal use is calculated according to the first position parameter and the second position parameter and is used
State parameter.
B25. the detection means as described in B24, it is characterised in that the second position parameter at least includes:It is described to use
Horizontal range L between middle terminal use's face feature point and terminal display characteristic point1, terminal use face in the use
Vertical dimension L between portion's characteristic point and terminal display characteristic point2。
B26. the detection means as described in B24 or B25, it is characterised in that described according to the first position parameter and institute
Stating the method for second position parameter calculating terminal use's use state parameter includes:
According to terminal use's face feature point in the use in the terminal front-facing camera imaging relations and described
It is flat with the terminal front-facing camera place straight line and terminal place that two location parameters obtain first user face feature point
The angle t in face3;
The first user face feature point is obtained with the terminal front-facing camera apart from MO and the second user face
Portion's characteristic point is with the terminal front-facing camera apart from CO;
Being prestored using described apart from MC between the first user face feature point and the second user face feature point
Terminal use's standard header as between image fisrt feature point and second feature point apart from M ' C ';
Wherein, the residing plane of terminal use's face and first user face feature point and the terminal front-facing camera
The angle of place straight line
Calculate the angle t of the residing plane of terminal use's face and plane residing for terminal2=t5-t3, the terminal use
Use state parameter t=90 °-(t1+t2)=90 °-(t1+t5-t3)。
B27. the detection means as described in B19, it is characterised in that the acquisition module is additionally operable to gather the terminal use
Environmental information during using terminal, at least including monochrome information, ultraviolet information, temperature information, air quality information.
B28. the detection means as described in B7, it is characterised in that the monochrome information acquisition method includes:Light sensor
Sensors sense environmental shading value simultaneously calculates terminal use's head portrait image intensity value during use.
B29. the detection means as described in B27, it is characterised in that the ultraviolet information, temperature information, air quality letter
Breath acquisition method includes:Networking detection is obtained.
B30. the detection means as described in B19, it is characterised in that the memory module is additionally operable to, record user is with bad
Terminal use's using terminal information of posture using terminal.
B31. the detection means as described in B30, it is characterised in that terminal use's using terminal information at least includes:
, with the persistent period of faulty posture using terminal, user is per twice with the interval time of faulty posture using terminal, user for user
Cervical vertebra forward leaning angle, environmental information during user's using terminal.Arbitrary technical scheme disclosed in the invention described above is except as otherwise sound
It is bright outer, if it discloses numerical range, then disclosed numerical range is preferred numerical range, the skill of any this area
Art personnel should be appreciated that:Preferred numerical range be only in many enforceable numerical value technique effect it is obvious or with generation
The numerical value of table.Due to numerical value it is more, it is impossible to it is exhaustive, thus the present invention just open component values illustrating skill of the invention
Art scheme, also, the above-mentioned numerical value enumerated should not constitute the restriction to the invention protection domain.
Finally it should be noted that:Above example is only to illustrate technical scheme rather than a limitation;To the greatest extent
Pipe has been described in detail with reference to preferred embodiment to the present invention, and those of ordinary skill in the art should be understood:Still
The specific embodiment of the present invention can be modified or equivalent is carried out to some technical characteristics;Without deviating from this
The spirit of bright technical scheme, it all should cover in the middle of the technical scheme scope being claimed in the present invention.
Claims (10)
1. a kind of detection method of terminal use state, by obtaining terminal use and terminal relative position relation, judges terminal
Use state, it is characterised in that including step:
Acquisition terminal status information and terminal use's status information, computing terminal user's use state parameter;
Judge whether terminal use's use state parameter exceedes threshold value;And
If terminal use's use state parameter exceedes threshold value, the terminal sends feedback signal;
Wherein, angle of the terminal status packet containing plane residing for terminal and characteristic face, residing for the terminal use
Angle of the status information comprising the residing plane of terminal use's face with plane residing for terminal;
Terminal use's use state parameter refers to the complementary angle of terminal use and the angle of characteristic face.
2. method according to claim 1, it is characterised in that including step:
Acquisition terminal status information and terminal use's status information, computing terminal user's use state parameter;
Judge whether terminal use's use state parameter exceedes threshold value;And
If terminal use's use state parameter exceedes threshold value, the terminal use is judged with faulty posture and uses the terminal,
The terminal sends feedback signal;
Wherein, angle t of the terminal status packet containing plane residing for terminal with characteristic face1, residing for the terminal use
Angle t of the status information comprising the residing plane of terminal use's face with plane residing for terminal2;
Terminal use's use state parameter refers to terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
3. method according to claim 1, it is characterised in that the terminal prestores terminal use's standard faces image
Information.
4. method according to claim 2, it is characterised in that the terminal prestores terminal use's standard faces image information
Afterwards, the first position parameter of terminal use's standard faces image information feature point that prestores, including user's specification facial image are gathered
The distance between information characteristics point.
5. method according to claim 1, it is characterised in that the method bag of the acquisition terminal user status information
Include:Terminal use's human face image information in the terminal front-facing camera collection use is enabled, from terminal use in the use
Terminal use's face feature point in the use is recognized in human face image information, terminal use's facial characteristics in the use are gathered
The second position parameter of point, calculates the terminal use and uses shape according to the first position parameter and the second position parameter
State parameter.
6. method according to claim 5, it is characterised in that the second position parameter at least includes:In the use
Horizontal range L between terminal use's face feature point and terminal display characteristic point1, terminal use's face in the use
Vertical dimension L between characteristic point and terminal display characteristic point2。
7. the method according to claim 5 or 6, it is characterised in that described according to the first position parameter and described
Two location parameters calculate the method for terminal use's use state parameter to be included:
According to terminal use's face feature point in the use in the terminal front-facing camera imaging relations and the second
Gain of parameter user's first facial characteristic point is put with the terminal front-facing camera place straight line and terminal place plane
Angle t3;
Obtain user's first facial characteristic point special apart from MO and the face of the user second with the terminal front-facing camera
Levy a little with the terminal front-facing camera apart from CO;
Between user's first facial characteristic point and the face feature point of the user second apart from CM using the terminal that prestores
Between user's specification head portrait image fisrt feature point and second feature point apart from C ' M ';
Wherein, the residing plane of terminal use's face and user's first facial characteristic point and the preposition shooting of the terminal in the use
The angle of head place straight line
Calculate the angle t of the residing plane of terminal use's face and plane residing for terminal in the use2=t5-t3, the terminal use
Family use state parameter t=90 °-(t1+t2)=90 °-(t1+t5-t3)。
8. a kind of detection means of terminal use state, by obtaining terminal use and terminal relative position relation, judges terminal
Use state, it is characterised in that include:
Acquisition module, for acquisition terminal status information and terminal use's status information;
Detection module, for computing terminal user's use state parameter;
Judge module, for judging terminal use whether with faulty posture using terminal;
Wherein, the judge module is specifically for judging whether terminal use's use state parameter exceedes predetermined threshold value;
Angle t of the terminal status packet containing plane residing for terminal with characteristic face1, terminal use's status
The angle t of the residing plane of packet face containing terminal use and plane residing for terminal2;
Terminal use's use state parameter refers to terminal use cervical vertebra forward leaning angle t, wherein t=90 °-(t1+t2)。
9. detection means according to claim 8, it is characterised in that also including memory module, for the terminal that prestores
User's specification human face image information;
Wherein, terminal use's standard faces image information specially gathers prestored user standard header as the of image characteristic point
The distance between one location parameter, including user's specification head portrait characteristic point.
10. detection means according to claim 8, it is characterised in that the acquisition terminal user status information
Method includes:Terminal use's human face image information in the terminal front-facing camera collection use is enabled, from the use eventually
Terminal use's face feature point in the use is recognized in end subscriber human face image information, terminal use face in the use is gathered
The second position parameter of portion's characteristic point, according to the first position parameter and the second position parameter terminal use is calculated
Use state parameter.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611189557.XA CN106650661A (en) | 2016-12-21 | 2016-12-21 | Terminal usage state detection method and apparatus |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611189557.XA CN106650661A (en) | 2016-12-21 | 2016-12-21 | Terminal usage state detection method and apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106650661A true CN106650661A (en) | 2017-05-10 |
Family
ID=58834300
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201611189557.XA Pending CN106650661A (en) | 2016-12-21 | 2016-12-21 | Terminal usage state detection method and apparatus |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106650661A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108986421A (en) * | 2018-07-23 | 2018-12-11 | 广州维纳斯家居股份有限公司 | Intelligent elevated table infrared reminding method, apparatus, intelligent elevated table and storage medium |
CN109120791A (en) * | 2018-08-31 | 2019-01-01 | 湖南人文科技学院 | A kind of method that smart phone carries out warning protection cervical vertebra |
CN110059670A (en) * | 2019-04-29 | 2019-07-26 | 杭州雅智医疗技术有限公司 | Human body Head And Face, limb activity angle and body appearance non-contact measurement method and equipment |
CN110236551A (en) * | 2019-05-21 | 2019-09-17 | 西藏纳旺网络技术有限公司 | Acquisition methods, device, electronic equipment and the medium of user's cervical vertebra tilt angle |
US20220319045A1 (en) * | 2021-04-01 | 2022-10-06 | MohammadSado Lulu | System For Posture Detection Using The Camera Of A Hand-Held Device |
CN115291678A (en) * | 2022-07-25 | 2022-11-04 | 深圳市三森科技有限公司 | Electronic equipment, and identification method and system for looking up at electronic equipment |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2010089989A1 (en) * | 2009-02-05 | 2012-08-09 | パナソニック株式会社 | Information display device and information display method |
CN103927012A (en) * | 2014-04-16 | 2014-07-16 | 北京尚德智产投资管理有限公司 | Mobile terminal achieving user posture detection through operating system and method |
CN103927250A (en) * | 2014-04-16 | 2014-07-16 | 北京尚德智产投资管理有限公司 | User posture detecting method achieved through terminal device |
CN103955273A (en) * | 2014-04-16 | 2014-07-30 | 北京尚德智产投资管理有限公司 | Mobile terminal and method for realizing user posture detection through operation system |
-
2016
- 2016-12-21 CN CN201611189557.XA patent/CN106650661A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2010089989A1 (en) * | 2009-02-05 | 2012-08-09 | パナソニック株式会社 | Information display device and information display method |
CN103927012A (en) * | 2014-04-16 | 2014-07-16 | 北京尚德智产投资管理有限公司 | Mobile terminal achieving user posture detection through operating system and method |
CN103927250A (en) * | 2014-04-16 | 2014-07-16 | 北京尚德智产投资管理有限公司 | User posture detecting method achieved through terminal device |
CN103955273A (en) * | 2014-04-16 | 2014-07-30 | 北京尚德智产投资管理有限公司 | Mobile terminal and method for realizing user posture detection through operation system |
Non-Patent Citations (1)
Title |
---|
林强等: "《行为识别与智能计算》", 30 November 2016, 西安电子科技大学出版社 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108986421A (en) * | 2018-07-23 | 2018-12-11 | 广州维纳斯家居股份有限公司 | Intelligent elevated table infrared reminding method, apparatus, intelligent elevated table and storage medium |
CN109120791A (en) * | 2018-08-31 | 2019-01-01 | 湖南人文科技学院 | A kind of method that smart phone carries out warning protection cervical vertebra |
CN109120791B (en) * | 2018-08-31 | 2021-01-01 | 湖南人文科技学院 | Method for warning and protecting cervical vertebra through smart phone |
CN110059670A (en) * | 2019-04-29 | 2019-07-26 | 杭州雅智医疗技术有限公司 | Human body Head And Face, limb activity angle and body appearance non-contact measurement method and equipment |
CN110059670B (en) * | 2019-04-29 | 2024-03-26 | 杭州雅智医疗技术有限公司 | Non-contact measuring method and equipment for head and face, limb movement angle and body posture of human body |
CN110236551A (en) * | 2019-05-21 | 2019-09-17 | 西藏纳旺网络技术有限公司 | Acquisition methods, device, electronic equipment and the medium of user's cervical vertebra tilt angle |
US20220319045A1 (en) * | 2021-04-01 | 2022-10-06 | MohammadSado Lulu | System For Posture Detection Using The Camera Of A Hand-Held Device |
CN115291678A (en) * | 2022-07-25 | 2022-11-04 | 深圳市三森科技有限公司 | Electronic equipment, and identification method and system for looking up at electronic equipment |
CN115291678B (en) * | 2022-07-25 | 2024-04-30 | 深圳市三森科技有限公司 | Electronic equipment, and identification method and system for looking up electronic equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106650661A (en) | Terminal usage state detection method and apparatus | |
CN109086726B (en) | Local image identification method and system based on AR intelligent glasses | |
US20210223862A1 (en) | Interactive Motion-Based Eye Tracking Calibration | |
US9254081B2 (en) | Fitting glasses frames to a user | |
CN105393192B (en) | Netted hierarchical menu for near-eye display shows configuration | |
CA3003550C (en) | Real-time visual feedback for user positioning with respect to a camera and a display | |
US9141761B2 (en) | Apparatus and method for assisting user to maintain correct posture | |
US9842246B2 (en) | Fitting glasses frames to a user | |
JP2023113596A (en) | Improved optical and sensory digital eyewear | |
KR102190812B1 (en) | Method for determining at least one value of a parameter for customising a visual compensation device | |
US7742624B2 (en) | Perspective improvement for image and video applications | |
US20130215293A1 (en) | Apparatus and method for generating image including multiple people | |
CN109782913A (en) | A kind of method and device that control screen content is shown | |
CN115599219A (en) | Eye protection control method, system, equipment and storage medium for display screen | |
CN102043942A (en) | Visual direction judging method, image processing method, image processing device and display device | |
JP2012022538A (en) | Attention position estimating method, image display method, attention content display method, attention position estimating device and image display device | |
CN109688325A (en) | A kind of image display method and terminal device | |
CN109640787A (en) | Measure the System and method for of interpupillary distance | |
WO2018219290A1 (en) | Information terminal | |
CN107495922A (en) | Method for detecting eye fatigue | |
CN107329570A (en) | A kind of method that display device angle is automatically adjusted according to individual beholder position | |
JP5242827B2 (en) | Face image processing apparatus, face image processing method, electronic still camera, digital image processing apparatus, and digital image processing method | |
EP4366600A2 (en) | Systems and methods for vision test and uses thereof | |
CN113032022B (en) | Equipment control method and related equipment | |
CN109977836A (en) | A kind of information collecting method and terminal |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170510 |