CN107272960A - A kind of display methods and display system - Google Patents
A kind of display methods and display system Download PDFInfo
- Publication number
- CN107272960A CN107272960A CN201710517327.XA CN201710517327A CN107272960A CN 107272960 A CN107272960 A CN 107272960A CN 201710517327 A CN201710517327 A CN 201710517327A CN 107272960 A CN107272960 A CN 107272960A
- Authority
- CN
- China
- Prior art keywords
- angle
- current
- data
- user
- sight
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0412—Digitisers structurally integrated in a display
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72448—User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Position Input By Displaying (AREA)
Abstract
The present embodiments relate to display technology field, more particularly to a kind of display methods and display system, one of which display methods, including:Judge whether current gaze angle matches pre- controlling angle;Judge whether current gaze angle matches pre- controlling angle;In the state of sight angle mismatches pre- controlling angle, the TP coordinates under adjustment current display state, so that TP coordinate matchings and current gaze drop point.Compared with prior art, the beneficial effects of the invention are as follows:TP coordinates are adjusted, so that sight drop point of the TP coordinates perpendicular to user, when user applies pressure to sight drop point region, the pressure puts on TP coordinates surface, the control that TP coordinate actual match user expects is to obtain the operation of user's expectation.
Description
Technical field
The present embodiments relate to display technology field, more particularly to a kind of display methods and display system.
Background technology
Touch-screen is also known as " touch screen ", " contact panel ", is a kind of induction type liquid for receiving the input signals such as contact
Crystal device, when contacting the graphic button on screen, haptic feedback system on screen can be according to the journey of preprogramming
Formula drives various link devices, is substituted for mechanical push button panel, and produces by liquid crystal display picture lively
Visual and sound effects.Touch-screen is as a kind of newest calculating input equipment, and it is most simple at present, conveniently, naturally one kind is man-machine
Interactive mode.
The composition structure of existing touch-screen is as shown in figure 1, at least include the control command to receive outside input
Touch control layer 1, the luminescent layer 3 to connect the transparent articulamentum 2 of touch control layer, show control, touch control layer 1 are formed by glass material, when
When viewpoint O observes luminescent layer 3 with non-vertical direction, 1 refractive index and the refractive index of air of the touch control layer of glass material formation
Difference, because refractive index is different, when the luminous point that viewpoint O is observed is in A ", control is physically located in A points, the drawbacks of now occurring
It is exactly:When user, which is intended to touch-control, is located at A points, because the position of the actual touch-control of error on line of sight user is located at A ", mobile terminal or display
Terminal can not obtain accurate touch command and then form bad command.Especially in the case where control is arranged with packed form,
The control of such as A points is C1, and the control of A " points is C2;Touch-control C1 controls are intended to when user is actual, when the thickness of touch control layer 1 is about thick,
Or articulamentum 2 thickness it is rear in the state of, such a error is about obvious.
The drawback is present not only in the application of mobile terminal, at the same occur the indoor and outdoor that occurs can touch-control interaction screen
On curtain, when user does not stand on optimal viewing point, the command signal and the command signal of user's expectation of the touch-control generation of user
Difference, substantially reduces the experience sense of user.
The content of the invention
Based on drawbacks described above, the present invention provides a kind of display methods and display system, it is intended to be in non-optimal position in sight
Under configuration state, the dispaly state matching viewpoint of adjustment display unit, specifically:
On the one hand, the present invention provides a kind of display methods, wherein:
Judge whether current gaze angle matches pre- controlling angle;
In the state of the sight angle mismatches pre- controlling angle, the TP coordinates under adjustment current display state, so that
TP coordinate matchings and current gaze drop point.
Preferably, above-mentioned display methods, wherein;Judge current gaze angle is also wrapped before whether matching pre- controlling angle
Include:
The current attitude data of operator is obtained, presently described sight angle is formed according to the attitude data.
Preferably, above-mentioned display methods, wherein;The current attitude data of operator is obtained to specifically include:
Obtain the face-image of current operator;According to the face-image obtain characteristic fisrt feature data,
Second feature data;
Positional information of the mobile terminal in current physical space is obtained, centre coordinate is formed according to the positional information
Data;
The sight angle is formed according to the centre coordinate data, fisrt feature data, second feature data.
Preferably, above-mentioned display methods, wherein;The current attitude data of operator is obtained to specifically include:
The pose presentation of current operator is obtained, the head position data of user are judged according to the pose presentation;
Presently described sight angle is formed according to the head position data of user.
Preferably, above-mentioned display methods, wherein;Also include,
In the state of the TP coordinate matchings and current gaze drop point, receive the operational order of user's input, and according to
The operational order formation operational order output.
On the other hand, the present invention provides a kind of display system again, wherein:
Judging unit, to judge whether current gaze angle matches pre- controlling angle;
Adjustment unit, in the state of mismatching pre- controlling angle in the sight angle, under adjustment current display state
TP coordinates so that TP coordinate matchings and current gaze drop point.
Preferably, above-mentioned display system, wherein;Also include:
First collecting unit, obtains the current attitude data of operator, forms presently described according to the attitude data
Sight angle.
Preferably, above-mentioned display system, wherein, the first collecting unit is specifically included:
First collector, obtains the face-image of current operator;The of characteristic is obtained according to the face-image
One characteristic, second feature data;
Second collector, obtains positional information of the mobile terminal in current physical space, according to the positional information
Form centre coordinate data;
First calculator, the sight is formed according to the centre coordinate data, fisrt feature data, second feature data
Angle.
Preferably, above-mentioned display system, wherein, first collecting unit is specifically included:
3rd collector, obtains the pose presentation of current operator, and the head position of user is judged according to the pose presentation
Put data;
Second calculator, forms presently described sight angle according to the head position data of user.
Preferably, above-mentioned display system, wherein, in addition to,
Control unit, in the state of the TP coordinate matchings and current gaze drop point, to receive the behaviour of user's input
Order, and operational order output is formed according to the operational order.
Compared with prior art, the beneficial effects of the invention are as follows:
Brief description of the drawings
A kind of structural representation for touch screen that Fig. 1 provides for embodiment in the present invention;
Fig. 2 is a kind of flow chart of display methods of embodiment offer in the present invention;
Fig. 3 is a kind of Image Adjusting structure chart of display methods of embodiment offer in the present invention;
Fig. 4 is a kind of flow chart of display methods of embodiment offer in the present invention;
Fig. 5 is a kind of flow chart of display methods of embodiment offer in the present invention.
Embodiment
The present invention is described in further detail with reference to the accompanying drawings and examples.It is understood that this place is retouched
The specific embodiment stated is used only for explaining the present invention, rather than limitation of the invention.It also should be noted that, in order to just
Part related to the present invention rather than entire infrastructure are illustrate only in description, accompanying drawing.
As shown in Figure 2, on the one hand, the present invention provides a kind of display methods, wherein:Including,
Step S110, judge whether current gaze angle theta matches pre- controlling angle;Sight angle is viewpoint to display screen surface
Angle, the scope of the angle is 0 ° -90 °;Further, it is preferable to ground, pre- controlling angle is 90 °, when sight angle is 90 °,
Judge that current gaze angle matches pre- controlling angle.
It should be noted that in the state of less than 10 inches, its pre- controlling angle can be 75 ° -90 °, less than 10 inches
Screen, the size of its screen is relatively small, so the thickness of articulamentum is relatively low, (screen size is smaller, touch control layer and hair
Attachment force between photosphere is relatively small), so refraction of the articulamentum to light is relatively small, then the body based on control in itself
Product, when sight angle is 75 °, the surface of user's touch-control matches the actual position of control substantially.
Step S120, in the state of the sight angle mismatches pre- controlling angle, the TP under adjustment current display state
Coordinate, so that TP coordinate matchings and current gaze drop point.
In the state of the sight angle mismatches pre- controlling angle, sight drop point is with TP coordinates not in same Vertical Square
Xiang Shi, in the state of user presses to sight drop point, it can not obtain control effect expected from user, based on this, adjust TP
Coordinate, so that sight drop point of the TP coordinates perpendicular to user, when user applies pressure to sight drop point region, the pressure
Put on TP coordinates surface, TP coordinate actual match user expect control with obtain user expectation operating efficiency.
For example, when user is from right side, i.e., sight drop point is located at the left side of actual control, then control is corresponding
TP coordinates are adjusted to the left, so that TP coordinates are arranged at directly over sight drop point.With should user from left sides when, by TP coordinates
Turn right adjustment.
With shown in Fig. 3, on the relatively thin display screen of hyaline layer, by taking viewpoint O as an example, sight angle is θ, then adjusts
Distance is:
Z=H*tan (90- θ)-H*tan { arcsin [(n2/n1) * sin (90- θ)] };
Wherein, Z:For the adjustment distance of TP coordinates;
θ:For the sight angle;
H:For the distance of luminescent layer to touch-control layer surface;
n1:For the refractive index of air;
n2:For the refractive index of touch control layer.
Step S130, in the state of the TP coordinate matchings and current gaze drop point, receive user input operation life
Order, and operational order output is formed according to the operational order.
The present invention operation principle be:According to the current TP coordinates of the position adjustment of sight drop point, so that TP coordinate matchings
With current gaze drop point.When i.e. user vertically presses, touch-control presses on the control expected to user, to improve the body of user
Test sense.
As further preferred embodiment, above-mentioned display methods, wherein;Judge before step S110 is performed current
Sight angle also includes before whether matching pre- controlling angle:
Step S109, obtains the current attitude data of operator, presently described sight is formed according to the attitude data
Angle.The current state of user is obtained by front camera, current state at least includes face-image, pre- according to current state
Estimate to form attitude data.
As shown in figure 4, specifically including:
Step S1091, obtains the face-image of current operator;The first of characteristic is obtained according to the face-image
Characteristic, second feature data;Wherein fisrt feature data can be the eyeball coordinate data of operator, and second feature data can
For ratio data, interpupillary distance that for example can be in face-image relatively obtains the ratio data with actual interpupillary distance.
Step S1092, obtains positional information of the mobile terminal in current physical space, according to the positional information shape
Into centre coordinate data;Positional information in current physical space can have what sensor was obtained, be not particularly limited herein.
Centre coordinate data can be the data of display unit center reference point, for example, when the state that display unit is rectangle
Under, center reference point is diagonal intersection point.
Step S1093, the sight is formed according to the centre coordinate data, fisrt feature data, second feature data
Angle.
Specifically, it can calculate at this 3 points in same according to fisrt feature data, second feature data, centre coordinate data
Sight angle of the operator to center coordinate points in plane.
Embodiment two
Embodiment one provides one kind and is based on eye feature data acquisition sight angle, and adjusts display according to sight angle
Image, so that the control in display image matches the sight drop point of user, it is desirable to provide the experience sense of user.But such a mode
It is relatively large in the presence of a weak point, i.e. data operation quantity, for the relatively small terminal device of calculation processing power, Ci Zhongfang
The when ductility of formula is relatively long, and based on this, the present invention provides a kind of display methods again, it is intended to reduce operand, reduces time delay.
The present invention provides a kind of display methods, wherein, including:As shown in figure 5,
Step S210, the pose presentation for obtaining current operator, the head position of user is judged according to the pose presentation
Data;The pose presentation of current operator is obtained by front camera.
Step S220, presently described sight angle is formed according to the head position data of user.
Step S230, judge whether current gaze angle theta matches pre- controlling angle.
Step S240, in the state of the sight angle mismatches pre- controlling angle, the TP under adjustment current display state
Coordinate, so that TP coordinate matchings and current gaze drop point.
Step S250, in the state of the TP coordinate matchings and current gaze drop point, receive user input operation life
Order, and operational order output is formed according to the operational order.
The operation principle of the present embodiment is:
Sight angle is estimated according to the facial deflection angle of operator, image is adjusted according to sight angle, so that the TP
In the state of coordinate matching and current gaze drop point.
Embodiment three
On the other hand, the present invention provides a kind of display system again, wherein:
Judging unit, to judge whether current gaze angle matches pre- controlling angle;
Adjustment unit, in the state of mismatching pre- controlling angle in the sight angle, under adjustment current display state
TP coordinates so that TP coordinate matchings and current gaze drop point.
As further preferred embodiment, above-mentioned display system, wherein;Also include:
First collecting unit, obtains the current attitude data of operator, forms presently described according to the attitude data
Sight angle.
As further preferred embodiment, above-mentioned display system, wherein, the first collecting unit is specifically included:
First collector, obtains the face-image of current operator;The of characteristic is obtained according to the face-image
One characteristic, second feature data;
Second collector, obtains positional information of the mobile terminal in current physical space, according to the positional information
Form centre coordinate data;
First calculator, the sight is formed according to the centre coordinate data, fisrt feature data, second feature data
Angle.
As further preferred embodiment, above-mentioned display system, wherein, first collecting unit is specifically included:
3rd collector, obtains the pose presentation of current operator, and the head position of user is judged according to the pose presentation
Put data;
Second calculator, forms presently described sight angle according to the head position data of user.
As further preferred embodiment, above-mentioned display system, wherein, in addition to,
Control unit, in the state of the TP coordinate matchings and current gaze drop point, to receive the behaviour of user's input
Order, and operational order output is formed according to the operational order.
The above method can perform the product that any embodiment of the present invention is provided, and possess the execution corresponding functional module of product
And beneficial effect.Therefore the operation principle to temprature control method is repeated no more herein.
Although various aspects of the invention are provided in the independent claim, the other side of the present invention includes coming from
The combination of the dependent claims of the feature of described embodiment and/or feature with independent claims, and not only
It is the combination clearly provided in claim.
Although it is to be noted here that the foregoing describing the example embodiment of the present invention, these descriptions are not
It should be understood in a limiting sense.Will without departing from such as appended right on the contrary, several change and modifications can be carried out
The scope of the present invention defined in asking.
Note, above are only presently preferred embodiments of the present invention and institute's application technology principle.It will be appreciated by those skilled in the art that
The invention is not restricted to specific embodiment described here, can carry out for a person skilled in the art it is various it is obvious change,
Readjust and substitute without departing from protection scope of the present invention.Therefore, although the present invention is carried out by above example
It is described in further detail, but the present invention is not limited only to above example, without departing from the inventive concept, also
Other more equivalent embodiments can be included, and the scope of the present invention is determined by scope of the appended claims.
Claims (10)
1. a kind of display methods, it is characterised in that:
Judge whether current gaze angle matches pre- controlling angle;
In the state of the sight angle mismatches pre- controlling angle, the TP coordinates under adjustment current display state, so that TP is sat
Mark matching and current gaze drop point.
2. display methods according to claim 1, it is characterised in that;Judge whether current gaze angle matches pre- controlling angle
Also include before:
The current attitude data of operator is obtained, presently described sight angle is formed according to the attitude data.
3. display methods according to claim 2, it is characterised in that obtain the current attitude data of operator and specifically wrap
Include:
Obtain the face-image of current operator;According to the fisrt feature data of face-image acquisition characteristic, second
Characteristic;
Positional information of the mobile terminal in current physical space is obtained, centre coordinate number is formed according to the positional information
According to;
The sight angle is formed according to the centre coordinate data, fisrt feature data, second feature data.
4. display methods according to claim 2, it is characterised in that obtain the current attitude data of operator and specifically wrap
Include:
The pose presentation of current operator is obtained, the head position data of user are judged according to the pose presentation;
Presently described sight angle is formed according to the head position data of user.
5. display methods according to claim 1, it is characterised in that:Also include,
In the state of the TP coordinate matchings and current gaze drop point, receive the operational order of user's input, and according to described
Operational order formation operational order output.
6. a kind of display system, it is characterised in that:
Judging unit, to judge whether current gaze angle matches pre- controlling angle;
Adjustment unit, in the state of mismatching pre- controlling angle in the sight angle, the TP under adjustment current display state
Coordinate, so that TP coordinate matchings and current gaze drop point.
7. display system according to claim 6, it is characterised in that;Also include:
First collecting unit, obtains the current attitude data of operator, presently described sight is formed according to the attitude data
Angle.
8. display system according to claim 7, it is characterised in that the first collecting unit is specifically included:
First collector, obtains the face-image of current operator;The first spy of characteristic is obtained according to the face-image
Levy data, second feature data;
Second collector, obtains positional information of the mobile terminal in current physical space, is formed according to the positional information
Centre coordinate data;
First calculator, forms the sight according to the centre coordinate data, fisrt feature data, second feature data and presss from both sides
Angle.
9. display system according to claim 7, it is characterised in that first collecting unit is specifically included:
3rd collector, obtains the pose presentation of current operator, the head position number of user is judged according to the pose presentation
According to;
Second calculator, forms presently described sight angle according to the head position data of user.
10. display system according to claim 6, it is characterised in that:Also include,
Control unit, the operation life in the state of the TP coordinate matchings and current gaze drop point, to receive user's input
Order, and operational order output is formed according to the operational order.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710517327.XA CN107272960A (en) | 2017-06-29 | 2017-06-29 | A kind of display methods and display system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710517327.XA CN107272960A (en) | 2017-06-29 | 2017-06-29 | A kind of display methods and display system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107272960A true CN107272960A (en) | 2017-10-20 |
Family
ID=60070812
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710517327.XA Pending CN107272960A (en) | 2017-06-29 | 2017-06-29 | A kind of display methods and display system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107272960A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109213323A (en) * | 2018-08-28 | 2019-01-15 | 北京航空航天大学青岛研究院 | A method of screen Attitude estimation is realized based on eye movement interaction technique |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110285947A1 (en) * | 2010-04-09 | 2011-11-24 | Nitto Denko Corporation | Optical compensation film |
CN102551764A (en) * | 2010-12-07 | 2012-07-11 | 富士胶片株式会社 | Radiographic system |
CN102591505A (en) * | 2011-01-12 | 2012-07-18 | 纬创资通股份有限公司 | Electronic device and touch position correction method thereof |
-
2017
- 2017-06-29 CN CN201710517327.XA patent/CN107272960A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110285947A1 (en) * | 2010-04-09 | 2011-11-24 | Nitto Denko Corporation | Optical compensation film |
CN102551764A (en) * | 2010-12-07 | 2012-07-11 | 富士胶片株式会社 | Radiographic system |
CN102591505A (en) * | 2011-01-12 | 2012-07-18 | 纬创资通股份有限公司 | Electronic device and touch position correction method thereof |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109213323A (en) * | 2018-08-28 | 2019-01-15 | 北京航空航天大学青岛研究院 | A method of screen Attitude estimation is realized based on eye movement interaction technique |
CN109213323B (en) * | 2018-08-28 | 2022-04-29 | 北京航空航天大学青岛研究院 | Method for realizing screen attitude estimation based on eye movement interaction technology |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN116348836A (en) | Gesture tracking for interactive game control in augmented reality | |
US10950205B2 (en) | Electronic device, augmented reality device for providing augmented reality service, and method of operating same | |
CN117120962A (en) | Controlling two-handed interactions between mapped hand regions of virtual and graphical elements | |
CN104238923B (en) | A kind of display equipment and its working method | |
US10943388B1 (en) | Intelligent stylus beam and assisted probabilistic input to element mapping in 2D and 3D graphical user interfaces | |
WO2021242634A1 (en) | Interactive augmented reality experiences using positional tracking | |
US20140285520A1 (en) | Wearable display device using augmented reality | |
WO2022006116A1 (en) | Augmented reality eyewear with speech bubbles and translation | |
EP4172955A1 (en) | Generating ground truth datasets for virtual reality experiences | |
WO2022012194A1 (en) | Interaction method and apparatus, display device, and storage medium | |
KR20170043911A (en) | Electronic apparatus and the controlling method thereof | |
CN115735177A (en) | Eyeglasses including shared object manipulation AR experience | |
EP4222550A1 (en) | Augmented reality gaming using virtual eyewear beams | |
CN205608290U (en) | Head -wearing display equipment | |
CN116324679A (en) | Contextually relevant eye-wear remote control | |
CN117321537A (en) | Dynamic power configuration for eyewear devices | |
CN107272960A (en) | A kind of display methods and display system | |
CN107357457A (en) | A kind of display methods and display system | |
US20220197032A1 (en) | Eyewear including a push-pull lens set | |
KR20200066962A (en) | Electronic device and method for providing content based on the motion of the user | |
CN115812189A (en) | Dynamic sensor selection for visual inertial odometer system | |
US20170300121A1 (en) | Input/output device, input/output program, and input/output method | |
CN205750115U (en) | Wearable computing device and the wearable device with it | |
US20230154368A1 (en) | Method and device for controlling luminance of augmented reality (ar) image | |
CN103076969A (en) | Input system for mobile terminal display screen and control method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20171020 |