CN109240493A - A kind of control method and electronic equipment - Google Patents
A kind of control method and electronic equipment Download PDFInfo
- Publication number
- CN109240493A CN109240493A CN201810959168.3A CN201810959168A CN109240493A CN 109240493 A CN109240493 A CN 109240493A CN 201810959168 A CN201810959168 A CN 201810959168A CN 109240493 A CN109240493 A CN 109240493A
- Authority
- CN
- China
- Prior art keywords
- user
- output precision
- control
- eyeball
- parameter
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 210000005252 bulbus oculi Anatomy 0.000 claims abstract description 108
- 210000003128 head Anatomy 0.000 claims abstract description 12
- 210000001747 pupil Anatomy 0.000 claims description 28
- 230000009471 action Effects 0.000 claims description 25
- 230000000007 visual effect Effects 0.000 claims description 17
- 238000012545 processing Methods 0.000 claims description 6
- 241000208340 Araliaceae Species 0.000 claims 1
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 claims 1
- 235000003140 Panax quinquefolius Nutrition 0.000 claims 1
- 235000008434 ginseng Nutrition 0.000 claims 1
- 208000003464 asthenopia Diseases 0.000 abstract description 8
- 239000011521 glass Substances 0.000 description 12
- 230000003287 optical effect Effects 0.000 description 11
- 238000012512 characterization method Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 210000001508 eye Anatomy 0.000 description 4
- 210000000744 eyelid Anatomy 0.000 description 4
- 230000002411 adverse Effects 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000013075 data extraction Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000003542 behavioural effect Effects 0.000 description 1
- 230000003862 health status Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000017260 vegetative to reproductive phase transition of meristem Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
This application discloses a kind of control method and electronic equipment, method includes: acquisition user behavior data;Based on the user behavior data, control parameter is determined;Based on the control parameter, module position parameter of the output precision relative to user eyeball is controlled, the output precision is arranged in user and is worn in the electronic equipment on head.It can be seen that, the application is after collecting user behavior data, control parameter is determined based on user behavior data, and then control the module position parameter that output precision that user is worn in the electronic equipment on head is set relative to user eyeball, come so that output precision is under module position corresponding to user behavior data, so that output precision is in the picture output position that eyeball can be allowed to export relative to user eyeball, and then slow down the visual fatigue that user wears and uses output precision, hence it is evident that improve user experience.
Description
Technical field
The present invention relates to augmented reality field more particularly to a kind of control method and electronic equipments.
Background technique
Currently, augmented reality AR (Augmented Reality) equipment such as AR glasses etc. using more and more, such as swim
Play, audio-visual experience etc..
But AR glasses are worn and used for a long time, it is easy to produce fatigue, causes adverse effect to user experience.
Summary of the invention
In view of this, the present invention provides a kind of control method and electronic equipment, wear for a long time in the prior art to solve
The technical issues of wearing and use AR glasses, visual fatigue is be easy to cause to cause adverse effect to user experience.
The present invention provides a kind of control methods, comprising:
Acquire user behavior data;
Based on the user behavior data, control parameter is determined;
Based on the control parameter, module position parameter of the output precision relative to user eyeball, the output group are controlled
Part is arranged in user and is worn in the electronic equipment on head.
The above method, it is preferable that the user behavior data includes: that user eyeball action data and/or user's input are dynamic
Make data.
The above method, it is preferable that be based on the user behavior data, determine control parameter, comprising:
Data processing is carried out to the user behavior data, to obtain the control parameter to control output precision.
The above method, it is preferable that be based on the control parameter, control the component parameter of output precision, comprising:
It is corresponding with the control parameter relative to the visual field angular dimensions of user eyeball to control output precision.
The above method, it is preferable that visual field angular dimensions and the control parameter of the control output precision relative to user eyeball
It is corresponding, comprising:
It controls the output precision to be rotated in the electronic equipment relative to user eyeball, until the output group
Part is corresponding with the control parameter relative to the upper field angle of user eyeball or lower visual field angular dimensions.
The above method, it is preferable that control the output precision and turned in the electronic equipment relative to user eyeball
It is dynamic, comprising:
It controls the output precision to be rotated in the electronic equipment using user eyeball center as origin, wherein turn
Dynamic radius is the sum of the distance of exit pupil of user eyeball radius and the output precision, so that the emergent pupil size of the output precision
Parameter is at least approximate constant.
Present invention also provides a kind of electronic equipment, comprising:
Output precision is worn on user's head;
Acquisition component, for acquiring user behavior data;
Controller determines control parameter, is based on the control parameter, control institute for being based on the user behavior data
State module position parameter of the output precision relative to user eyeball.
Above-mentioned electronic equipment, preferably:
The controller is specifically used for: control output precision is joined relative to the visual field angular dimensions of user eyeball and the control
Number is corresponding.
Above-mentioned electronic equipment, it is preferable that further include:
Rotational structure is connected with the output precision and the controller respectively, the band under the control of the controller
It moves the output precision to be rotated relative to user eyeball, until upper field angle of the output precision relative to user eyeball
Or lower visual field angular dimensions is corresponding with the control parameter.
Above-mentioned electronic equipment, it is preferable that the rotational structure specifically drives under the control of the controller described defeated
Component is rotated in the electronic equipment using user eyeball center as origin out, wherein the radius of gyration is user eyeball half
The sum of the distance of exit pupil of diameter and the output precision, so that the emergent pupil size parameter of the output precision is at least approximate constant.
It can be seen from the above technical proposal that a kind of control method disclosed in the present application and electronic equipment, are collecting use
After the behavioral data of family, control parameter is determined based on user behavior data, and then control is arranged in user and is worn on head
Module position parameter of the output precision relative to user eyeball in electronic equipment, comes so that output precision is in user behavior number
According under corresponding module position, the picture of eyeball output can be allowed defeated relative to user eyeball so that output precision is in
Out position, and then slow down the visual fatigue that user wears and uses output precision, hence it is evident that improve user experience.
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below
There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this
Some embodiments of invention for those of ordinary skill in the art without creative efforts, can be with
It obtains other drawings based on these drawings.
Fig. 1 is a kind of flow chart for control method that the embodiment of the present application one provides;
Fig. 2~Fig. 4 is respectively the application exemplary diagram of the embodiment of the present application;
Fig. 5 is the structural schematic diagram for a kind of electronic equipment that the embodiment of the present application two provides;
Fig. 6 and Fig. 7 is respectively the partial structure diagram of the embodiment of the present application two;
Fig. 8 and Fig. 9 is respectively the other application exemplary diagram of the embodiment of the present application.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete
Site preparation description, it is clear that described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.It is based on
Embodiment in the present invention, it is obtained by those of ordinary skill in the art without making creative efforts every other
Embodiment shall fall within the protection scope of the present invention.
It is a kind of implementation flow chart for control method that the embodiment of the present application one provides with reference to Fig. 1, the control in the present embodiment
Method processed suitable for the electronic equipment for being worn on user's head, such as AR/ Virtual Reality (Virtual Reality) glasses or
Other terminal devices with output precision.Wherein, output precision can export picture data, to be imaged in user eyeball,
The lighting apparatus component in the visual experience of virtual reality, such as AR glasses is provided for user, and picture data is exported in the form of light, and
It is output in user eyeball, provides game experiencing for user.
In the present embodiment, which may comprise steps of:
Step 101: acquisition user behavior data.
Wherein, user behavior data refers to, characterization user watches the data for the use intention being intended to or to electronic equipment,
User behavior data can be acquired by the sensor of setting on an electronic device in the present embodiment.
Step 102: being based on user behavior data, determine control parameter.
Wherein, control parameter can be characterization user to the device requirements parameter of output precision in electronic equipment, such as export
Output parameter etc. when component location parameter in the electronic device and/or output precision output data.
Step 103: being based on control parameter, control module position parameter of the output precision relative to user eyeball.
That is, controlling group of the output precision relative to user eyeball in the present embodiment after determining control parameter
Part location parameter is corresponding with control parameter, so that user eyeball is in when receiving the light data that output precision is exported
Comfortable state will not cause visual fatigue.
By above scheme it is found that a kind of control method that the embodiment of the present application one provides, is collecting user behavior data
Later, control parameter is determined based on user behavior data, and then control is arranged in user and is worn in the electronic equipment on head
Module position parameter of the output precision relative to user eyeball, come so that output precision is in corresponding to user behavior data
Under module position, so that output precision is in the picture output position that eyeball can be allowed to export relative to user eyeball, into
And slow down the visual fatigue that user wears and uses output precision, hence it is evident that improve user experience.
In one implementation, user behavior data may include having: the action data of user eyeball and/or user are defeated
Enter action data.
Wherein, the action data of user eyeball is it is to be understood that movement of the user eyeball when watching scenery such as rotates shape
The data such as the state after state or rotation, to characterize the eyeball viewed status of relative comfort locating for user eyeball, for example, user
The action data of eyeball can be with are as follows: angle that user's eyelid opens, eyeball downward or upward rotational angle, sight moving direction or
The data such as move angle, as shown in the rotation of apparent time under eyeball in Fig. 2, user does not need that eyelid is forced to open dominant bit at this time
The sight viewing angle set and eyeball is made to be at relative comfort.
Specifically, the action data of user eyeball can be the real-time eyeball action data of active user, it is also possible to root
The action data that user eyeball is under the sight viewing angle of the most comfortable is obtained according to historical data such as biological experiment data.
Correspondingly, in the present embodiment real-time user can be acquired by the iris sensor being arranged on electronic equipment such as AR glasses
The action data of eyeball;Alternatively, using data-interface etc. on network or in preset server obtain active user or
The action data of ordinary user's eyeball in masses.
Wherein, user's input action data are it is to be understood that user watches the output of output precision in use electronic equipment
When picture in advance or the setting data that input in real time, such as the setting parameter of output precision, to characterize user to the need of viewing
It asks.User's input action data can be acquired by input equipment such as touch screen or keyboard etc. in the present embodiment.
In one implementation, the present embodiment step 102 can specifically be realized by following steps:
Data processing is carried out to user behavior data, to obtain the control parameter to control output precision.
It wherein, can be by carrying out the processing such as data extraction or parsing to user behavior data, to obtain in the present embodiment
To the parameters of characterization user demand, the i.e. control parameter to control output precision.For example, to user eyeball action data
Parsed, i.e., from user eyeball upward or downward rotational angle and sight move up or down and the data such as move angle in
Extract control parameter, such as be aligned with user's sight location parameter or output angle, the field angle relative to user eyeball
Parameter.
In one implementation, step 103 can specifically be realized by following steps in the present embodiment:
It is corresponding with control parameter relative to the visual field angular dimensions of user eyeball to control output precision, so that output precision
The location status that picture data is comfortably watched by user eyeball can be exported by being under the control of current control parameter.
Specifically, rotated in the electronic device relative to user eyeball in the present embodiment by controlling output precision,
Until output precision is corresponding with control parameter relative to the upper field angle of user eyeball or lower visual field angular dimensions, such as institute in Fig. 3
Showing, output precision reduces by 5 degree relative to the upper field angle of eyeball in control parameter, and lower field angle increases by 5 degree accordingly, by
This, controls output precision and rotates down in the electronic device relative to user eyeball, until output precision is opposite in the present embodiment
In user eyeball upper field angle compared to rotation before reduce by 5 degree, lower field angle increases by 5 degree, reaches and exports picture data and allow user
The purpose that eyeball is comfortably watched.
And in order to guarantee output precision emergent pupil size parameter will not because of the rotation of output precision in the electronic device and
It changes or changes greatly, turn of output precision in the electronic device can be controlled in the present embodiment in the following manner
It is dynamic, specifically:
Control output precision is rotated using user eyeball center as origin in the electronic device, and the radius of gyration is user
The sum of eyeball radius and the distance of exit pupil of output precision, as shown in Figure 4, the thus emergent pupil size parameter of output precision and rotation
It is before at least approximate constant, viewing effect of the user eyeball to the picture data that output precision exports will not be generated so bad
It influences, will not such as reduce emergent pupil size, and then the visual effect that user watches picture would not be reduced.
With reference to Fig. 5, for the structural schematic diagram for a kind of electronic equipment that the embodiment of the present application two provides, which can be with
For equipment such as AR or VR glasses, it is worn on to provide the viewing experiences such as picture such as game or video for user.
In the present embodiment, which may include with flowering structure:
Output precision 501 is worn on user's head, and as shown in Figure 6, output precision 501 corresponds specifically to user's head
The setting of eyeball position, to export picture data, such as game picture or video pictures to user eyeball.
Acquisition component 502, for acquiring user behavior data.
Wherein, user behavior data refers to, characterization user watches the data for the use intention being intended to or to electronic equipment,
User behavior data can be acquired by the sensor of setting on an electronic device in the present embodiment.That is, this
Acquisition component 502 in embodiment can be the various sensors that can acquire user behavior data.
It should be noted that user behavior data may include having: the action data of user eyeball and/or user's input are dynamic
Make data.
Wherein, the action data of user eyeball is it is to be understood that movement of the user eyeball when watching scenery such as rotates shape
The data such as the state after state or rotation, to characterize the eyeball viewed status of relative comfort locating for user eyeball, for example, user
The action data of eyeball can be with are as follows: angle that user's eyelid opens, eyeball downward or upward rotational angle, sight moving direction or
The data such as move angle, as shown in Figure 2, user does not need that eyelid is forced to open maximum position and enable eyeball at this time
Sight viewing angle in relative comfort.
Specifically, the action data of user eyeball can be the real-time eyeball action data of active user, it is also possible to root
The action data that user eyeball is under the sight viewing angle of the most comfortable is obtained according to historical data such as biological experiment data.
Correspondingly, in the present embodiment can by the acquisition component 502 that is arranged on electronic equipment such as AR glasses such as iris sensor come
Inspection acquires the action data of real-time user eyeball;Alternatively, using data-interface etc. on network or in preset server
Obtain the action data of ordinary user's eyeball in active user or masses, wherein ordinary user can refer to that augen state is
The user of common eyeball standard health status.
Wherein, user's input action data are it is to be understood that user watches the output of output precision in use electronic equipment
When picture in advance or the setting data that input in real time, such as the setting parameter of output precision, to characterize user to the need of viewing
It asks.It in the present embodiment can include but is not limited to that touch screen or keyboard acquisition are used by such as input equipment of acquisition component 502
Family input action data.
Controller 503 is connected with acquisition component 502 and output precision 501, for being collected based on acquisition component 502
User behavior data, determine control parameter, and the component based on control parameter control output precision 501 relative to user eyeball
Location parameter.
Wherein, wherein control parameter can be characterization user to the device requirements parameter of output precision in electronic equipment, such as
Output parameter etc. when output precision location parameter in the electronic device and/or output precision output data.Controller 503
It can be by carrying out data processing to user behavior data, to obtain the control parameter to control output precision.Specifically, control
Device 503 processed can be by carrying out the processing such as data extraction or parsing to user behavior data, to obtain characterization user demand
Parameters, the i.e. control parameter to control output precision.For example, parsed to user eyeball action data, i.e., from
Family eyeball upward or downward rotational angle and sight move up or down and the data such as move angle in extract control parameter,
The location parameter or output angle that are such as aligned with user's sight, the field angle parameter relative to user eyeball.
That is, in the present embodiment controller 503 determine control parameter after, control output precision 501 relative to
The module position parameter of family eyeball is corresponding with control parameter, so that user eyeball receives what output precision 501 was exported
It is in comfortable state when light data, visual fatigue will not be caused.
By above scheme it is found that a kind of electronic equipment that the embodiment of the present application two provides, is collecting user behavior data
Later, control parameter is determined based on user behavior data, and then control is arranged in user and is worn in the electronic equipment on head
Module position parameter of the output precision relative to user eyeball, come so that output precision is in corresponding to user behavior data
Under module position, so that output precision is in the picture output position that eyeball can be allowed to export relative to user eyeball, into
And slow down the visual fatigue that user wears and uses output precision, hence it is evident that improve user experience.
In one implementation, the controller 503 in electronic equipment is controlling 501 phase of output precision based on control parameter
For user eyeball module position parameter when, can specifically be accomplished by the following way:
It is corresponding with control parameter relative to the visual field angular dimensions of user eyeball to control output precision 501, so that output
Component 501 is under the control of current control parameter can export picture data by the comfortable state of user eyeball.
Specifically, as shown in Figure 7, can be turned an angle by being arranged in the electronic device in the present embodiment
Rotational structure 504 realize the control of output precision 501, it is specific:
Rotational structure 503 is connected with output precision 501 and controller 503 respectively, is arranged in electronic equipment such as AR glasses
In, also, drive output precision 501 to be rotated relative to user eyeball under the control of controller 503, until output precision
501 is corresponding with control parameter relative to the upper field angle of user eyeball or lower visual field angular dimensions.As shown in Figure 3, it is controlling
Output precision reduces by 5 degree relative to the upper field angle of eyeball in parameter, and lower field angle increases by 5 degree accordingly, as a result, the present embodiment
Middle controller 503 controls output precision 501 and rotates down in the electronic device relative to user eyeball, until output precision 501
5 degree are reduced before rotating compared with the upper field angle of user eyeball, lower field angle increases by 5 degree, reaches output picture data and allows
The purpose that user eyeball is comfortably watched.
It is used to controlling it should be noted that driving device can be arranged in the present embodiment in the electronic device in the concrete realization
Rotational structure 504 is driven to rotate under the control of device 503 processed, such as electric drive or Mechanical Driven, and then the rotational structure for being
504 are able to drive output precision 501 rotates in the electronic device, is turning to output precision 501 relative to the upper of user eyeball
It stops operating when in field angle or lower visual field angular dimensions position corresponding with control parameter, as a result, at output precision 501
In picture data can be exported by the comfortable state of user eyeball, avoids user from wearing for a long time with this and use electronic equipment institute
Caused visual fatigue brings better viewing experience for user.
And in order to guarantee that the emergent pupil size parameter of output precision 501 will not be because of output precision 501 in the electronic device
It rotates and changes or change greatly, in the present embodiment, rotation result 504 is specifically the band under the control of controller 503
Dynamic output precision 501 is rotated using user eyeball center as origin in the electronic device, and the radius of gyration is user eyeball half
The sum of the distance of exit pupil of diameter and output precision 501, the emergent pupil size parameter of output precision 501 and rotation are at least close before as a result,
Like constant, the viewing effect generation adverse effect for the picture data that will not be exported in this way to user eyeball to output precision 501, such as
Emergent pupil size, which will not be reduced, would not reduce the visual effect that user watches picture.
The scheme in the embodiment of the present application is carried out in the concrete application in AR glasses below simple for example:
As shown in Figure 8, output precision such as the field angle FOV (Field of the optical system in AR glasses under normal conditions
Of view) be it is symmetrical, i.e., upper field angle is identical as lower field angle, and in use, user is in order to see for this optical system
The content of complete upper field angle just needs to be rotated up eyeball, but people's ordinary practice, in head-up or lower view, this makes user
It is more comfortable in use, so to reduce field angle to the greatest extent, increase lower field angle, with reference to example in Fig. 3.
And the usage experience of user is improved in the present embodiment, it is realized using the scheme of rotary optical system and increases lower view
Rink corner, the method for field angle in reduction.Certain rotary optical system is not arbitrarily, to select suitable rotation reference, this
Sample just can guarantee that the optical parameter of original design does not change.
Specifically, optical system exit end is referred to as the distance of exit pupil ER of optical system to the distance between pupil of human
(exit relief), at all convergences of rays and emergent pupil, emergent pupil size and distance of exit pupil are two significant designs of optical system
Index.For comfortable wearing, distance of exit pupil is typically larger than 20mm, and emergent pupil size is generally greater than the pupil size of human eye, this
Sample can make human eye easily see full images when observing image.
As shown in Figure 9, using eyeball center as origin, optical system is rotated down, eyeball center is usually after eye pupil
13.5mm, so the radius of gyration is ER+13.5, when human eye ball rotates down, the emergent pupil of optical system and the pupil of human eye
Hole is overlapped, and will not lose emergent pupil size.Rotational angle is determined according to optical system FOV, usually upper 5 ° or so of field angle, to protect
It holds and regards habit under user, alleviate the fatigue for wearing AR glasses for a long time, achieving the purpose that, which improves user, wears experience.
Each embodiment in this specification is described in a progressive manner, the highlights of each of the examples are with other
The difference of embodiment, the same or similar parts in each embodiment may refer to each other.For device disclosed in embodiment
For, since it is corresponded to the methods disclosed in the examples, so being described relatively simple, related place is said referring to method part
It is bright.
Professional further appreciates that, unit described in conjunction with the examples disclosed in the embodiments of the present disclosure
And algorithm steps, can be realized with electronic hardware, computer software, or a combination of the two, in order to clearly demonstrate hardware and
The interchangeability of software generally describes each exemplary composition and step according to function in the above description.These
Function is implemented in hardware or software actually, the specific application and design constraint depending on technical solution.Profession
Technical staff can use different methods to achieve the described function each specific application, but this realization is not answered
Think beyond the scope of this invention.
The step of method described in conjunction with the examples disclosed in this document or algorithm, can directly be held with hardware, processor
The combination of capable software module or the two is implemented.Software module can be placed in random access memory (RAM), memory, read-only deposit
Reservoir (ROM), electrically programmable ROM, electrically erasable ROM, register, hard disk, moveable magnetic disc, CD-ROM or technology
In any other form of storage medium well known in field.
The foregoing description of the disclosed embodiments enables those skilled in the art to implement or use the present invention.
Various modifications to these embodiments will be readily apparent to those skilled in the art, as defined herein
General Principle can be realized in other embodiments without departing from the spirit or scope of the present invention.Therefore, of the invention
It is not intended to be limited to the embodiments shown herein, and is to fit to and the principles and novel features disclosed herein phase one
The widest scope of cause.
Claims (10)
1. a kind of control method, comprising:
Acquire user behavior data;
Based on the user behavior data, control parameter is determined;
Based on the control parameter, module position parameter of the output precision relative to user eyeball is controlled, the output precision is set
It sets in the electronic equipment that user is worn on head.
2. the method according to claim 1, wherein the user behavior data includes: user eyeball movement number
According to and/or user's input action data.
3. method according to claim 1 or 2, which is characterized in that be based on the user behavior data, determine control ginseng
Number, comprising:
Data processing is carried out to the user behavior data, to obtain the control parameter to control output precision.
4. method according to claim 1 or 2, which is characterized in that be based on the control parameter, control the group of output precision
Part parameter, comprising:
It is corresponding with the control parameter relative to the visual field angular dimensions of user eyeball to control output precision.
5. according to the method described in claim 4, it is characterized in that, control output precision is joined relative to the field angle of user eyeball
Number is corresponding with the control parameter, comprising:
It controls the output precision to be rotated in the electronic equipment relative to user eyeball, until the output precision phase
Upper field angle or lower visual field angular dimensions for user eyeball is corresponding with the control parameter.
6. according to the method described in claim 5, it is characterized in that, to control the output precision opposite in the electronic equipment
It is rotated in user eyeball, comprising:
It controls the output precision to be rotated in the electronic equipment using user eyeball center as origin, wherein rotation half
Diameter is the sum of the distance of exit pupil of user eyeball radius and the output precision, so that the emergent pupil size parameter of the output precision
It is at least approximate constant.
7. a kind of electronic equipment, comprising:
Output precision is worn on user's head;
Acquisition component, for acquiring user behavior data;
Controller determines control parameter, is based on the control parameter, control described defeated for being based on the user behavior data
Module position parameter of the component relative to user eyeball out.
8. electronic equipment according to claim 7, it is characterised in that:
The controller is specifically used for: visual field angular dimensions and the control parameter phase of the control output precision relative to user eyeball
It is corresponding.
9. electronic equipment according to claim 8, which is characterized in that further include:
Rotational structure is connected with the output precision and the controller respectively, and institute is driven under the control of the controller
Output precision is stated to be rotated relative to user eyeball, until the output precision relative to the upper field angle of user eyeball or under
Visual field angular dimensions is corresponding with the control parameter.
10. electronic equipment according to claim 9, which is characterized in that the rotational structure is specifically in the controller
Control under drive the output precision to be rotated in the electronic equipment using user eyeball center as origin, wherein turn
Dynamic radius is the sum of the distance of exit pupil of user eyeball radius and the output precision, so that the emergent pupil size of the output precision
Parameter is at least approximate constant.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810959168.3A CN109240493A (en) | 2018-08-22 | 2018-08-22 | A kind of control method and electronic equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810959168.3A CN109240493A (en) | 2018-08-22 | 2018-08-22 | A kind of control method and electronic equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109240493A true CN109240493A (en) | 2019-01-18 |
Family
ID=65068422
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810959168.3A Pending CN109240493A (en) | 2018-08-22 | 2018-08-22 | A kind of control method and electronic equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109240493A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112261408A (en) * | 2020-09-16 | 2021-01-22 | 青岛小鸟看看科技有限公司 | Image processing method and device for head-mounted display equipment and electronic equipment |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN2819243Y (en) * | 2005-08-19 | 2006-09-20 | 宝陆科技股份有限公司 | Adjusting structure of helmet displaying device |
US20130328925A1 (en) * | 2012-06-12 | 2013-12-12 | Stephen G. Latta | Object focus in a mixed reality environment |
CN105334961A (en) * | 2015-10-27 | 2016-02-17 | 惠州Tcl移动通信有限公司 | Method for controlling mobile terminal based on eyeball tracking and mobile terminal |
CN106708270A (en) * | 2016-12-29 | 2017-05-24 | 宇龙计算机通信科技(深圳)有限公司 | Display method and apparatus for virtual reality device, and virtual reality device |
CN107409189A (en) * | 2015-04-07 | 2017-11-28 | 奥林巴斯株式会社 | Glasses type display device |
CN108089332A (en) * | 2017-12-15 | 2018-05-29 | 歌尔科技有限公司 | VR wears display device and display methods |
WO2018120554A1 (en) * | 2016-12-26 | 2018-07-05 | 华为技术有限公司 | Image display method and head-mounted display device |
-
2018
- 2018-08-22 CN CN201810959168.3A patent/CN109240493A/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN2819243Y (en) * | 2005-08-19 | 2006-09-20 | 宝陆科技股份有限公司 | Adjusting structure of helmet displaying device |
US20130328925A1 (en) * | 2012-06-12 | 2013-12-12 | Stephen G. Latta | Object focus in a mixed reality environment |
CN107409189A (en) * | 2015-04-07 | 2017-11-28 | 奥林巴斯株式会社 | Glasses type display device |
CN105334961A (en) * | 2015-10-27 | 2016-02-17 | 惠州Tcl移动通信有限公司 | Method for controlling mobile terminal based on eyeball tracking and mobile terminal |
WO2018120554A1 (en) * | 2016-12-26 | 2018-07-05 | 华为技术有限公司 | Image display method and head-mounted display device |
CN106708270A (en) * | 2016-12-29 | 2017-05-24 | 宇龙计算机通信科技(深圳)有限公司 | Display method and apparatus for virtual reality device, and virtual reality device |
CN108089332A (en) * | 2017-12-15 | 2018-05-29 | 歌尔科技有限公司 | VR wears display device and display methods |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112261408A (en) * | 2020-09-16 | 2021-01-22 | 青岛小鸟看看科技有限公司 | Image processing method and device for head-mounted display equipment and electronic equipment |
WO2022057782A1 (en) * | 2020-09-16 | 2022-03-24 | 青岛小鸟看看科技有限公司 | Image processing method and apparatus for head-mounted display device, and electronic device |
CN112261408B (en) * | 2020-09-16 | 2023-04-25 | 青岛小鸟看看科技有限公司 | Image processing method and device for head-mounted display equipment and electronic equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11836289B2 (en) | Use of eye tracking to adjust region-of-interest (ROI) for compressing images for transmission | |
WO2017124794A1 (en) | Display control device, control method therefor, and display control system | |
US20220397959A1 (en) | Methods and Systems for Enabling the Remote Testing of Vision and Diagnosis of Vision-Related Issues | |
WO2018211673A1 (en) | Image generation device, image display system, and image generation method | |
EP3549095B1 (en) | Methods and devices for displaying image with changed field of view | |
JP2020042807A (en) | Foveated rendering system and method | |
AU2023285715A1 (en) | Wearable image manipulation and control system with correction for vision defects and augmentation of vision and sensing | |
CN111447868A (en) | Method and apparatus for treating double vision and insufficient convergence disorders | |
CN103513768A (en) | Control method and device based on posture changes of mobile terminal and mobile terminal | |
WO2018211672A1 (en) | Image generation device, image display system, and image generation method | |
JP7001832B2 (en) | Image adjustment of the optotype tracking system | |
CN111880654A (en) | Image display method and device, wearable device and storage medium | |
CN106913449A (en) | Interactive augmented reality eyesight convalescence device | |
CN107307981B (en) | Control method of head-mounted display device | |
EP3615986A1 (en) | Wearable image manipulation and control system with correction for vision defects and augmentation of vision and sensing | |
CN104766056A (en) | Human-computer interaction method, device and virtual headset device | |
CN105943327A (en) | Vision-exercising health caring system with anti-dizziness device | |
CN106200908A (en) | A kind of control method and electronic equipment | |
CN109240493A (en) | A kind of control method and electronic equipment | |
CN207216145U (en) | A kind of Wearable | |
CN110840721B (en) | Near-eye display-based eyesight prevention, control and training method and near-eye display equipment | |
CN107137211A (en) | The 3D vision training methods moved based on eye | |
CN113960788B (en) | Image display method, device, AR glasses and storage medium | |
CA2963252C (en) | Multiple data sources of captured data into single newly rendered video feed | |
WO2023276566A1 (en) | Image display system and image display method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190118 |