CN104793749B - Intelligent glasses and its control method, device - Google Patents
Intelligent glasses and its control method, device Download PDFInfo
- Publication number
- CN104793749B CN104793749B CN201510219994.0A CN201510219994A CN104793749B CN 104793749 B CN104793749 B CN 104793749B CN 201510219994 A CN201510219994 A CN 201510219994A CN 104793749 B CN104793749 B CN 104793749B
- Authority
- CN
- China
- Prior art keywords
- gesture identification
- identification region
- gesture
- area
- intelligent glasses
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Landscapes
- User Interface Of Digital Computer (AREA)
Abstract
The disclosure is directed to intelligent glasses and its control method, device, this method to include:In the viewfinder range of the intelligent glasses, gesture identification region is determined;Acquire the gesture information in the gesture identification region;Zoom instructions are generated based on the gesture information.Using the embodiment of the present disclosure, finger, which has only been placed on specific regional location, can trigger the acquisition and subsequent Zoom control process of gesture information, be greatly lowered the probability of maloperation appearance in this way, improve the precision of gesture Zoom control.
Description
Technical field
This disclosure relates to wearable device technical field more particularly to a kind of intelligent glasses and its control method, device.
Background technique
With making rapid progress for science and technology, wearable device because its function multiplicity and what is used portable progress into people's
Daily life.For example, intelligent glasses can by installing camera and circuit devcie relevant to camera on mirror holder
Realize camera function.
In the related technology, user can carry out sliding touch or hand by any position in intelligent glasses viewfinder range
Gesture control, come to intelligent glasses camera focal length or electronics focal length be adjusted, this mode easily causes maloperation, such as
When shooting, occurring acquaintance in viewfinder range, user lifts hand greeting and is possible to by intelligent glasses erroneous judgement be control gesture,
And lead to maloperation, it makes troubles for user.
Summary of the invention
Present disclose provides a kind of intelligent glasses and its control methods, device, to solve in the related technology by finding a view
Any position in range touch or gesture operation is come the problem of controlling focal length, easily cause maloperation.
According to the first aspect of the embodiments of the present disclosure, a kind of control method of intelligent glasses is provided, including:
In the viewfinder range of the intelligent glasses, gesture identification region is determined;
Acquire the gesture information in the gesture identification region;
Zoom instructions are generated based on the gesture information.
Optionally, described in the viewfinder range of the intelligent glasses, determine gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
In the setting position of the viewfinder range, gesture identification region is shown based on preset shapes and sizes.
Optionally, described in the viewfinder range of the intelligent glasses, determine gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
It focuses to the scenery to be shot in the viewfinder range, determines focusing area;
In the viewfinder range in addition to the focusing area, in non-focusing area that area ratio preset area is big,
Show gesture identification region.
Optionally, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, it is described non-right to randomly choose one of those
Burnt region shows the gesture identification region.
Optionally, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, preset positional priority is read;
It is greater than non-focusing area location, selection in the viewfinder range of setting area based on each area
The highest non-focusing area of positional priority shows the gesture identification region.
Optionally, before the determining gesture identification region, the method also includes:
Luminaire is opened, so that the luminaire emits the light of the first color along the edge of the viewfinder range.
Optionally, after the determining gesture identification region, the method also includes:
Luminaire is opened, so that the luminaire emits the second color along the edge in the gesture identification region
Light.
Optionally, the shape in the gesture identification region is rectangle, circle or hand shape.
Optionally, before the gesture information in the acquisition gesture identification region, the method also includes:
Acquire the hand shape information in the gesture identification region;
By collected hand shape information compare with pre-stored hand shape template, obtain similarity;
Then the gesture information in the acquisition gesture identification region is executed when the similarity is greater than hand shape threshold value.
Optionally, the gesture information in the acquisition gesture identification region, including:
Acquire the motion information of the finger in the gesture identification region;
The path curves of the finger as the gesture information are generated based on the motion information.
Optionally, zoom instructions are generated based on the gesture information, including:
The path curves are compared respectively with pre-stored amplification track template, obtain amplification similarity;
When the amplification similarity is greater than default amplification threshold value, magnification ratio is generated based on the path curves;
Amplification zoom instructions are generated based on the magnification ratio.
Optionally, the path curves are that the path curves of two fingers openings or a finger are drawn a circle
Path curves.
Optionally, zoom instructions are generated based on the gesture information, including:
The path curves are compared respectively with pre-stored diminution track template, obtain reducing similarity;
When the diminution similarity is greater than default diminution threshold value, diminution ratio is generated based on the path curves;
It is generated based on the diminution ratio and reduces zoom instructions.
Optionally, the path curves are the path curves or a finger click that two fingers are mediated
Path curves.
Optionally, after the generation zoom instructions based on the gesture information, the method also includes:
Zoom control is carried out to the intelligent glasses based on the zoom instructions.
Optionally, after the generation zoom instructions based on the gesture information, the method also includes:
Zoom control is carried out based on the zoom instructions pair and the camera that the intelligent glasses are wirelessly connected.
According to the second aspect of an embodiment of the present disclosure, a kind of control device of intelligent glasses is provided, including:
Gesture identification area determination module, for determining gesture identification region in the viewfinder range of the intelligent glasses;
Gesture information acquisition module, for acquiring the gesture information in the gesture identification region;
Zoom instructions generation module, for generating zoom instructions based on the gesture information.
Optionally, the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Gesture identification region display sub-module is based on preset shape for the setting position in the viewfinder range
Gesture identification region is shown with size.
Optionally, the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Focusing area determines submodule, for focusing to the scenery to be shot in the viewfinder range, determining pair
Burnt region;
Gesture identification region display sub-module, in the viewfinder range in addition to the focusing area, area
In the non-focusing area bigger than preset area, gesture identification region is shown.
Optionally, gesture identification region display sub-module, including:
Submodule is selected, when the non-focusing area for being greater than preset area when area has multiple, randomly chooses it
In non-focusing area show the gesture identification region.
Optionally, gesture identification region display sub-module, including:
Positional priority reading submodule, when the non-focusing area for being greater than preset area when area has multiple,
Read preset positional priority;
Submodule is selected, for being greater than the non-focusing area of setting area based on each area in the viewfinder range
Location selects the highest non-focusing area of the positional priority to show the gesture identification region.
Optionally, described device further includes:
First light emitting module, for emitting the light of the first color along the edge of the viewfinder range.
Optionally, described device further includes:
Second light emitting module emits the light of the second color for the edge along the gesture identification region.
Optionally, the shape in the gesture identification region that the gesture identification area determination module determines be rectangle,
Round or hand shape.
Optionally, described device further includes:
Hand shape information acquisition module, for acquiring the hand shape information in the gesture identification region;
Hand shape comparison module, for by collected hand shape information compare with pre-stored hand shape template, obtain phase
Like degree;
Then the gesture information acquisition module acquires the gesture information in the gesture identification region and compares in the hand shape
The execution when similarity that module is compared is greater than hand shape threshold value.
Optionally, the gesture information acquisition module includes:
Motion information acquisition submodule, for acquiring the motion information of the finger in the gesture identification region;
Path curves generate submodule, for being generated based on the motion information as described in the gesture information
The path curves of finger.
Optionally, the zoom instructions generation module includes:
Amplify comparer module, for comparing the path curves and pre-stored amplification track template respectively
It is right, obtain amplification similarity;
Magnification ratio generates submodule, for being based on the fortune when the amplification similarity is greater than default amplification threshold value
Dynamic rail trace curve generates magnification ratio;
Amplify zoom instructions and generate submodule, for generating amplification zoom instructions based on the magnification ratio.
Optionally, it is that two fingers open that the path curves, which generate the path curves that submodule generates,
Path curves or the path curves drawn a circle of a finger.
Optionally, the zoom instructions generation module includes:
Comparer module is reduced, for comparing the path curves and pre-stored diminution track template respectively
It is right, it obtains reducing similarity;
Diminution ratio generates submodule, for being based on the fortune when the diminution similarity is greater than default diminution threshold value
Dynamic rail trace curve generates diminution ratio;
It reduces zoom instructions and generates submodule, reduce zoom instructions for generating based on the diminution ratio.
Optionally, it is that two fingers are mediated that the path curves, which generate the path curves that submodule generates,
Path curves or a finger click path curves.
Optionally, described device further includes:
First Zoom control module, for carrying out Zoom control to the intelligent glasses based on the zoom instructions.
Optionally, described device further includes:
Second Zoom control module, camera for being wirelessly connected based on the zoom instructions pair with the intelligent glasses into
Row Zoom control.
According to the second aspect of an embodiment of the present disclosure, a kind of intelligent glasses are provided, including:Mirror holder and above-mentioned Brilliant Eyes
The control device of mirror, the control device of the intelligent glasses are mounted on the mirror holder.
The technical scheme provided by this disclosed embodiment can include the following benefits:
The control device of intelligent glasses can determine gesture identification region first in the disclosure, then to gesture identification region
In gesture information be acquired, and based on the gesture information of acquisition to intelligent glasses carry out Zoom control.That is, finger
The acquisition and subsequent Zoom control process of gesture information can be triggered by being only placed on specific regional location, thus greatly
The probability for reducing maloperation appearance greatly, improves the precision of gesture Zoom control.
Control device can show that gesture is known according to the position, shape and size of setting in viewfinder range in the disclosure
Other region.Focusing is also based on as a result, determining in non-focusing area and showing gesture identification region, so that user is passing through
When gesture focuses, the main subject in viewfinder range will not be stopped, be conducive to user and carry out focus adjustment and shooting.
Control device is also based on positional priority to select the non-focusing area in display gesture identification region in the disclosure
Domain, so as to select user be most accustomed to most convenient operation region come show gesture identification region and execute gesture.
Control device identifies the edge of viewfinder range by the light of readily identified color in the disclosure, is conducive to user
Range based on viewfinder range adjustment intelligent glasses.Pass through readily identified color and is different from identifying the color of viewfinder range
Light identify the edge in gesture identification region, be conducive to user and identify gesture identification region, execute control in correct position
The gesture of focal length.Moreover, the shape in gesture identification region can be flexible in conjunction with carry out such as habit, the sizes of viewfinder range of user
Setting.
Control device first identifies the hand shape for entering gesture identification region in the disclosure, the hand of only specified hand shape
Gesture information can be only used for carrying out focus controlling, to further avoid maloperation.
Control device can be with movement that any two fingers or specified two fingers open or a finger in the disclosure
The movement drawn a circle is as amplification instruction.It can be with movement that any two fingers or specified two fingers are mediated or a finger
The movement of click is as amplification instruction.It is easy to operate, it is easy to accomplish.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not
The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows the implementation for meeting the disclosure
Example, and together with specification for explaining the principles of this disclosure.
Fig. 1 is a kind of disclosure control method flow chart of intelligent glasses shown according to an exemplary embodiment.
Fig. 2 (A), (B) are the two of disclosure viewfinder range shown according to an exemplary embodiment and gesture identification region
Width schematic diagram.
Fig. 3 (A), (B) are two width schematic diagrames of disclosure zoom gesture shown according to an exemplary embodiment.
Fig. 4 is a kind of disclosure control device block diagram of intelligent glasses shown according to an exemplary embodiment.
Fig. 5 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 6 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 7 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 8 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 9 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 10 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 11 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 12 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 13 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 14 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 15 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 16 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 17 is a kind of disclosure structural schematic diagram of intelligent glasses shown according to an exemplary embodiment.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to
When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment
Described in embodiment do not represent all implementations consistent with this disclosure.On the contrary, they be only with it is such as appended
The example of the consistent device and method of some aspects be described in detail in claims, the disclosure.
It is only to be not intended to be limiting the disclosure merely for for the purpose of describing particular embodiments in the term that the disclosure uses.
The "an" of the singular used in disclosure and the accompanying claims book, " described " and "the" are also intended to including majority
Form, unless the context clearly indicates other meaning.It is also understood that term "and/or" used herein refers to and wraps
It may be combined containing one or more associated any or all of project listed.
It will be appreciated that though various information, but this may be described using term first, second, third, etc. in the disclosure
A little information should not necessarily be limited by these terms.These terms are only used to for same type of information being distinguished from each other out.For example, not departing from
In the case where disclosure range, the first information can also be referred to as the second information, and similarly, the second information can also be referred to as
One information.Depending on context, word as used in this " if " can be construed to " ... when " or " when ...
When " or " in response to determination ".
As shown in Figure 1, Fig. 1 is a kind of control method flow chart of intelligent glasses shown according to an exemplary embodiment,
This method can be used in intelligent glasses, include the following steps:
Step 101, in the viewfinder range of intelligent glasses, determine gesture identification region.
In disclosure step, firstly, opening the shooting function of intelligent glasses, viewfinder range is determined.
When user presses shutter to the intelligent glasses worn, or otherwise start the shooting function of intelligent glasses
When, as shown in Fig. 2 (A) and (B), intelligent glasses automatically determine viewfinder range with the program set.For example, by the practical covering of camera lens
The region of the 75%-90% of range is determined as viewfinder range.The movement found a view usually by the view finder that is arranged on mirror holder Lai
It realizes, which can be using common view finder in the related technology, and details are not described herein.
In the embodiment of the present disclosure, the shape in the gesture identification region can be the rectangle as shown in Fig. 2 (A), can be with
It can also be the hand shape as shown in Fig. 2 (B) for circle or other shapes, to guide user that specified finger is put into hand shape region,
Carry out zoom operation, it should be noted that show the entire hand shape including five fingers in Fig. 2 (B), in fact or
The shape of the finger of two or other numbers.
It is also provided with laser or other luminaires in disclosure step, on mirror holder, is opening shooting function
When, while laser is opened, and make laser along the edge transmitting red of viewfinder range or other readily identified face
The laser of color shoots the things in viewfinder range so that viewfinder range is definitely convenient for user.
Then, gesture identification region is determined in viewfinder range.
In disclosure step, the position in gesture identification region can be preset, which can be set
At multiple positions in viewfinder range, such as:Gesture identification region is arranged using the habit of left or right hand according to user
Lower right corner or the upper left corner of viewfinder range etc..And the gesture identification region can be preset shapes and sizes.
In disclosure step, it is also based on the scenery setting gesture identification region in current viewfinder range.
In this case, firstly, focusing to the main scenery to be shot in viewfinder range, auto-focusing is determined
The region (Auto Focus, AF).
Main scenery in viewfinder range includes people, animals and plants, building etc..The focusing mode can be in the related technology
Single point focalizing, multipoint focalizing, manual focus, auto-focusing, continuous focusing and scenery mobile tracking focusing etc., herein no longer
It repeats.
Then, according to focusing as a result, selection gesture is known in the region (the hereinafter referred to as non-region AF) other than the region AF
Other region.
In disclosure step, each non-region AF can be calculated with gesture identification region, such as positioned at viewfinder range upper left
Angle, the lower left corner, the upper right corner and the lower right corner the non-region AF area, then the area in the non-region AF and preset area are carried out
Compare, and selects the maximum non-region AF of area to show gesture identification region.
Gesture identification region, it is more than one if it is greater than the non-region AF of default gesture identification region area, then it can be
A region is randomly choosed in the qualified non-region AF to show gesture identification region;It is excellent that preset position can also be read
First grade selects the highest region of positional priority to show gesture identification region from the qualified non-region AF, such as:It should
Positional priority can for the lower right corner highest priority or the right priority be higher than left side priority, the lower right corner it is preferential
Grade is higher than the priority in the upper right corner, etc..
After determining gesture identification region, another luminaire can also be opened, so that the luminaire is known along gesture
The edge in other region, transmitting is different from the light of another color of the color of the light of directive viewfinder range, so that gesture identification area
The range in domain definitely, convenient for user carries out gesture operation in the area.
Gesture information in step 102, acquisition gesture identification region.
When user needs to carry out zoom to the scenery taken, gesture identification region is first placed a digit into, then finger is held
The preset operation of row.
In disclosure step, before acquiring gesture information, intelligent glasses can be first to the hand for entering gesture identification region
Shape identified, acquire hand shape information.
Then collected hand shape information is compared intelligent glasses with pre-stored hand shape information.
Wherein, which can also include the number of finger, can be for example, any two fingers
Shape, such as thumb and index finger, index finger and middle finger etc. can also be the shape of any one finger.
Then intelligent glasses calculate the similarity of hand shape information collected and pre-stored hand shape information, if this is similar
Degree is greater than hand shape threshold value, then is judged as that the hand shape for being currently entering gesture identification region is otherwise the shape of specified finger is judged as
The hand shape for being currently entering gesture operation is not specified finger, the then acquisition without the gesture information of next step.
For example, the hand shape information for being currently entering gesture identification region is three fingers, then intelligent glasses are judged as and do not refer to
Determine finger, current operation may be maloperation.
It should be noted that if gesture identification region is set as non-hand shape, such as it is rectangular or round, then intelligent glasses
The similarity of hand shape information collected and pre-stored hand shape information is calculated according to above-mentioned steps, and is based on similarity judgement
No is specified finger.If gesture identification region is set as hand shape, intelligent glasses calculate collected according to above-mentioned steps
The registration of hand shape information and set hand shape information, and specified finger is judged whether it is based on registration.That is, only
When the similarity or registration of finger shape collected and the finger shape of setting are more than given threshold, can just trigger next
The identification process of the gesture information of step.
Then, it when intelligent glasses are judged as that the shape of hand is the shape of specified finger, acquires in gesture identification region
The gesture information of finger.
The gesture operation can be:Using the movement that thumb and index finger open as the zoom of " amplification " as shown in Fig. 3 (A)
Instruction, and as the zoom of movement as " diminution " mediated using any two finger such as thumb and index finger of Fig. 3 (B) refers to
Show, wherein arrow indicates the direction of action of finger.Can also be:Referred to using the zoom of the movement that some finger is drawn a circle as " amplification "
Show, the movement clicked is made as the zoom of " diminution " using some finger and is indicated.
Intelligent glasses acquire the action message (also referred to as gesture information) of specified finger, analyze the motion information
Processing obtains the path curves of specified finger.
Step 103 generates zoom instructions based on gesture information.
Path curves obtained in previous step are compared intelligent glasses with preset motion profile template, such as
Fruit similarity is more than gesture threshold value, then intelligent glasses are judged as current gesture operation for zoom instruction.Specifically, if movement
The similarity of geometric locus and amplification track template is greater than amplification threshold value, it is determined that current operation is amplification instruction, if movement
Geometric locus and the similarity for reducing track template, which are greater than, reduces threshold value, it is determined that current operation is to reduce instruction.If similar
Degree is no more than threshold value, then intelligent glasses are judged as maloperation, the movement without next step
Then, intelligent glasses are based on gesture information and generate corresponding magnification ratio or diminution ratio.
In disclosure step, such as when being indicated using the expansion action of two fingers as amplification, it can refer to based on two and open
The angle between amplitude, distance or two fingers opened is become to determine magnification ratio based on the corresponding generation amplification of the magnification ratio
Coke instruction;When indicating using the kneading action of two fingers as diminution, it can refer to the amplitude reduced, distance or two fingers based on two
Between angle determine diminution ratio, and generated based on the diminution ratio and reduce zoom instructions;When drawing a circle with finger
When movement is as amplification instruction, magnification ratio can be determined based on the radius of circle, and be based on the corresponding life of the magnification ratio
At amplification zoom instructions;When being indicated using the click action of a finger as diminution, can be determined based on the quantity of click
Diminution ratio, and generated based on the diminution ratio and reduce zoom instructions.
Finally, intelligent glasses be based on diminutions zoom instructions generated or amplification zoom instructions adjusting intelligent glasses or with
The focal length for the camera that intelligent glasses are wirelessly connected, to realize the scaling processing to shown scenery.
Corresponding with the control method embodiment of aforementioned intelligent glasses, the disclosure additionally provides the control device of intelligent glasses
And its embodiment of applied intelligent glasses.
As shown in figure 4, Fig. 4 is a kind of control device block diagram of intelligent glasses shown according to an exemplary embodiment, it should
Method can be used in intelligent glasses, may include:Gesture identification area determination module 410,420 and of gesture information acquisition module
Zoom instructions generation module 430.
Wherein, gesture identification area determination module 410 is configured as in the viewfinder range of intelligent glasses, determines gesture
Identification region;
Gesture information acquisition module 420 is configured as gesture determined by acquisition gesture identification area determination module 410 and knows
Gesture information in other region;
Zoom control module 430, the gesture information for being configured as acquiring based on gesture information acquisition module 420 generate zoom
Instruction.
In the above-described embodiments, the control device of intelligent glasses determines gesture identification region first, then to gesture identification
Gesture information in region is acquired, and the gesture information based on acquisition generates zoom instructions.That is, finger is only put
The acquisition and subsequent Zoom control process of gesture information can be triggered in specific regional location, is thus substantially reduced
The probability that maloperation occurs, improves the precision of gesture Zoom control.
As shown in figure 5, Fig. 5 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram, on the basis of aforementioned embodiment illustrated in fig. 4, gesture identification area determination module 410 may include the embodiment:It finds a view
Range determination submodule 411 and gesture identification region display sub-module 412.
Wherein, viewfinder range determines submodule 411, is configured as opening the shooting function of intelligent glasses, determines model of finding a view
It encloses;
Gesture identification region display sub-module 412 is configured as determining in viewfinder range and find a view determined by submodule 411
The setting position of range shows gesture identification region based on preset shapes and sizes.
In the above-described embodiments, control device can be shown in viewfinder range according to the position, shape and size of setting
Gesture identification region.
As shown in fig. 6, Fig. 6 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram, on the basis of aforementioned embodiment illustrated in fig. 4, gesture identification area determination module 410 may include the embodiment:It finds a view
Range determination submodule 413, focusing area determine submodule 414 and gesture identification region display sub-module 415.
Viewfinder range determines submodule 413, is configured as opening the shooting function of intelligent glasses, determines viewfinder range;
Focusing area determines submodule 414, is configured as determining viewfinder range viewfinder range determined by submodule 413
The middle scenery to be shot is focused, and determines focusing area;
Gesture identification region display sub-module 415 is configured as in viewfinder range except focusing area determines submodule 414
In non-focusing area other than determining focusing area, area ratio preset area is big, gesture identification region is shown.
In the above-described embodiments, control device can be based on focusing as a result, knowing in non-focusing area determination and display gesture
Other region is conducive to user so that user when focusing by gesture, will not stop the main subject in viewfinder range
Carry out focus adjustment and shooting.
As shown in fig. 7, Fig. 7 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram, on the basis of aforementioned embodiment illustrated in fig. 6, gesture identification region display sub-module 415 may include the embodiment:Choosing
Select submodule 416.
Submodule 416 is selected, the random selection when the non-focusing area that area is greater than preset area has multiple is configured as
One of those non-focusing area shows gesture identification region.
In the above-described embodiments, can an optional area be greater than the non-focusing area of preset area and show gesture identification
Region.
As shown in figure 8, Fig. 8 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram, on the basis of aforementioned embodiment illustrated in fig. 6, gesture identification region display sub-module 415 may include the embodiment:Position
Set priority reading submodule 417 and selection submodule 418.
Wherein, positional priority reading submodule 417 is configured as being greater than the non-focusing area of preset area when area
When domain has multiple, preset positional priority is read;
Submodule 418 is selected, is configured as being greater than based on each area the non-focusing area of setting area in viewfinder range
Location selects the highest non-focusing area of positional priority to show gesture identification region.
In the above-described embodiments, positional priority is also based on to select the non-focusing area in display gesture identification region
Domain.So as to select user be most accustomed to most convenient operation region come show gesture identification region and execute gesture.
As shown in figure 9, Fig. 9 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram, on the basis of aforementioned embodiment illustrated in fig. 4, which can also include the embodiment:First light emitting module 440.
Wherein, the first light emitting module 440 is configured as emitting the light of the first color along the edge of viewfinder range.
In the above-described embodiments, the edge that viewfinder range is identified by the light of readily identified color, is conducive to user
Range based on viewfinder range adjustment intelligent glasses.
As shown in Figure 10, Figure 10 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, which can also include the embodiment:Second light emitting module 450.
Second light emitting module 450 is configured as emitting the light of the second color along the edge in the gesture identification region.
The shape in the gesture identification region that gesture identification area determination module 410 determines is rectangle, circle or hand shape.
In the above-described embodiments, pass through readily identified color and be different from the light of the color of mark viewfinder range to mark
The edge for knowing gesture identification region is conducive to user and identifies gesture identification region, executes the hand of control focal length in correct position
Gesture.Moreover, the shape in gesture identification region can carry out flexible setting in conjunction with habit, size of viewfinder range of user etc..
As shown in figure 11, Figure 10 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, which can also include the embodiment:Hand shape information acquisition module
460 and hand shape comparison module 470.
Wherein, hand shape information acquisition module 460 is configured as the hand that acquisition gesture identification area determination module 410 determines
Hand shape information in gesture identification region;
Hand shape comparison module 470, be configured as by the collected hand shape information of the institute of hand shape information acquisition module 460 with prestore
The hand shape template of storage compares, and obtains similarity;
Then gesture information acquisition module 420 acquires the gesture information in gesture identification region in 470 institute of hand shape comparison module
The similarity of comparison executes when being greater than hand shape threshold value.
In the above-described embodiments, the hand shape for entering gesture identification region is identified first, the hand of only specified hand shape
Gesture information can be only used for carrying out focus controlling, to further avoid maloperation.
As shown in figure 12, Figure 12 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, gesture information acquisition module 420 may include the embodiment:Movement letter
Breath acquisition submodule 421 and path curves generate submodule 422.
Wherein, motion information acquisition submodule 421 is configured as the movement letter of the finger in acquisition gesture identification region
Breath;
Path curves generate submodule 422, are configured as the movement acquired based on motion information acquisition submodule 421
Information generates the path curves of the finger as gesture information.
In the above-described embodiments, gesture information is motion track information, and control device passes through the motion information of acquisition finger,
Generate motion track information.
As shown in figure 13, Figure 13 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, zoom instructions generation module 430 may include the embodiment:Amplification ratio
Submodule 432 is generated to submodule 431, magnification ratio and amplification zoom instructions generate submodule 433.
Wherein, amplify comparer module 431, be configured as the movement for generating path curves generation submodule 422
Geometric locus is compared respectively with pre-stored amplification track template, obtains amplification similarity;
Magnification ratio generates submodule 432, is configured as being greater than when the amplification similarity that amplification comparer module 431 obtains
When default amplification threshold value, magnification ratio is generated based on path curves;
Amplify zoom instructions and generate submodule 433, is configured as generating the amplification that submodule 432 generates based on magnification ratio
Ratio generates amplification zoom instructions.
Wherein, it is the movement that two fingers open that path curves, which generate the path curves that submodule 422 generates,
The path curves that geometric locus or a finger are drawn a circle.
It in the above-described embodiments, can be with movement that any two fingers or specified two fingers open or a finger
The movement drawn a circle is as amplification instruction.
As shown in figure 14, Figure 14 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, zoom instructions generation module 430 may include the embodiment:Drawdown ratio
To submodule 434, reduces ratio generation submodule 435 and reduce zoom instructions generation submodule 436.
Wherein, comparer module 434 is reduced, the movement for generating path curves generation submodule 422 is configured as
Geometric locus is compared respectively with pre-stored diminution track template, obtains reducing similarity;
Diminution ratio generates submodule 435, is configured as being greater than when the diminution similarity that diminution comparer module 434 determines
When default diminution threshold value, diminution ratio is generated based on path curves;
It reduces zoom instructions and generates submodule 436, be configured as generating the diminution that submodule 335 generates based on diminution ratio
Ratio, which generates, reduces zoom instructions.
Wherein, it is the movement that two fingers are mediated that path curves, which generate the path curves that submodule 422 generates,
The path curves that geometric locus or a finger are clicked.
It in the above-described embodiments, can be with movement that any two fingers or specified two fingers are mediated or a finger
The movement of click is as amplification instruction.
As shown in figure 15, Figure 15 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of earlier figures 4-14 any illustrated embodiment, which further includes the embodiment:First Zoom control mould
Block 480.
Wherein, the first Zoom control module 480 is configured as carrying out Zoom control to intelligent glasses based on zoom instructions.
As shown in figure 16, Figure 16 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment
Block diagram is set, on the basis of earlier figures 4-14 any illustrated embodiment, which further includes the embodiment:Second Zoom control mould
Block 490.
Wherein, the second Zoom control module 490 is configured as the phase being wirelessly connected based on zoom instructions pair with intelligent glasses
Machine carries out Zoom control.
As shown in figure 17, Figure 17 is a kind of disclosure structural representation of intelligent glasses shown according to an exemplary embodiment
Figure.Intelligent glasses in the present embodiment refer mainly to the intelligent glasses with camera function, by sliding or gesture motion to intelligence
During the focal length of glasses is controlled, the movement by appearing in viewfinder range any position can all be adopted by intelligent glasses
Collection and identification, and be possible to arbitrarily acting user into the movement for being judged as adjustment focal length, so as to cause maloperation, thus it is necessary to
Gesture identification region is limited in viewfinder range, and only qualified hand shape in gesture identification region and gesture are judged.
The intelligent glasses include:The control device 520 of intelligent glasses in mirror holder 510 and above-described embodiment, wherein control device 520
It is mounted on mirror holder 510, may be mounted at any position on mirror holder 510.
Intelligent glasses in the embodiment of the present disclosure, by the control device that is arranged on mirror holder to the gesture in viewfinder range
Identification region is defined, and the gesture information for entering gesture identification region is acquired and is judged, can effectively avoid accidentally grasping
Make, and improves the accuracy of gesture Zoom control.
Those skilled in the art after considering the specification and implementing the invention disclosed here, will readily occur to its of the disclosure
Its embodiment.The disclosure is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or
Person's adaptive change follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure
Or conventional techniques.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are by following
Claim is pointed out.
The foregoing is merely the preferred embodiments of the disclosure, not to limit the disclosure, all essences in the disclosure
Within mind and principle, any modification, equivalent substitution, improvement and etc. done be should be included within the scope of disclosure protection.
Claims (31)
1. a kind of control method of intelligent glasses, which is characterized in that including
In the viewfinder range of the intelligent glasses, gesture identification region is determined;
Acquire the hand shape information in the gesture identification region;
By collected hand shape information compare with pre-stored hand shape template, obtain similarity;
When the similarity is greater than hand shape threshold value, the gesture information in the gesture identification region is acquired;
Determine the corresponding ratio of the gesture information;
Zoom instructions are generated based on the gesture information and the ratio.
2. the method according to claim 1, wherein described in the viewfinder range of the intelligent glasses, determination
Gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
In the setting position of the viewfinder range, gesture identification region is shown based on preset shapes and sizes.
3. the method according to claim 1, wherein described in the viewfinder range of the intelligent glasses, determination
Gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
It focuses to the scenery to be shot in the viewfinder range, determines focusing area;
In the viewfinder range in addition to the focusing area, in non-focusing area that area ratio preset area is big, display
Gesture identification region.
4. according to the method described in claim 3, it is characterized in that, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, one of those described non-focusing area is randomly choosed
Domain shows the gesture identification region.
5. according to the method described in claim 3, it is characterized in that, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, preset positional priority is read;
It is greater than the non-focusing area for setting area the location of in the viewfinder range based on each area, described in selection
The positional priority highest non-focusing area shows the gesture identification region.
6. the method according to claim 1, wherein the method is also before the determining gesture identification region
Including:
Luminaire is opened, so that the luminaire emits the light of the first color along the edge of the viewfinder range.
7. the method according to claim 1, wherein the method is also after the determining gesture identification region
Including:
Luminaire is opened, so that the luminaire emits the light of the second color along the edge in the gesture identification region.
8. the method according to claim 1, wherein the shape in the gesture identification region is rectangle, circle
Or hand shape.
9. the method according to claim 1, wherein the gesture letter in the acquisition gesture identification region
Breath, including:
Acquire the motion information of the finger in the gesture identification region;
The path curves of the finger as the gesture information are generated based on the motion information.
10. according to the method described in claim 9, it is characterized in that, based on the gesture information generate zoom instructions, including:
The path curves are compared respectively with pre-stored amplification track template, obtain amplification similarity;
When the amplification similarity is greater than default amplification threshold value, magnification ratio is generated based on the path curves;
Amplification zoom instructions are generated based on the magnification ratio.
11. according to the method described in claim 10, it is characterized in that, the path curves are the fortune that two fingers open
The path curves that dynamic rail trace curve or a finger are drawn a circle.
12. according to the method described in claim 9, it is characterized in that, based on the gesture information generate zoom instructions, including:
The path curves are compared respectively with pre-stored diminution track template, obtain reducing similarity;
When the diminution similarity is greater than default diminution threshold value, diminution ratio is generated based on the path curves;
It is generated based on the diminution ratio and reduces zoom instructions.
13. according to the method for claim 12, which is characterized in that the path curves are the fortune that two fingers are mediated
The path curves that dynamic rail trace curve or a finger are clicked.
14. -13 described in any item methods according to claim 1, which is characterized in that described generated based on the gesture information is become
After coke instruction, the method also includes:
Zoom control is carried out to the intelligent glasses based on the zoom instructions.
15. -13 described in any item methods according to claim 1, which is characterized in that described generated based on the gesture information is become
After coke instruction, the method also includes:
Zoom control is carried out based on the zoom instructions pair and the camera that the intelligent glasses are wirelessly connected.
16. a kind of control device of intelligent glasses, which is characterized in that including:
Gesture identification area determination module, for determining gesture identification region in the viewfinder range of the intelligent glasses;
Hand shape information acquisition module, for acquiring the hand shape information in the gesture identification region;
Hand shape comparison module, for by collected hand shape information compare with pre-stored hand shape template, obtain similarity;
Gesture information acquisition module, when the similarity for being compared in the hand shape comparison module is greater than hand shape threshold value, acquisition
Gesture information in the gesture identification region;
Ratio-dependent module, for determining the corresponding ratio of the gesture information;
Zoom instructions generation module, for generating zoom instructions based on the gesture information and the ratio.
17. device according to claim 16, which is characterized in that the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Gesture identification region display sub-module, for the setting position in the viewfinder range, based on preset shape and greatly
Small display gesture identification region.
18. device according to claim 16, which is characterized in that the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Focusing area determines submodule, for focusing to the scenery to be shot in the viewfinder range, determines focusing area
Domain;
Gesture identification region display sub-module, in the viewfinder range in addition to the focusing area, area ratio it is pre-
If showing gesture identification region in the big non-focusing area of area.
19. device according to claim 18, which is characterized in that gesture identification region display sub-module, including:
Submodule is selected, when the non-focusing area for being greater than preset area when area has multiple, is randomly choosed therein
One non-focusing area shows the gesture identification region.
20. device according to claim 18, which is characterized in that gesture identification region display sub-module, including:
Positional priority reading submodule when the non-focusing area for being greater than preset area when area has multiple, is read
Preset positional priority;
Submodule is selected, the non-focusing area for being greater than setting area based on each area is locating in the viewfinder range
Position, select the highest non-focusing area of the positional priority to show the gesture identification region.
21. device according to claim 16, which is characterized in that described device further includes:
First light emitting module, for emitting the light of the first color along the edge of the viewfinder range.
22. device according to claim 16, which is characterized in that described device further includes:
Second light emitting module emits the light of the second color for the edge along the gesture identification region.
23. device according to claim 16, which is characterized in that the gesture identification area determination module determines described
The shape in gesture identification region is rectangle, circle or hand shape.
24. device according to claim 16, which is characterized in that the gesture information acquisition module includes:
Motion information acquisition submodule, for acquiring the motion information of the finger in the gesture identification region;
Path curves generate submodule, for generating the finger as the gesture information based on the motion information
Path curves.
25. device according to claim 24, which is characterized in that the zoom instructions generation module includes:
Amplify comparer module, for the path curves to be compared respectively with pre-stored amplification track template,
Obtain amplification similarity;
Magnification ratio generates submodule, for being based on the movement rail when the amplification similarity is greater than default amplification threshold value
Trace curve generates magnification ratio;
Amplify zoom instructions and generate submodule, for generating amplification zoom instructions based on the magnification ratio.
26. device according to claim 25, which is characterized in that the path curves generate the institute that submodule generates
Stating path curves is the path curves that two fingers open or the path curves that a finger is drawn a circle.
27. device according to claim 24, which is characterized in that the zoom instructions generation module includes:
Comparer module is reduced, for the path curves to be compared respectively with pre-stored diminution track template,
It obtains reducing similarity;
Diminution ratio generates submodule, for being based on the movement rail when the diminution similarity is greater than default diminution threshold value
Trace curve generates diminution ratio;
It reduces zoom instructions and generates submodule, reduce zoom instructions for generating based on the diminution ratio.
28. device according to claim 27, which is characterized in that the path curves generate the institute that submodule generates
Stating path curves is the path curves that two fingers are mediated or the path curves that a finger is clicked.
29. the described in any item devices of 6-28 according to claim 1, which is characterized in that described device further includes:
First Zoom control module, for carrying out Zoom control to the intelligent glasses based on the zoom instructions.
30. the described in any item devices of 6-28 according to claim 1, which is characterized in that described device further includes:
Second Zoom control module, the camera for being wirelessly connected based on the zoom instructions pair with the intelligent glasses are become
Coke control.
31. a kind of intelligent glasses, which is characterized in that including mirror holder and the described in any item intelligent glasses of claim 16-30
Control device, the control device of the intelligent glasses are mounted on the mirror holder.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510219994.0A CN104793749B (en) | 2015-04-30 | 2015-04-30 | Intelligent glasses and its control method, device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510219994.0A CN104793749B (en) | 2015-04-30 | 2015-04-30 | Intelligent glasses and its control method, device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104793749A CN104793749A (en) | 2015-07-22 |
CN104793749B true CN104793749B (en) | 2018-11-30 |
Family
ID=53558618
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510219994.0A Active CN104793749B (en) | 2015-04-30 | 2015-04-30 | Intelligent glasses and its control method, device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104793749B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105807931B (en) * | 2016-03-16 | 2019-09-17 | 成都电锯互动科技有限公司 | A kind of implementation method of virtual reality |
CN105847679B (en) * | 2016-03-28 | 2020-01-31 | 联想(北京)有限公司 | Image acquisition method and electronic equipment |
CN110096926A (en) * | 2018-01-30 | 2019-08-06 | 北京亮亮视野科技有限公司 | A kind of method and intelligent glasses of scaling intelligent glasses screen |
CN109151312A (en) * | 2018-09-04 | 2019-01-04 | 广州视源电子科技股份有限公司 | Focusing method, device and video presenter |
CN111405181B (en) * | 2020-03-25 | 2022-01-28 | 维沃移动通信有限公司 | Focusing method and electronic equipment |
CN113709410A (en) * | 2020-05-21 | 2021-11-26 | 幻蝎科技(武汉)有限公司 | Method, system and equipment for enhancing human eye visual ability based on MR glasses |
CN112565602A (en) * | 2020-11-30 | 2021-03-26 | 北京地平线信息技术有限公司 | Method and apparatus for controlling image photographing apparatus, and computer-readable storage medium |
CN114007140A (en) * | 2021-10-29 | 2022-02-01 | 海信视像科技股份有限公司 | Method for controlling position of controlled role through gesture and display device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103324329A (en) * | 2012-03-23 | 2013-09-25 | 联想(北京)有限公司 | Touch control method and device |
CN103698904A (en) * | 2013-12-04 | 2014-04-02 | 全蕊 | Smart glasses and control method thereof |
CN104156138A (en) * | 2014-07-31 | 2014-11-19 | 北京智谷睿拓技术服务有限公司 | Shooting controlling method and device |
CN204203600U (en) * | 2014-09-28 | 2015-03-11 | 湖南省骁睿铁路科技发展有限公司 | A kind of capture apparatus of laser determination viewfinder range |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130088413A1 (en) * | 2011-10-05 | 2013-04-11 | Google Inc. | Method to Autofocus on Near-Eye Display |
-
2015
- 2015-04-30 CN CN201510219994.0A patent/CN104793749B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103324329A (en) * | 2012-03-23 | 2013-09-25 | 联想(北京)有限公司 | Touch control method and device |
CN103698904A (en) * | 2013-12-04 | 2014-04-02 | 全蕊 | Smart glasses and control method thereof |
CN104156138A (en) * | 2014-07-31 | 2014-11-19 | 北京智谷睿拓技术服务有限公司 | Shooting controlling method and device |
CN204203600U (en) * | 2014-09-28 | 2015-03-11 | 湖南省骁睿铁路科技发展有限公司 | A kind of capture apparatus of laser determination viewfinder range |
Also Published As
Publication number | Publication date |
---|---|
CN104793749A (en) | 2015-07-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104793749B (en) | Intelligent glasses and its control method, device | |
KR102559028B1 (en) | Method and apparatus for recognizing handwriting | |
CN105487652B (en) | Display device and its control method | |
US8199208B2 (en) | Operation input apparatus, operation input method, and computer readable medium for determining a priority between detected images | |
CN105933607A (en) | Photographing effect adjusting method of mobile terminal and mobile terminal | |
US11812132B2 (en) | Imaging device, control method therefor, and recording medium | |
CN105376487B (en) | A kind of focusing method and device | |
US20120275648A1 (en) | Imaging device and imaging method and program | |
CN103945121A (en) | Information processing method and electronic equipment | |
JP2008016289A (en) | Lighting control system | |
CN109219957A (en) | Motion picture acquisition equipment with double camera | |
KR100918094B1 (en) | Guesture recognition system having mobile video camera | |
JP2013242571A (en) | Subject detection and recognition in defocus state | |
JP2003520528A (en) | Multimodal video target acquisition and redirection system and method | |
EP2549741A1 (en) | Image capturing apparatus and image-capture control program product | |
CN202815718U (en) | Individual carried-with device | |
CN104935698A (en) | Photographing method of smart terminal, photographing device and smart phone | |
KR20160094933A (en) | Non-contact palmprint authentication method, device and mobile terminal | |
CN109076164B (en) | Method, apparatus and computer-readable storage medium for switching focus | |
CN112099639A (en) | Display attribute adjusting method and device, display equipment and storage medium | |
CN104702848B (en) | Show the method and device of framing information | |
CN104486548A (en) | Information processing method and electronic equipment | |
JP2014033265A (en) | Imaging device, imaging method, and program | |
CN110413239A (en) | Parameter adjusting method, device and storage medium is arranged in terminal | |
US10488923B1 (en) | Gaze detection, identification and control method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
EXSB | Decision made by sipo to initiate substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |