CN104793749B - Intelligent glasses and its control method, device - Google Patents

Intelligent glasses and its control method, device Download PDF

Info

Publication number
CN104793749B
CN104793749B CN201510219994.0A CN201510219994A CN104793749B CN 104793749 B CN104793749 B CN 104793749B CN 201510219994 A CN201510219994 A CN 201510219994A CN 104793749 B CN104793749 B CN 104793749B
Authority
CN
China
Prior art keywords
gesture identification
identification region
gesture
area
intelligent glasses
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510219994.0A
Other languages
Chinese (zh)
Other versions
CN104793749A (en
Inventor
唐明勇
陈涛
刘华君
刘华一君
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiaomi Inc
Original Assignee
Xiaomi Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiaomi Inc filed Critical Xiaomi Inc
Priority to CN201510219994.0A priority Critical patent/CN104793749B/en
Publication of CN104793749A publication Critical patent/CN104793749A/en
Application granted granted Critical
Publication of CN104793749B publication Critical patent/CN104793749B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • User Interface Of Digital Computer (AREA)

Abstract

The disclosure is directed to intelligent glasses and its control method, device, this method to include:In the viewfinder range of the intelligent glasses, gesture identification region is determined;Acquire the gesture information in the gesture identification region;Zoom instructions are generated based on the gesture information.Using the embodiment of the present disclosure, finger, which has only been placed on specific regional location, can trigger the acquisition and subsequent Zoom control process of gesture information, be greatly lowered the probability of maloperation appearance in this way, improve the precision of gesture Zoom control.

Description

Intelligent glasses and its control method, device
Technical field
This disclosure relates to wearable device technical field more particularly to a kind of intelligent glasses and its control method, device.
Background technique
With making rapid progress for science and technology, wearable device because its function multiplicity and what is used portable progress into people's Daily life.For example, intelligent glasses can by installing camera and circuit devcie relevant to camera on mirror holder Realize camera function.
In the related technology, user can carry out sliding touch or hand by any position in intelligent glasses viewfinder range Gesture control, come to intelligent glasses camera focal length or electronics focal length be adjusted, this mode easily causes maloperation, such as When shooting, occurring acquaintance in viewfinder range, user lifts hand greeting and is possible to by intelligent glasses erroneous judgement be control gesture, And lead to maloperation, it makes troubles for user.
Summary of the invention
Present disclose provides a kind of intelligent glasses and its control methods, device, to solve in the related technology by finding a view Any position in range touch or gesture operation is come the problem of controlling focal length, easily cause maloperation.
According to the first aspect of the embodiments of the present disclosure, a kind of control method of intelligent glasses is provided, including:
In the viewfinder range of the intelligent glasses, gesture identification region is determined;
Acquire the gesture information in the gesture identification region;
Zoom instructions are generated based on the gesture information.
Optionally, described in the viewfinder range of the intelligent glasses, determine gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
In the setting position of the viewfinder range, gesture identification region is shown based on preset shapes and sizes.
Optionally, described in the viewfinder range of the intelligent glasses, determine gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
It focuses to the scenery to be shot in the viewfinder range, determines focusing area;
In the viewfinder range in addition to the focusing area, in non-focusing area that area ratio preset area is big, Show gesture identification region.
Optionally, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, it is described non-right to randomly choose one of those Burnt region shows the gesture identification region.
Optionally, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, preset positional priority is read;
It is greater than non-focusing area location, selection in the viewfinder range of setting area based on each area The highest non-focusing area of positional priority shows the gesture identification region.
Optionally, before the determining gesture identification region, the method also includes:
Luminaire is opened, so that the luminaire emits the light of the first color along the edge of the viewfinder range.
Optionally, after the determining gesture identification region, the method also includes:
Luminaire is opened, so that the luminaire emits the second color along the edge in the gesture identification region Light.
Optionally, the shape in the gesture identification region is rectangle, circle or hand shape.
Optionally, before the gesture information in the acquisition gesture identification region, the method also includes:
Acquire the hand shape information in the gesture identification region;
By collected hand shape information compare with pre-stored hand shape template, obtain similarity;
Then the gesture information in the acquisition gesture identification region is executed when the similarity is greater than hand shape threshold value.
Optionally, the gesture information in the acquisition gesture identification region, including:
Acquire the motion information of the finger in the gesture identification region;
The path curves of the finger as the gesture information are generated based on the motion information.
Optionally, zoom instructions are generated based on the gesture information, including:
The path curves are compared respectively with pre-stored amplification track template, obtain amplification similarity;
When the amplification similarity is greater than default amplification threshold value, magnification ratio is generated based on the path curves;
Amplification zoom instructions are generated based on the magnification ratio.
Optionally, the path curves are that the path curves of two fingers openings or a finger are drawn a circle Path curves.
Optionally, zoom instructions are generated based on the gesture information, including:
The path curves are compared respectively with pre-stored diminution track template, obtain reducing similarity;
When the diminution similarity is greater than default diminution threshold value, diminution ratio is generated based on the path curves;
It is generated based on the diminution ratio and reduces zoom instructions.
Optionally, the path curves are the path curves or a finger click that two fingers are mediated Path curves.
Optionally, after the generation zoom instructions based on the gesture information, the method also includes:
Zoom control is carried out to the intelligent glasses based on the zoom instructions.
Optionally, after the generation zoom instructions based on the gesture information, the method also includes:
Zoom control is carried out based on the zoom instructions pair and the camera that the intelligent glasses are wirelessly connected.
According to the second aspect of an embodiment of the present disclosure, a kind of control device of intelligent glasses is provided, including:
Gesture identification area determination module, for determining gesture identification region in the viewfinder range of the intelligent glasses;
Gesture information acquisition module, for acquiring the gesture information in the gesture identification region;
Zoom instructions generation module, for generating zoom instructions based on the gesture information.
Optionally, the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Gesture identification region display sub-module is based on preset shape for the setting position in the viewfinder range Gesture identification region is shown with size.
Optionally, the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Focusing area determines submodule, for focusing to the scenery to be shot in the viewfinder range, determining pair Burnt region;
Gesture identification region display sub-module, in the viewfinder range in addition to the focusing area, area In the non-focusing area bigger than preset area, gesture identification region is shown.
Optionally, gesture identification region display sub-module, including:
Submodule is selected, when the non-focusing area for being greater than preset area when area has multiple, randomly chooses it In non-focusing area show the gesture identification region.
Optionally, gesture identification region display sub-module, including:
Positional priority reading submodule, when the non-focusing area for being greater than preset area when area has multiple, Read preset positional priority;
Submodule is selected, for being greater than the non-focusing area of setting area based on each area in the viewfinder range Location selects the highest non-focusing area of the positional priority to show the gesture identification region.
Optionally, described device further includes:
First light emitting module, for emitting the light of the first color along the edge of the viewfinder range.
Optionally, described device further includes:
Second light emitting module emits the light of the second color for the edge along the gesture identification region.
Optionally, the shape in the gesture identification region that the gesture identification area determination module determines be rectangle, Round or hand shape.
Optionally, described device further includes:
Hand shape information acquisition module, for acquiring the hand shape information in the gesture identification region;
Hand shape comparison module, for by collected hand shape information compare with pre-stored hand shape template, obtain phase Like degree;
Then the gesture information acquisition module acquires the gesture information in the gesture identification region and compares in the hand shape The execution when similarity that module is compared is greater than hand shape threshold value.
Optionally, the gesture information acquisition module includes:
Motion information acquisition submodule, for acquiring the motion information of the finger in the gesture identification region;
Path curves generate submodule, for being generated based on the motion information as described in the gesture information The path curves of finger.
Optionally, the zoom instructions generation module includes:
Amplify comparer module, for comparing the path curves and pre-stored amplification track template respectively It is right, obtain amplification similarity;
Magnification ratio generates submodule, for being based on the fortune when the amplification similarity is greater than default amplification threshold value Dynamic rail trace curve generates magnification ratio;
Amplify zoom instructions and generate submodule, for generating amplification zoom instructions based on the magnification ratio.
Optionally, it is that two fingers open that the path curves, which generate the path curves that submodule generates, Path curves or the path curves drawn a circle of a finger.
Optionally, the zoom instructions generation module includes:
Comparer module is reduced, for comparing the path curves and pre-stored diminution track template respectively It is right, it obtains reducing similarity;
Diminution ratio generates submodule, for being based on the fortune when the diminution similarity is greater than default diminution threshold value Dynamic rail trace curve generates diminution ratio;
It reduces zoom instructions and generates submodule, reduce zoom instructions for generating based on the diminution ratio.
Optionally, it is that two fingers are mediated that the path curves, which generate the path curves that submodule generates, Path curves or a finger click path curves.
Optionally, described device further includes:
First Zoom control module, for carrying out Zoom control to the intelligent glasses based on the zoom instructions.
Optionally, described device further includes:
Second Zoom control module, camera for being wirelessly connected based on the zoom instructions pair with the intelligent glasses into Row Zoom control.
According to the second aspect of an embodiment of the present disclosure, a kind of intelligent glasses are provided, including:Mirror holder and above-mentioned Brilliant Eyes The control device of mirror, the control device of the intelligent glasses are mounted on the mirror holder.
The technical scheme provided by this disclosed embodiment can include the following benefits:
The control device of intelligent glasses can determine gesture identification region first in the disclosure, then to gesture identification region In gesture information be acquired, and based on the gesture information of acquisition to intelligent glasses carry out Zoom control.That is, finger The acquisition and subsequent Zoom control process of gesture information can be triggered by being only placed on specific regional location, thus greatly The probability for reducing maloperation appearance greatly, improves the precision of gesture Zoom control.
Control device can show that gesture is known according to the position, shape and size of setting in viewfinder range in the disclosure Other region.Focusing is also based on as a result, determining in non-focusing area and showing gesture identification region, so that user is passing through When gesture focuses, the main subject in viewfinder range will not be stopped, be conducive to user and carry out focus adjustment and shooting.
Control device is also based on positional priority to select the non-focusing area in display gesture identification region in the disclosure Domain, so as to select user be most accustomed to most convenient operation region come show gesture identification region and execute gesture.
Control device identifies the edge of viewfinder range by the light of readily identified color in the disclosure, is conducive to user Range based on viewfinder range adjustment intelligent glasses.Pass through readily identified color and is different from identifying the color of viewfinder range Light identify the edge in gesture identification region, be conducive to user and identify gesture identification region, execute control in correct position The gesture of focal length.Moreover, the shape in gesture identification region can be flexible in conjunction with carry out such as habit, the sizes of viewfinder range of user Setting.
Control device first identifies the hand shape for entering gesture identification region in the disclosure, the hand of only specified hand shape Gesture information can be only used for carrying out focus controlling, to further avoid maloperation.
Control device can be with movement that any two fingers or specified two fingers open or a finger in the disclosure The movement drawn a circle is as amplification instruction.It can be with movement that any two fingers or specified two fingers are mediated or a finger The movement of click is as amplification instruction.It is easy to operate, it is easy to accomplish.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows the implementation for meeting the disclosure Example, and together with specification for explaining the principles of this disclosure.
Fig. 1 is a kind of disclosure control method flow chart of intelligent glasses shown according to an exemplary embodiment.
Fig. 2 (A), (B) are the two of disclosure viewfinder range shown according to an exemplary embodiment and gesture identification region Width schematic diagram.
Fig. 3 (A), (B) are two width schematic diagrames of disclosure zoom gesture shown according to an exemplary embodiment.
Fig. 4 is a kind of disclosure control device block diagram of intelligent glasses shown according to an exemplary embodiment.
Fig. 5 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 6 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 7 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 8 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Fig. 9 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 10 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 11 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 12 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 13 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 14 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 15 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 16 is the control device block diagram of the disclosure another intelligent glasses shown according to an exemplary embodiment.
Figure 17 is a kind of disclosure structural schematic diagram of intelligent glasses shown according to an exemplary embodiment.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment Described in embodiment do not represent all implementations consistent with this disclosure.On the contrary, they be only with it is such as appended The example of the consistent device and method of some aspects be described in detail in claims, the disclosure.
It is only to be not intended to be limiting the disclosure merely for for the purpose of describing particular embodiments in the term that the disclosure uses. The "an" of the singular used in disclosure and the accompanying claims book, " described " and "the" are also intended to including majority Form, unless the context clearly indicates other meaning.It is also understood that term "and/or" used herein refers to and wraps It may be combined containing one or more associated any or all of project listed.
It will be appreciated that though various information, but this may be described using term first, second, third, etc. in the disclosure A little information should not necessarily be limited by these terms.These terms are only used to for same type of information being distinguished from each other out.For example, not departing from In the case where disclosure range, the first information can also be referred to as the second information, and similarly, the second information can also be referred to as One information.Depending on context, word as used in this " if " can be construed to " ... when " or " when ... When " or " in response to determination ".
As shown in Figure 1, Fig. 1 is a kind of control method flow chart of intelligent glasses shown according to an exemplary embodiment, This method can be used in intelligent glasses, include the following steps:
Step 101, in the viewfinder range of intelligent glasses, determine gesture identification region.
In disclosure step, firstly, opening the shooting function of intelligent glasses, viewfinder range is determined.
When user presses shutter to the intelligent glasses worn, or otherwise start the shooting function of intelligent glasses When, as shown in Fig. 2 (A) and (B), intelligent glasses automatically determine viewfinder range with the program set.For example, by the practical covering of camera lens The region of the 75%-90% of range is determined as viewfinder range.The movement found a view usually by the view finder that is arranged on mirror holder Lai It realizes, which can be using common view finder in the related technology, and details are not described herein.
In the embodiment of the present disclosure, the shape in the gesture identification region can be the rectangle as shown in Fig. 2 (A), can be with It can also be the hand shape as shown in Fig. 2 (B) for circle or other shapes, to guide user that specified finger is put into hand shape region, Carry out zoom operation, it should be noted that show the entire hand shape including five fingers in Fig. 2 (B), in fact or The shape of the finger of two or other numbers.
It is also provided with laser or other luminaires in disclosure step, on mirror holder, is opening shooting function When, while laser is opened, and make laser along the edge transmitting red of viewfinder range or other readily identified face The laser of color shoots the things in viewfinder range so that viewfinder range is definitely convenient for user.
Then, gesture identification region is determined in viewfinder range.
In disclosure step, the position in gesture identification region can be preset, which can be set At multiple positions in viewfinder range, such as:Gesture identification region is arranged using the habit of left or right hand according to user Lower right corner or the upper left corner of viewfinder range etc..And the gesture identification region can be preset shapes and sizes.
In disclosure step, it is also based on the scenery setting gesture identification region in current viewfinder range.
In this case, firstly, focusing to the main scenery to be shot in viewfinder range, auto-focusing is determined The region (Auto Focus, AF).
Main scenery in viewfinder range includes people, animals and plants, building etc..The focusing mode can be in the related technology Single point focalizing, multipoint focalizing, manual focus, auto-focusing, continuous focusing and scenery mobile tracking focusing etc., herein no longer It repeats.
Then, according to focusing as a result, selection gesture is known in the region (the hereinafter referred to as non-region AF) other than the region AF Other region.
In disclosure step, each non-region AF can be calculated with gesture identification region, such as positioned at viewfinder range upper left Angle, the lower left corner, the upper right corner and the lower right corner the non-region AF area, then the area in the non-region AF and preset area are carried out Compare, and selects the maximum non-region AF of area to show gesture identification region.
Gesture identification region, it is more than one if it is greater than the non-region AF of default gesture identification region area, then it can be A region is randomly choosed in the qualified non-region AF to show gesture identification region;It is excellent that preset position can also be read First grade selects the highest region of positional priority to show gesture identification region from the qualified non-region AF, such as:It should Positional priority can for the lower right corner highest priority or the right priority be higher than left side priority, the lower right corner it is preferential Grade is higher than the priority in the upper right corner, etc..
After determining gesture identification region, another luminaire can also be opened, so that the luminaire is known along gesture The edge in other region, transmitting is different from the light of another color of the color of the light of directive viewfinder range, so that gesture identification area The range in domain definitely, convenient for user carries out gesture operation in the area.
Gesture information in step 102, acquisition gesture identification region.
When user needs to carry out zoom to the scenery taken, gesture identification region is first placed a digit into, then finger is held The preset operation of row.
In disclosure step, before acquiring gesture information, intelligent glasses can be first to the hand for entering gesture identification region Shape identified, acquire hand shape information.
Then collected hand shape information is compared intelligent glasses with pre-stored hand shape information.
Wherein, which can also include the number of finger, can be for example, any two fingers Shape, such as thumb and index finger, index finger and middle finger etc. can also be the shape of any one finger.
Then intelligent glasses calculate the similarity of hand shape information collected and pre-stored hand shape information, if this is similar Degree is greater than hand shape threshold value, then is judged as that the hand shape for being currently entering gesture identification region is otherwise the shape of specified finger is judged as The hand shape for being currently entering gesture operation is not specified finger, the then acquisition without the gesture information of next step.
For example, the hand shape information for being currently entering gesture identification region is three fingers, then intelligent glasses are judged as and do not refer to Determine finger, current operation may be maloperation.
It should be noted that if gesture identification region is set as non-hand shape, such as it is rectangular or round, then intelligent glasses The similarity of hand shape information collected and pre-stored hand shape information is calculated according to above-mentioned steps, and is based on similarity judgement No is specified finger.If gesture identification region is set as hand shape, intelligent glasses calculate collected according to above-mentioned steps The registration of hand shape information and set hand shape information, and specified finger is judged whether it is based on registration.That is, only When the similarity or registration of finger shape collected and the finger shape of setting are more than given threshold, can just trigger next The identification process of the gesture information of step.
Then, it when intelligent glasses are judged as that the shape of hand is the shape of specified finger, acquires in gesture identification region The gesture information of finger.
The gesture operation can be:Using the movement that thumb and index finger open as the zoom of " amplification " as shown in Fig. 3 (A) Instruction, and as the zoom of movement as " diminution " mediated using any two finger such as thumb and index finger of Fig. 3 (B) refers to Show, wherein arrow indicates the direction of action of finger.Can also be:Referred to using the zoom of the movement that some finger is drawn a circle as " amplification " Show, the movement clicked is made as the zoom of " diminution " using some finger and is indicated.
Intelligent glasses acquire the action message (also referred to as gesture information) of specified finger, analyze the motion information Processing obtains the path curves of specified finger.
Step 103 generates zoom instructions based on gesture information.
Path curves obtained in previous step are compared intelligent glasses with preset motion profile template, such as Fruit similarity is more than gesture threshold value, then intelligent glasses are judged as current gesture operation for zoom instruction.Specifically, if movement The similarity of geometric locus and amplification track template is greater than amplification threshold value, it is determined that current operation is amplification instruction, if movement Geometric locus and the similarity for reducing track template, which are greater than, reduces threshold value, it is determined that current operation is to reduce instruction.If similar Degree is no more than threshold value, then intelligent glasses are judged as maloperation, the movement without next step
Then, intelligent glasses are based on gesture information and generate corresponding magnification ratio or diminution ratio.
In disclosure step, such as when being indicated using the expansion action of two fingers as amplification, it can refer to based on two and open The angle between amplitude, distance or two fingers opened is become to determine magnification ratio based on the corresponding generation amplification of the magnification ratio Coke instruction;When indicating using the kneading action of two fingers as diminution, it can refer to the amplitude reduced, distance or two fingers based on two Between angle determine diminution ratio, and generated based on the diminution ratio and reduce zoom instructions;When drawing a circle with finger When movement is as amplification instruction, magnification ratio can be determined based on the radius of circle, and be based on the corresponding life of the magnification ratio At amplification zoom instructions;When being indicated using the click action of a finger as diminution, can be determined based on the quantity of click Diminution ratio, and generated based on the diminution ratio and reduce zoom instructions.
Finally, intelligent glasses be based on diminutions zoom instructions generated or amplification zoom instructions adjusting intelligent glasses or with The focal length for the camera that intelligent glasses are wirelessly connected, to realize the scaling processing to shown scenery.
Corresponding with the control method embodiment of aforementioned intelligent glasses, the disclosure additionally provides the control device of intelligent glasses And its embodiment of applied intelligent glasses.
As shown in figure 4, Fig. 4 is a kind of control device block diagram of intelligent glasses shown according to an exemplary embodiment, it should Method can be used in intelligent glasses, may include:Gesture identification area determination module 410,420 and of gesture information acquisition module Zoom instructions generation module 430.
Wherein, gesture identification area determination module 410 is configured as in the viewfinder range of intelligent glasses, determines gesture Identification region;
Gesture information acquisition module 420 is configured as gesture determined by acquisition gesture identification area determination module 410 and knows Gesture information in other region;
Zoom control module 430, the gesture information for being configured as acquiring based on gesture information acquisition module 420 generate zoom Instruction.
In the above-described embodiments, the control device of intelligent glasses determines gesture identification region first, then to gesture identification Gesture information in region is acquired, and the gesture information based on acquisition generates zoom instructions.That is, finger is only put The acquisition and subsequent Zoom control process of gesture information can be triggered in specific regional location, is thus substantially reduced The probability that maloperation occurs, improves the precision of gesture Zoom control.
As shown in figure 5, Fig. 5 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram, on the basis of aforementioned embodiment illustrated in fig. 4, gesture identification area determination module 410 may include the embodiment:It finds a view Range determination submodule 411 and gesture identification region display sub-module 412.
Wherein, viewfinder range determines submodule 411, is configured as opening the shooting function of intelligent glasses, determines model of finding a view It encloses;
Gesture identification region display sub-module 412 is configured as determining in viewfinder range and find a view determined by submodule 411 The setting position of range shows gesture identification region based on preset shapes and sizes.
In the above-described embodiments, control device can be shown in viewfinder range according to the position, shape and size of setting Gesture identification region.
As shown in fig. 6, Fig. 6 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram, on the basis of aforementioned embodiment illustrated in fig. 4, gesture identification area determination module 410 may include the embodiment:It finds a view Range determination submodule 413, focusing area determine submodule 414 and gesture identification region display sub-module 415.
Viewfinder range determines submodule 413, is configured as opening the shooting function of intelligent glasses, determines viewfinder range;
Focusing area determines submodule 414, is configured as determining viewfinder range viewfinder range determined by submodule 413 The middle scenery to be shot is focused, and determines focusing area;
Gesture identification region display sub-module 415 is configured as in viewfinder range except focusing area determines submodule 414 In non-focusing area other than determining focusing area, area ratio preset area is big, gesture identification region is shown.
In the above-described embodiments, control device can be based on focusing as a result, knowing in non-focusing area determination and display gesture Other region is conducive to user so that user when focusing by gesture, will not stop the main subject in viewfinder range Carry out focus adjustment and shooting.
As shown in fig. 7, Fig. 7 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram, on the basis of aforementioned embodiment illustrated in fig. 6, gesture identification region display sub-module 415 may include the embodiment:Choosing Select submodule 416.
Submodule 416 is selected, the random selection when the non-focusing area that area is greater than preset area has multiple is configured as One of those non-focusing area shows gesture identification region.
In the above-described embodiments, can an optional area be greater than the non-focusing area of preset area and show gesture identification Region.
As shown in figure 8, Fig. 8 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram, on the basis of aforementioned embodiment illustrated in fig. 6, gesture identification region display sub-module 415 may include the embodiment:Position Set priority reading submodule 417 and selection submodule 418.
Wherein, positional priority reading submodule 417 is configured as being greater than the non-focusing area of preset area when area When domain has multiple, preset positional priority is read;
Submodule 418 is selected, is configured as being greater than based on each area the non-focusing area of setting area in viewfinder range Location selects the highest non-focusing area of positional priority to show gesture identification region.
In the above-described embodiments, positional priority is also based on to select the non-focusing area in display gesture identification region Domain.So as to select user be most accustomed to most convenient operation region come show gesture identification region and execute gesture.
As shown in figure 9, Fig. 9 is the control device of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram, on the basis of aforementioned embodiment illustrated in fig. 4, which can also include the embodiment:First light emitting module 440.
Wherein, the first light emitting module 440 is configured as emitting the light of the first color along the edge of viewfinder range.
In the above-described embodiments, the edge that viewfinder range is identified by the light of readily identified color, is conducive to user Range based on viewfinder range adjustment intelligent glasses.
As shown in Figure 10, Figure 10 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, which can also include the embodiment:Second light emitting module 450.
Second light emitting module 450 is configured as emitting the light of the second color along the edge in the gesture identification region.
The shape in the gesture identification region that gesture identification area determination module 410 determines is rectangle, circle or hand shape.
In the above-described embodiments, pass through readily identified color and be different from the light of the color of mark viewfinder range to mark The edge for knowing gesture identification region is conducive to user and identifies gesture identification region, executes the hand of control focal length in correct position Gesture.Moreover, the shape in gesture identification region can carry out flexible setting in conjunction with habit, size of viewfinder range of user etc..
As shown in figure 11, Figure 10 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, which can also include the embodiment:Hand shape information acquisition module 460 and hand shape comparison module 470.
Wherein, hand shape information acquisition module 460 is configured as the hand that acquisition gesture identification area determination module 410 determines Hand shape information in gesture identification region;
Hand shape comparison module 470, be configured as by the collected hand shape information of the institute of hand shape information acquisition module 460 with prestore The hand shape template of storage compares, and obtains similarity;
Then gesture information acquisition module 420 acquires the gesture information in gesture identification region in 470 institute of hand shape comparison module The similarity of comparison executes when being greater than hand shape threshold value.
In the above-described embodiments, the hand shape for entering gesture identification region is identified first, the hand of only specified hand shape Gesture information can be only used for carrying out focus controlling, to further avoid maloperation.
As shown in figure 12, Figure 12 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, gesture information acquisition module 420 may include the embodiment:Movement letter Breath acquisition submodule 421 and path curves generate submodule 422.
Wherein, motion information acquisition submodule 421 is configured as the movement letter of the finger in acquisition gesture identification region Breath;
Path curves generate submodule 422, are configured as the movement acquired based on motion information acquisition submodule 421 Information generates the path curves of the finger as gesture information.
In the above-described embodiments, gesture information is motion track information, and control device passes through the motion information of acquisition finger, Generate motion track information.
As shown in figure 13, Figure 13 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, zoom instructions generation module 430 may include the embodiment:Amplification ratio Submodule 432 is generated to submodule 431, magnification ratio and amplification zoom instructions generate submodule 433.
Wherein, amplify comparer module 431, be configured as the movement for generating path curves generation submodule 422 Geometric locus is compared respectively with pre-stored amplification track template, obtains amplification similarity;
Magnification ratio generates submodule 432, is configured as being greater than when the amplification similarity that amplification comparer module 431 obtains When default amplification threshold value, magnification ratio is generated based on path curves;
Amplify zoom instructions and generate submodule 433, is configured as generating the amplification that submodule 432 generates based on magnification ratio Ratio generates amplification zoom instructions.
Wherein, it is the movement that two fingers open that path curves, which generate the path curves that submodule 422 generates, The path curves that geometric locus or a finger are drawn a circle.
It in the above-described embodiments, can be with movement that any two fingers or specified two fingers open or a finger The movement drawn a circle is as amplification instruction.
As shown in figure 14, Figure 14 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of aforementioned embodiment illustrated in fig. 4, zoom instructions generation module 430 may include the embodiment:Drawdown ratio To submodule 434, reduces ratio generation submodule 435 and reduce zoom instructions generation submodule 436.
Wherein, comparer module 434 is reduced, the movement for generating path curves generation submodule 422 is configured as Geometric locus is compared respectively with pre-stored diminution track template, obtains reducing similarity;
Diminution ratio generates submodule 435, is configured as being greater than when the diminution similarity that diminution comparer module 434 determines When default diminution threshold value, diminution ratio is generated based on path curves;
It reduces zoom instructions and generates submodule 436, be configured as generating the diminution that submodule 335 generates based on diminution ratio Ratio, which generates, reduces zoom instructions.
Wherein, it is the movement that two fingers are mediated that path curves, which generate the path curves that submodule 422 generates, The path curves that geometric locus or a finger are clicked.
It in the above-described embodiments, can be with movement that any two fingers or specified two fingers are mediated or a finger The movement of click is as amplification instruction.
As shown in figure 15, Figure 15 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of earlier figures 4-14 any illustrated embodiment, which further includes the embodiment:First Zoom control mould Block 480.
Wherein, the first Zoom control module 480 is configured as carrying out Zoom control to intelligent glasses based on zoom instructions.
As shown in figure 16, Figure 16 is the control dress of the disclosure another intelligent glasses shown according to an exemplary embodiment Block diagram is set, on the basis of earlier figures 4-14 any illustrated embodiment, which further includes the embodiment:Second Zoom control mould Block 490.
Wherein, the second Zoom control module 490 is configured as the phase being wirelessly connected based on zoom instructions pair with intelligent glasses Machine carries out Zoom control.
As shown in figure 17, Figure 17 is a kind of disclosure structural representation of intelligent glasses shown according to an exemplary embodiment Figure.Intelligent glasses in the present embodiment refer mainly to the intelligent glasses with camera function, by sliding or gesture motion to intelligence During the focal length of glasses is controlled, the movement by appearing in viewfinder range any position can all be adopted by intelligent glasses Collection and identification, and be possible to arbitrarily acting user into the movement for being judged as adjustment focal length, so as to cause maloperation, thus it is necessary to Gesture identification region is limited in viewfinder range, and only qualified hand shape in gesture identification region and gesture are judged. The intelligent glasses include:The control device 520 of intelligent glasses in mirror holder 510 and above-described embodiment, wherein control device 520 It is mounted on mirror holder 510, may be mounted at any position on mirror holder 510.
Intelligent glasses in the embodiment of the present disclosure, by the control device that is arranged on mirror holder to the gesture in viewfinder range Identification region is defined, and the gesture information for entering gesture identification region is acquired and is judged, can effectively avoid accidentally grasping Make, and improves the accuracy of gesture Zoom control.
Those skilled in the art after considering the specification and implementing the invention disclosed here, will readily occur to its of the disclosure Its embodiment.The disclosure is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or Person's adaptive change follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure Or conventional techniques.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are by following Claim is pointed out.
The foregoing is merely the preferred embodiments of the disclosure, not to limit the disclosure, all essences in the disclosure Within mind and principle, any modification, equivalent substitution, improvement and etc. done be should be included within the scope of disclosure protection.

Claims (31)

1. a kind of control method of intelligent glasses, which is characterized in that including
In the viewfinder range of the intelligent glasses, gesture identification region is determined;
Acquire the hand shape information in the gesture identification region;
By collected hand shape information compare with pre-stored hand shape template, obtain similarity;
When the similarity is greater than hand shape threshold value, the gesture information in the gesture identification region is acquired;
Determine the corresponding ratio of the gesture information;
Zoom instructions are generated based on the gesture information and the ratio.
2. the method according to claim 1, wherein described in the viewfinder range of the intelligent glasses, determination Gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
In the setting position of the viewfinder range, gesture identification region is shown based on preset shapes and sizes.
3. the method according to claim 1, wherein described in the viewfinder range of the intelligent glasses, determination Gesture identification region, including:
The shooting function for opening the intelligent glasses, determines viewfinder range;
It focuses to the scenery to be shot in the viewfinder range, determines focusing area;
In the viewfinder range in addition to the focusing area, in non-focusing area that area ratio preset area is big, display Gesture identification region.
4. according to the method described in claim 3, it is characterized in that, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, one of those described non-focusing area is randomly choosed Domain shows the gesture identification region.
5. according to the method described in claim 3, it is characterized in that, the display gesture identification region, including:
When the non-focusing area that area is greater than preset area has multiple, preset positional priority is read;
It is greater than the non-focusing area for setting area the location of in the viewfinder range based on each area, described in selection The positional priority highest non-focusing area shows the gesture identification region.
6. the method according to claim 1, wherein the method is also before the determining gesture identification region Including:
Luminaire is opened, so that the luminaire emits the light of the first color along the edge of the viewfinder range.
7. the method according to claim 1, wherein the method is also after the determining gesture identification region Including:
Luminaire is opened, so that the luminaire emits the light of the second color along the edge in the gesture identification region.
8. the method according to claim 1, wherein the shape in the gesture identification region is rectangle, circle Or hand shape.
9. the method according to claim 1, wherein the gesture letter in the acquisition gesture identification region Breath, including:
Acquire the motion information of the finger in the gesture identification region;
The path curves of the finger as the gesture information are generated based on the motion information.
10. according to the method described in claim 9, it is characterized in that, based on the gesture information generate zoom instructions, including:
The path curves are compared respectively with pre-stored amplification track template, obtain amplification similarity;
When the amplification similarity is greater than default amplification threshold value, magnification ratio is generated based on the path curves;
Amplification zoom instructions are generated based on the magnification ratio.
11. according to the method described in claim 10, it is characterized in that, the path curves are the fortune that two fingers open The path curves that dynamic rail trace curve or a finger are drawn a circle.
12. according to the method described in claim 9, it is characterized in that, based on the gesture information generate zoom instructions, including:
The path curves are compared respectively with pre-stored diminution track template, obtain reducing similarity;
When the diminution similarity is greater than default diminution threshold value, diminution ratio is generated based on the path curves;
It is generated based on the diminution ratio and reduces zoom instructions.
13. according to the method for claim 12, which is characterized in that the path curves are the fortune that two fingers are mediated The path curves that dynamic rail trace curve or a finger are clicked.
14. -13 described in any item methods according to claim 1, which is characterized in that described generated based on the gesture information is become After coke instruction, the method also includes:
Zoom control is carried out to the intelligent glasses based on the zoom instructions.
15. -13 described in any item methods according to claim 1, which is characterized in that described generated based on the gesture information is become After coke instruction, the method also includes:
Zoom control is carried out based on the zoom instructions pair and the camera that the intelligent glasses are wirelessly connected.
16. a kind of control device of intelligent glasses, which is characterized in that including:
Gesture identification area determination module, for determining gesture identification region in the viewfinder range of the intelligent glasses;
Hand shape information acquisition module, for acquiring the hand shape information in the gesture identification region;
Hand shape comparison module, for by collected hand shape information compare with pre-stored hand shape template, obtain similarity;
Gesture information acquisition module, when the similarity for being compared in the hand shape comparison module is greater than hand shape threshold value, acquisition Gesture information in the gesture identification region;
Ratio-dependent module, for determining the corresponding ratio of the gesture information;
Zoom instructions generation module, for generating zoom instructions based on the gesture information and the ratio.
17. device according to claim 16, which is characterized in that the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Gesture identification region display sub-module, for the setting position in the viewfinder range, based on preset shape and greatly Small display gesture identification region.
18. device according to claim 16, which is characterized in that the gesture identification area determination module includes:
Viewfinder range determines submodule, for opening the shooting function of the intelligent glasses, determines viewfinder range;
Focusing area determines submodule, for focusing to the scenery to be shot in the viewfinder range, determines focusing area Domain;
Gesture identification region display sub-module, in the viewfinder range in addition to the focusing area, area ratio it is pre- If showing gesture identification region in the big non-focusing area of area.
19. device according to claim 18, which is characterized in that gesture identification region display sub-module, including:
Submodule is selected, when the non-focusing area for being greater than preset area when area has multiple, is randomly choosed therein One non-focusing area shows the gesture identification region.
20. device according to claim 18, which is characterized in that gesture identification region display sub-module, including:
Positional priority reading submodule when the non-focusing area for being greater than preset area when area has multiple, is read Preset positional priority;
Submodule is selected, the non-focusing area for being greater than setting area based on each area is locating in the viewfinder range Position, select the highest non-focusing area of the positional priority to show the gesture identification region.
21. device according to claim 16, which is characterized in that described device further includes:
First light emitting module, for emitting the light of the first color along the edge of the viewfinder range.
22. device according to claim 16, which is characterized in that described device further includes:
Second light emitting module emits the light of the second color for the edge along the gesture identification region.
23. device according to claim 16, which is characterized in that the gesture identification area determination module determines described The shape in gesture identification region is rectangle, circle or hand shape.
24. device according to claim 16, which is characterized in that the gesture information acquisition module includes:
Motion information acquisition submodule, for acquiring the motion information of the finger in the gesture identification region;
Path curves generate submodule, for generating the finger as the gesture information based on the motion information Path curves.
25. device according to claim 24, which is characterized in that the zoom instructions generation module includes:
Amplify comparer module, for the path curves to be compared respectively with pre-stored amplification track template, Obtain amplification similarity;
Magnification ratio generates submodule, for being based on the movement rail when the amplification similarity is greater than default amplification threshold value Trace curve generates magnification ratio;
Amplify zoom instructions and generate submodule, for generating amplification zoom instructions based on the magnification ratio.
26. device according to claim 25, which is characterized in that the path curves generate the institute that submodule generates Stating path curves is the path curves that two fingers open or the path curves that a finger is drawn a circle.
27. device according to claim 24, which is characterized in that the zoom instructions generation module includes:
Comparer module is reduced, for the path curves to be compared respectively with pre-stored diminution track template, It obtains reducing similarity;
Diminution ratio generates submodule, for being based on the movement rail when the diminution similarity is greater than default diminution threshold value Trace curve generates diminution ratio;
It reduces zoom instructions and generates submodule, reduce zoom instructions for generating based on the diminution ratio.
28. device according to claim 27, which is characterized in that the path curves generate the institute that submodule generates Stating path curves is the path curves that two fingers are mediated or the path curves that a finger is clicked.
29. the described in any item devices of 6-28 according to claim 1, which is characterized in that described device further includes:
First Zoom control module, for carrying out Zoom control to the intelligent glasses based on the zoom instructions.
30. the described in any item devices of 6-28 according to claim 1, which is characterized in that described device further includes:
Second Zoom control module, the camera for being wirelessly connected based on the zoom instructions pair with the intelligent glasses are become Coke control.
31. a kind of intelligent glasses, which is characterized in that including mirror holder and the described in any item intelligent glasses of claim 16-30 Control device, the control device of the intelligent glasses are mounted on the mirror holder.
CN201510219994.0A 2015-04-30 2015-04-30 Intelligent glasses and its control method, device Active CN104793749B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510219994.0A CN104793749B (en) 2015-04-30 2015-04-30 Intelligent glasses and its control method, device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510219994.0A CN104793749B (en) 2015-04-30 2015-04-30 Intelligent glasses and its control method, device

Publications (2)

Publication Number Publication Date
CN104793749A CN104793749A (en) 2015-07-22
CN104793749B true CN104793749B (en) 2018-11-30

Family

ID=53558618

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510219994.0A Active CN104793749B (en) 2015-04-30 2015-04-30 Intelligent glasses and its control method, device

Country Status (1)

Country Link
CN (1) CN104793749B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105807931B (en) * 2016-03-16 2019-09-17 成都电锯互动科技有限公司 A kind of implementation method of virtual reality
CN105847679B (en) * 2016-03-28 2020-01-31 联想(北京)有限公司 Image acquisition method and electronic equipment
CN110096926A (en) * 2018-01-30 2019-08-06 北京亮亮视野科技有限公司 A kind of method and intelligent glasses of scaling intelligent glasses screen
CN109151312A (en) * 2018-09-04 2019-01-04 广州视源电子科技股份有限公司 Focusing method, device and video presenter
CN111405181B (en) * 2020-03-25 2022-01-28 维沃移动通信有限公司 Focusing method and electronic equipment
CN113709410A (en) * 2020-05-21 2021-11-26 幻蝎科技(武汉)有限公司 Method, system and equipment for enhancing human eye visual ability based on MR glasses
CN112565602A (en) * 2020-11-30 2021-03-26 北京地平线信息技术有限公司 Method and apparatus for controlling image photographing apparatus, and computer-readable storage medium
CN114007140A (en) * 2021-10-29 2022-02-01 海信视像科技股份有限公司 Method for controlling position of controlled role through gesture and display device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103324329A (en) * 2012-03-23 2013-09-25 联想(北京)有限公司 Touch control method and device
CN103698904A (en) * 2013-12-04 2014-04-02 全蕊 Smart glasses and control method thereof
CN104156138A (en) * 2014-07-31 2014-11-19 北京智谷睿拓技术服务有限公司 Shooting controlling method and device
CN204203600U (en) * 2014-09-28 2015-03-11 湖南省骁睿铁路科技发展有限公司 A kind of capture apparatus of laser determination viewfinder range

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130088413A1 (en) * 2011-10-05 2013-04-11 Google Inc. Method to Autofocus on Near-Eye Display

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103324329A (en) * 2012-03-23 2013-09-25 联想(北京)有限公司 Touch control method and device
CN103698904A (en) * 2013-12-04 2014-04-02 全蕊 Smart glasses and control method thereof
CN104156138A (en) * 2014-07-31 2014-11-19 北京智谷睿拓技术服务有限公司 Shooting controlling method and device
CN204203600U (en) * 2014-09-28 2015-03-11 湖南省骁睿铁路科技发展有限公司 A kind of capture apparatus of laser determination viewfinder range

Also Published As

Publication number Publication date
CN104793749A (en) 2015-07-22

Similar Documents

Publication Publication Date Title
CN104793749B (en) Intelligent glasses and its control method, device
KR102559028B1 (en) Method and apparatus for recognizing handwriting
CN105487652B (en) Display device and its control method
US8199208B2 (en) Operation input apparatus, operation input method, and computer readable medium for determining a priority between detected images
CN105933607A (en) Photographing effect adjusting method of mobile terminal and mobile terminal
US11812132B2 (en) Imaging device, control method therefor, and recording medium
CN105376487B (en) A kind of focusing method and device
US20120275648A1 (en) Imaging device and imaging method and program
CN103945121A (en) Information processing method and electronic equipment
JP2008016289A (en) Lighting control system
CN109219957A (en) Motion picture acquisition equipment with double camera
KR100918094B1 (en) Guesture recognition system having mobile video camera
JP2013242571A (en) Subject detection and recognition in defocus state
JP2003520528A (en) Multimodal video target acquisition and redirection system and method
EP2549741A1 (en) Image capturing apparatus and image-capture control program product
CN202815718U (en) Individual carried-with device
CN104935698A (en) Photographing method of smart terminal, photographing device and smart phone
KR20160094933A (en) Non-contact palmprint authentication method, device and mobile terminal
CN109076164B (en) Method, apparatus and computer-readable storage medium for switching focus
CN112099639A (en) Display attribute adjusting method and device, display equipment and storage medium
CN104702848B (en) Show the method and device of framing information
CN104486548A (en) Information processing method and electronic equipment
JP2014033265A (en) Imaging device, imaging method, and program
CN110413239A (en) Parameter adjusting method, device and storage medium is arranged in terminal
US10488923B1 (en) Gaze detection, identification and control method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
EXSB Decision made by sipo to initiate substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant