WO2019128400A1 - 控制智能眼镜拍摄视角移动的方法 - Google Patents

控制智能眼镜拍摄视角移动的方法 Download PDF

Info

Publication number
WO2019128400A1
WO2019128400A1 PCT/CN2018/110887 CN2018110887W WO2019128400A1 WO 2019128400 A1 WO2019128400 A1 WO 2019128400A1 CN 2018110887 W CN2018110887 W CN 2018110887W WO 2019128400 A1 WO2019128400 A1 WO 2019128400A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
eyeball
area
movement
smart glasses
Prior art date
Application number
PCT/CN2018/110887
Other languages
English (en)
French (fr)
Inventor
刘天一
Original Assignee
北京亮亮视野科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京亮亮视野科技有限公司 filed Critical 北京亮亮视野科技有限公司
Publication of WO2019128400A1 publication Critical patent/WO2019128400A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2628Alteration of picture size, shape, position or orientation, e.g. zooming, rotation, rolling, perspective, translation

Definitions

  • the present invention relates to the field of smart glasses, and in particular to a method for controlling the movement of a smart eyeglass to capture a viewing angle.
  • an imaging function component is started to be added to a device such as smart glasses to realize recording of contents and events seen by the eyes at any time.
  • Smart glasses refer to glasses that have an operating system like a smart phone and use a lens as a display device. Users of smart glasses can install software, games, and other software service providers to provide map navigation and video through voice or motion operations. Audio recording, video and audio calls with friends, audio and video playback, etc., and Internet access via mobile communication networks.
  • the image area of the image sensor is generally cut into an area of 20% of the entire area, and the user controls the angle of view by rotating the head.
  • this method of moving the head to control the viewing angle is not satisfactory: the head rotates at a slower speed, and does not necessarily accurately correspond to the position where the eyeball is being observed each time, resulting in the camera not capturing the desired angle of view.
  • the user needs to keep the head in a fixed position for a long time, which is easy to cause fatigue.
  • the embodiment of the invention provides a method for controlling the movement of the angle of view of the smart glasses to precisely control the movement of the shooting angle of the smart glasses.
  • the method includes:
  • the movement information of the user's eyeball is obtained, and the movement of the cropping area is controlled according to the movement information of the user's eyeball.
  • the method further includes:
  • the boundary of the cutting area is controlled not to exceed the boundary of the entire imaging area.
  • the movement information of the user's eyeball includes a moving direction, a moving speed, and a moving distance of the user's eyeball;
  • the controlling the movement of the cutting area according to the movement information of the user's eyeball comprises:
  • Controlling the moving speed of the cutting area is proportional to the moving speed of the user's eyeball
  • the moving distance of the cutting area is controlled to be proportional to the moving distance of the user's eyeball.
  • the method further includes:
  • the enlargement and reduction of the cropping area are controlled according to the tilting information of the user's eyes.
  • the embodiment of the present invention further provides a smart glasses for improving a video shooting experience.
  • the smart glasses are equipped with an eye tracker for acquiring initial position information and movement information of a user's eyeball;
  • the smart glasses are used to: cut an area for imaging in all imaging areas of the image sensor; and move the cut area to the user's visual direction according to the initial position information of the user's eyeball acquired by the eye tracker a position; controlling movement of the cropping area according to movement information of a user's eyeball acquired by the eye tracker.
  • the smart glasses are further used to:
  • the boundary of the cutting area is controlled not to exceed the boundary of the entire imaging area.
  • the movement information of the user's eyeball includes a moving direction, a moving speed, and a moving distance of the user's eyeball;
  • the controlling the movement of the cutting area according to the movement information of the eyeball of the user acquired by the eye tracker comprises:
  • Controlling the moving speed of the cutting area is proportional to the moving speed of the user's eyeball
  • the moving distance of the cutting area is controlled to be proportional to the moving distance of the user's eyeball.
  • the smart glasses are further used to:
  • the enlargement and reduction of the cropping area are controlled according to the tilting information of the user's eyes.
  • An embodiment of the present invention further provides a computer device, including a memory, a processor, and a computer program stored on the memory and operable on the processor, where the processor controls the smart glasses to capture a viewing angle when the computer program is executed Methods.
  • the embodiment of the invention further provides a computer readable storage medium storing a computer program for performing a method of controlling the movement of the smart glasses to capture a viewing angle.
  • the method for controlling the movement of the smart glasses to capture the angle of view is to cut an area for imaging in the entire imaging area of the image sensor, and obtain the initial position information of the user's eyeball, and then move the cutting area to the initial of the user.
  • the visual position is used to control the movement of the cropping area by the user's eye movement.
  • the method can make the shooting angle of view move with the movement of the user's eyeball, so that the camera accurately captures the things observed by the human eye, and obtains a more realistic shooting experience; and the user is not prone to fatigue when shooting the video with the smart glasses for a long time.
  • FIG. 1 is a schematic flow chart of a method for controlling movement of a viewing angle of a smart glasses according to an embodiment of the present invention
  • FIG. 2A is a schematic diagram of acquiring a cutting area according to an embodiment of the present invention.
  • 2B is a schematic diagram of controlling movement of a cutting area according to an embodiment of the present invention.
  • FIG. 3 is a schematic diagram of smart glasses and an eye tracker according to an embodiment of the present invention.
  • an embodiment of the present invention provides a method for controlling movement of a smart eyeglass to capture a viewing angle, which is used to precisely control a shooting angle of movement of the smart glasses.
  • the method includes:
  • 103 Obtain movement information of the user's eyeball, and control movement of the cropping area according to movement information of the user's eyeball.
  • FIG. 2A and FIG. 2B The working principle of the method for controlling the movement of the viewing angle of the smart glasses in the implementation of the present invention is as shown in FIG. 2A and FIG. 2B: a cropped area is determined in all the imaging regions of the CMOS sensor, and the cutting area is displayed. Only a part of the entire imaging area is displayed, and then the position of the user's eyeball is monitored by the eye tracker to obtain the eye movement information of the user's eyeball, and then the cutting area is moved to the user's visual direction according to the initial position information of the user's eyeball. position.
  • the eye tracker monitors the movement of the user's eyeball in real time, obtains the movement information of the user's eyeball, and finally controls the movement of the cropping area according to the movement information of the user's eyeball.
  • a center point can be determined in the cutting area, and the center point of the cutting area can be moved to the visual position of the eyeball.
  • the center point and the user of the cutting area can be adjusted.
  • the visual position coincides; if the user's eye moves to the edge of the entire imaging area of the image sensor, the center point of the crop area also moves to the edge of the entire imaging area, at which point only half of the crop area is located in the entire imaging area.
  • the imaging area is halved. Therefore, when the cutting area is moved, the boundary of the cutting area can be controlled not to exceed the boundary of the entire imaging area to ensure the imaging area of the cutting area.
  • the movement information of the user's eyeball may contain a variety of information.
  • the movement information of the user's eyeball may include the moving direction, moving speed, and moving distance of the user's eyeball; when the movement of the cropping area is controlled according to the movement information of the user's eyeball, various embodiments are possible.
  • the moving direction of the cutting area can be controlled to be consistent with the moving direction of the user's eyeball; controlling the moving speed of the cutting area is proportional to the moving speed of the user's eyeball; controlling the moving distance of the cutting area is proportional to the moving distance of the user's eyeball.
  • the moving speed of the cutting area can be controlled to be the same as the moving speed of the user's eyeball, and the moving distance of the controlled cutting area is the same as the moving distance of the user's eyeball.
  • the moving distance, the moving speed, and the moving direction of the cutting area are the same as the movement of the eye of the user, and the things observed by the human eye can be quickly captured, and the angle of view of the video shooting is exactly the same as the human eye.
  • the user often needs to adjust the focal length and perform digital zoom when shooting video, that is, the size of the cropping area needs to be adjusted.
  • the size of the cropping area needs to be adjusted.
  • the enlargement and reduction of the crop area can be controlled according to the tilt information of the user's eyes.
  • the human eye since the human eye needs to be shaken at the same time, in order to prevent misoperation, it is possible to set two or three consecutive rapid swaying as a control signal.
  • the eye is moved twice to enlarge the control signal of the cutting area. Three times of shaking is to reduce the control signal of the cropping area.
  • the left eye flapping twice may be a control signal for enlarging the cropping area
  • the right eye flapping twice is a control signal for reducing the cropping area.
  • an embodiment of the present invention further provides a smart glasses, as described in the following embodiments. Since the principle of solving the problem of the smart glasses is similar to the method of controlling the movement of the smart glasses to capture the angle of view, the implementation of the smart glasses can be referred to the implementation of the method for controlling the movement of the angle of view of the smart glasses, and the repeated description will not be repeated.
  • FIG. 3 is a smart glasses 300 according to an embodiment of the present invention.
  • the smart glasses 300 are equipped with an eye tracker 301, which is used to acquire initial position information and movement information of a user's eyeball.
  • the smart glasses 300 are configured to: cut an area for imaging in all imaging areas of the image sensor; and move the cut area to the user's visual direction according to the initial position information of the user's eyeball acquired by the eye tracker 301 Position; control the movement of the cropping area according to the movement information of the user's eyeball acquired by the eye tracker 301.
  • the smart glasses 300 are further configured to:
  • the boundary of the control crop area is controlled not to exceed the boundary of the entire image area.
  • the movement information of the user's eyeball includes a moving direction, a moving speed, and a moving distance of the user's eyeball;
  • the controlling the movement of the cropping area according to the movement information of the user's eyeball acquired by the eye tracker includes:
  • Controlling the moving speed of the cutting area is proportional to the moving speed of the user's eyeball
  • the moving distance of the control cropping area is proportional to the moving distance of the user's eyeball.
  • the smart glasses 300 are further configured to:
  • the enlargement and reduction of the crop area are controlled according to the tilt information of the user's eyes.
  • An embodiment of the present invention further provides a computer device, including a memory, a processor, and a computer program stored on the memory and operable on the processor, where the processor controls the smart glasses to capture a viewing angle when the computer program is executed Methods.
  • the embodiment of the invention further provides a computer readable storage medium storing a computer program for performing a method of controlling the movement of the smart glasses to capture a viewing angle.
  • the method for controlling the movement of the smart glasses to capture the angle of view is to cut an area for imaging in all the imaging regions of the image sensor, and obtain the initial position information of the user's eyeball, and then cut the area. Move to the user's initial visual position, and then control the movement of the cropping area by the user's eye movement.
  • the method can make the shooting angle of view move with the movement of the user's eyeball, so that the camera accurately captures the things observed by the human eye, and obtains a more realistic shooting experience; and the user is less prone to fatigue when using the smart glasses for a long time to shoot the video.
  • embodiments of the present invention can be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment, or a combination of software and hardware. Moreover, the invention can take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) including computer usable program code.
  • computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
  • the computer program instructions can also be stored in a computer readable memory that can direct a computer or other programmable data processing device to operate in a particular manner, such that the instructions stored in the computer readable memory produce an article of manufacture comprising the instruction device.
  • the apparatus implements the functions specified in one or more blocks of a flow or a flow and/or block diagram of the flowchart.
  • These computer program instructions can also be loaded onto a computer or other programmable data processing device such that a series of operational steps are performed on a computer or other programmable device to produce computer-implemented processing for execution on a computer or other programmable device.
  • the instructions provide steps for implementing the functions specified in one or more of the flow or in a block or blocks of a flow diagram.

Abstract

本发明实施例提供了一种控制智能眼镜拍摄视角移动的方法,该方法包括:在图像传感器的全部成像区域内裁切出一区域用于成像;获得用户眼球的初始位置信息,根据用户眼球的初始位置信息移动裁切区域至用户的目视位置;获得用户眼球的移动信息,根据用户眼球的移动信息控制裁切区域的移动。该方法可以使拍摄视角随着用户眼球的移动而移动,使得摄像头准确捕捉到人眼观察的事物,获得更加逼真的拍摄体验;且用户长时间使用智能眼镜拍摄视频时不易产生疲劳。

Description

控制智能眼镜拍摄视角移动的方法 技术领域
本发明涉及智能眼镜领域,特别是涉及一种控制智能眼镜拍摄视角移动的方法。
背景技术
随着科技发展,用户越来越多地希望能够通过电子设备记录生活中发生的事件,相关技术中,用户通过手机或者相机进行拍摄,但是该些传统的电子设备,体积较大,并且在使用时,需要较长的反应时间,例如在使用相机时,需要先开机,再对焦,拍摄等动作,往往一系列动作完成后,需要记录的事件,已经结束,这给用户带来较差的体验。因此,相关技术中,开始在智能眼镜等设备中加入摄像功能组件,以实现随时记录眼睛看到的内容和事件。智能眼镜是指像智能手机一样具有操作系统,并以镜片作为显示装置的眼镜,智能眼镜的用户可以安装软件、游戏等软件服务商提供的程序,可通过语音或者动作等操作实现地图导航、视频音频录制、与好友进行视频音频通话、影音播放等功能,且可以通过移动通信网络进行互联网接入。
目前在利用智能眼镜进行视频录制时,一般是将图像传感器的成像区域裁切为全部面积20%的区域,用户通过转动头部来控制拍摄的视角。但是,这种移动头部控制拍摄视角的方法并不能令人满意:头部转动的速度较慢,且不一定每次都能准确对应眼球正在观察的位置,导致摄像头拍不到想要的视角,或瞄准想拍的内容时用户需要头部长时间保持固定姿势,容易产生疲劳感。
发明内容
本发明实施例提供了一种控制智能眼镜拍摄视角移动的方法,用以精确控制智能眼镜的拍摄视角移动,该方法包括:
在图像传感器的全部成像区域内裁切出一裁切区域用于成像;
获得用户眼球的初始位置信息,根据用户眼球的初始位置信息移动所述裁切区域至用户的目视位置;
获得用户眼球的移动信息,根据用户眼球的移动信息控制所述裁切区域的移动。
优选地,所述方法,进一步包括:
调整所述裁切区域的中心点与用户的目视位置重合;
在移动所述裁切区域时,控制所述裁切区域的边界不超出所述全部成像区域的边界。
优选地,所述用户眼球的移动信息包括用户眼球的移动方向、移动速度和移动距离;
所述根据用户眼球的移动信息控制所述裁切区域的移动,包括:
控制所述裁切区域的移动方向与用户眼球的移动方向一致;
控制所述裁切区域的移动速度与用户眼球的移动速度成正比;
控制所述裁切区域的移动距离与用户眼球的移动距离成正比。
优选地,所述方法,进一步包括:
根据用户眼睛的眨动信息控制所述裁切区域的放大与缩小。
本发明实施例还提供了一种智能眼镜,用以提升视频拍摄体验,该智能眼镜搭载眼动仪,所述眼动仪用于获取用户眼球的初始位置信息及移动信息;
所述智能眼镜用于:在图像传感器的全部成像区域内裁切出一区域用于成像;根据所述眼动仪获取的用户眼球的初始位置信息,移动所述裁切区域至用户的目视位置;根据所述眼动仪获取的用户眼球的移动信息,控制所述裁切区域的移动。
优选地,所述智能眼镜进一步用于:
调整所述裁切区域的中心点与用户的目视位置重合;
在移动所述裁切区域时,控制所述裁切区域的边界不超出所述全部成像区域的边界。
优选地,所述用户眼球的移动信息包括用户眼球的移动方向、移动速度和移动距离;
所述根据所述眼动仪获取的用户眼球的移动信息,控制所述裁切区域的移动,包括:
控制所述裁切区域的移动方向与用户眼球的移动方向一致;
控制所述裁切区域的移动速度与用户眼球的移动速度成正比;
控制所述裁切区域的移动距离与用户眼球的移动距离成正比。
优选的,所述智能眼镜进一步用于:
根据用户眼睛的眨动信息控制所述裁切区域的放大与缩小。
本发明实施例还提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现控制智能眼镜拍摄视角移动的方法。
本发明实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质存储有执行控制智能眼镜拍摄视角移动的方法的计算机程序。
本发明实施例中的控制智能眼镜拍摄视角移动的方法,在图像传感器的全部成像区域内裁切出一区域进行成像,同时获取用户眼球的初始位置信息,然后将裁切区域移动至用户的初始目视位置,进而通过用户的眼球移动控制裁切区域的移动。该方法可以使拍摄视角随着用户眼球的移动而移动,使得摄像头准确捕捉到人眼观察的事物,获得更加逼真的拍摄体验;且用户长时间使用智能眼镜拍摄视频时不易产生疲劳。
附图说明
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得 其他的附图。在附图中:
图1为本发明实施例中控制智能眼镜拍摄视角移动的方法的流程示意图;
图2A为本发明实施例中获取裁切区域的示意图;
图2B为本发明实施例中控制裁切区域移动的示意图;
图3为本发明实施例中智能眼镜及眼动仪的示意图。
具体实施方式
为使本发明实施例的目的、技术方案和优点更加清楚明白,下面结合附图对本发明实施例做进一步详细说明。在此,本发明的示意性实施例及其说明用于解释本发明,但并不作为对本发明的限定。
如图1所示,本发明实施例提供了一种控制智能眼镜拍摄视角移动的方法,用以精确控制智能眼镜的拍摄视角移动,该方法包括:
101:在图像传感器的全部成像区域内裁切出一区域用于成像;
102:获得用户眼球的初始位置信息,根据用户眼球的初始位置信息移动裁切区域至用户的目视位置;
103:获得用户眼球的移动信息,根据用户眼球的移动信息控制裁切区域的移动。
本发明实施中的控制智能眼镜拍摄视角移动的方法的工作原理如图2A、图2B所示:在CMOS传感器的全部成像区域内确定一裁切区域(Cropped area),该裁切区域在显示时只显示全部成像区域内的一部分,再通过眼动仪监测用户眼球的位置,获取用户眼球的初始位置信息(Eye movement detection),进而根据用户眼球的初始位置信息移动裁切区域至用户的目视位置。眼动仪实时监测用户眼球的移动,获得用户眼球的移动信息,最终根据用户眼球的移动信息控制裁切区域的移动。
一个实施例中,移动裁切区域至用户的目视位置可以有多种实施方案。例如,为了精准地控制裁切区域,可以在裁切区域内确定一中心点,并将裁切区域的中心点移动至眼球的目视位置,实施中,可以调整裁切 区域的中心点与用户的目视位置重合;如果用户的眼球移动到图像传感器的全部成像区域的边缘时,裁切区域的中心点也会移动至全部成像区域的边缘,此时,裁切区域只有一半位于全部成像区域内,成像面积减半,因此,在移动裁切区域时,可以控制裁切区域的边界不超出全部成像区域的边界,以保证裁切区域的成像面积。
一个实施例中,用户眼球的移动信息可以包含多种信息。例如,用户眼球的移动信息可以包括用户眼球的移动方向、移动速度和移动距离;当根据用户眼球的移动信息控制裁切区域的移动时,可以有多种实施方案。例如,可以控制裁切区域的移动方向与用户眼球的移动方向一致;控制裁切区域的移动速度与用户眼球的移动速度成正比;控制裁切区域的移动距离与用户眼球的移动距离成正比。再例如,为了提升视频拍摄的实时性,可以控制裁切区域的移动速度与用户眼球的移动速度相同,控制裁切区域的移动距离与用户眼球的移动距离相同。此时,裁切区域的移动距离、移动速度、移动方向均与用户的眼球移动相同,可以迅速捕捉到人眼观察到的事物,且视频拍摄的视角与人眼完全相同。
一个实施例中,用户在拍摄视频时往往需要调整焦距,进行数码变焦,也就是需要对裁切区域的大小进行调整。调整裁切区域的大小可以有多种实施方案。例如,可以根据用户眼睛的眨动信息控制裁切区域的放大与缩小。实施中,因人眼在平时也需要进行眨动,为了防止误操作,可以设定连续两次或三次快速眨动为控制信号,例如,眼睛眨动两次为放大裁切区域的控制信号,眨动三次则为缩小裁切区域的控制信号。再例如,左眼眨动两次可以为放大裁切区域的控制信号,右眼眨动两次为缩小裁切区域的控制信号。
基于同一发明构思,本发明实施例中还提供了一种智能眼镜,如下面的实施例所述。由于该智能眼镜解决问题的原理与控制智能眼镜拍摄视角移动的方法相似,因此智能眼镜的实施可以参见控制智能眼镜拍摄视角移动的方法的实施,重复之处不再赘述。
图3为本发明实施例提供的智能眼镜300,该智能眼镜300搭载眼动 仪301,所述眼动仪301用于获取用户眼球的初始位置信息及移动信息;
所述智能眼镜300用于:在图像传感器的全部成像区域内裁切出一区域用于成像;根据所述眼动仪301获取的用户眼球的初始位置信息,移动裁切区域至用户的目视位置;根据所述眼动仪301获取的用户眼球的移动信息,控制裁切区域的移动。
一个实施例中,所述智能眼镜300进一步用于:
调整裁切区域的中心点与用户的目视位置重合;
在移动裁切区域时,控制裁切区域的边界不超出全部成像区域的边界。
一个实施例中,所述用户眼球的移动信息包括用户眼球的移动方向、移动速度和移动距离;
所述根据所述眼动仪获取的用户眼球的移动信息,控制裁切区域的移动,包括:
控制裁切区域的移动方向与用户眼球的移动方向一致;
控制裁切区域的移动速度与用户眼球的移动速度成正比;
控制裁切区域的移动距离与用户眼球的移动距离成正比。
一个实施例中,所述智能眼镜300进一步用于:
根据用户眼睛的眨动信息控制裁切区域的放大与缩小。
本发明实施例还提供了一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现控制智能眼镜拍摄视角移动的方法。
本发明实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质存储有执行控制智能眼镜拍摄视角移动的方法的计算机程序。
综上所述,本发明实施例中的控制智能眼镜拍摄视角移动的方法,在图像传感器的全部成像区域内裁切出一区域进行成像,同时获取用户眼球的初始位置信息,然后将裁切区域移动至用户的初始目视位置,进而通过用户的眼球移动控制裁切区域的移动。该方法可以使拍摄视角随着用户眼球的移动而移动,使得摄像头准确捕捉到人眼观察的事物,获 得更加逼真的拍摄体验;且用户长时间使用智能眼镜拍摄视频时不易产生疲劳。
本领域内的技术人员应明白,本发明的实施例可提供为方法、系统、或计算机程序产品。因此,本发明可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本发明可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。
本发明是参照根据本发明实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。
以上所述的具体实施例,对本发明的目的、技术方案和有益效果进行了进一步详细说明,所应理解的是,以上所述仅为本发明的具体实施例而已,并不用于限定本发明的保护范围,凡在本发明的精神和原则之 内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。

Claims (10)

  1. 一种控制智能眼镜拍摄视角移动的方法,其特征在于,包括:
    在图像传感器的全部成像区域内裁切出一裁切区域用于成像;
    获得用户眼球的初始位置信息,根据用户眼球的初始位置信息移动所述裁切区域至用户的目视位置;
    获得用户眼球的移动信息,根据用户眼球的移动信息控制所述裁切区域的移动。
  2. 如权利要求1所述的方法,其特征在于,进一步包括:
    调整所述裁切区域的中心点与用户的目视位置重合;
    在移动所述裁切区域时,控制所述裁切区域的边界不超出所述全部成像区域的边界。
  3. 如权利要求2所述的方法,其特征在于,所述用户眼球的移动信息包括用户眼球的移动方向、移动速度和移动距离;
    所述根据用户眼球的移动信息控制所述裁切区域的移动,包括:
    控制所述裁切区域的移动方向与用户眼球的移动方向一致;
    控制所述裁切区域的移动速度与用户眼球的移动速度成正比;
    控制所述裁切区域的移动距离与用户眼球的移动距离成正比。
  4. 如权利要求1至3任一项所述的方法,其特征在于,进一步包括:
    根据用户眼睛的眨动信息控制所述裁切区域的放大与缩小。
  5. 一种智能眼镜,其特征在于,所述智能眼镜搭载眼动仪,所述眼动仪用于获取用户眼球的初始位置信息及移动信息;
    所述智能眼镜用于:在图像传感器的全部成像区域内裁切出一裁切区域用于成像;根据所述眼动仪获取的用户眼球的初始位置信息,移动所述裁切区域至用户的目视位置;根据所述眼动仪获取的用户眼球的移 动信息,控制所述裁切区域的移动。
  6. 如权利要求5所述的智能眼镜,其特征在于,所述智能眼镜进一步用于:
    调整所述裁切区域的中心点与用户的目视位置重合;
    在移动所述裁切区域时,控制所述裁切区域的边界不超出所述全部成像区域的边界。
  7. 如权利要求6所述的智能眼镜,其特征在于,所述用户眼球的移动信息包括用户眼球的移动方向、移动速度和移动距离;
    所述根据所述眼动仪获取的用户眼球的移动信息,控制所述裁切区域的移动,包括:
    控制所述裁切区域的移动方向与用户眼球的移动方向一致;
    控制所述裁切区域的移动速度与用户眼球的移动速度成正比;
    控制所述裁切区域的移动距离与用户眼球的移动距离成正比。
  8. 如权利要求5至7任一项所述的智能眼镜,其特征在于,所述智能眼镜进一步用于:
    根据用户眼睛的眨动信息控制所述裁切区域的放大与缩小。
  9. 一种计算机设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现权利要求1至4任一所述方法。
  10. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质存储有执行权利要求1至4任一所述方法的计算机程序。
PCT/CN2018/110887 2017-12-29 2018-10-18 控制智能眼镜拍摄视角移动的方法 WO2019128400A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201711482557.3A CN109995986A (zh) 2017-12-29 2017-12-29 控制智能眼镜拍摄视角移动的方法
CN201711482557.3 2017-12-29

Publications (1)

Publication Number Publication Date
WO2019128400A1 true WO2019128400A1 (zh) 2019-07-04

Family

ID=67065066

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/110887 WO2019128400A1 (zh) 2017-12-29 2018-10-18 控制智能眼镜拍摄视角移动的方法

Country Status (2)

Country Link
CN (1) CN109995986A (zh)
WO (1) WO2019128400A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112399124A (zh) * 2019-08-14 2021-02-23 大唐移动通信设备有限公司 视频通讯方法及装置
CN113079315A (zh) * 2021-03-25 2021-07-06 联想(北京)有限公司 图像处理方法、装置、电子设备及计算机可读存储介质

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109190425B (zh) * 2018-07-25 2021-12-28 维沃移动通信有限公司 一种信息识别方法及终端设备
CN111147743B (zh) * 2019-12-30 2021-08-24 维沃移动通信有限公司 摄像头控制方法及电子设备

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101682694A (zh) * 2007-05-30 2010-03-24 伊斯曼柯达公司 可配置用于自主操作的照相机
CN102813574A (zh) * 2012-08-03 2012-12-12 上海交通大学 基于眼动跟踪的视觉假体图像采集装置
CN105138119A (zh) * 2015-08-04 2015-12-09 湖南七迪视觉科技有限公司 一种基于人体生物特征控制自动变焦的立体视觉系统
KR20160036166A (ko) * 2014-09-24 2016-04-04 주식회사 이랜텍 의료용 스마트 안경
US9679538B2 (en) * 2014-06-26 2017-06-13 Intel IP Corporation Eye display interface for a touch display device
CN107272224A (zh) * 2017-08-03 2017-10-20 苏州医视医疗科技有限公司 具有双向调节功能的智能眼镜
CN107306352A (zh) * 2016-04-25 2017-10-31 张翔宇 头戴式人眼偏转控制拍摄的显示、拍摄设备系统

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102547209B (zh) * 2012-02-06 2015-07-22 华为技术有限公司 视讯设备控制方法、装置及视讯系统
CN203070206U (zh) * 2012-07-31 2013-07-17 北京播思软件技术有限公司 一种人眼控制鼠标
JP5825328B2 (ja) * 2013-11-07 2015-12-02 コニカミノルタ株式会社 透過型hmdを有する情報表示システム及び表示制御プログラム
CN104090659B (zh) * 2014-07-08 2017-04-05 重庆金瓯科技发展有限责任公司 基于人眼图像和视线追踪的操作指针指示控制设备
CN104731340B (zh) * 2015-03-31 2016-08-17 努比亚技术有限公司 光标位置确定方法及终端设备
CN104754497B (zh) * 2015-04-02 2018-10-16 清华大学 一种视觉注意力驱动的通信连接建立方法
CN106708251A (zh) * 2015-08-12 2017-05-24 天津电眼科技有限公司 一种基于眼球追踪技术的智能眼镜控制方法
CN205121351U (zh) * 2015-11-24 2016-03-30 杨珊珊 一种用于控制无人机的智能眼镜及系统
CN107305430A (zh) * 2016-04-25 2017-10-31 张翔宇 人眼偏转控制显示内容变化的方案
CN107003744B (zh) * 2016-12-01 2019-05-10 深圳前海达闼云端智能科技有限公司 视点确定方法、装置和电子设备
CN106774885A (zh) * 2016-12-15 2017-05-31 上海眼控科技股份有限公司 一种车载眼动控制系统
CN107092346B (zh) * 2017-03-01 2021-12-24 联想(北京)有限公司 一种信息处理方法及电子设备

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101682694A (zh) * 2007-05-30 2010-03-24 伊斯曼柯达公司 可配置用于自主操作的照相机
CN102813574A (zh) * 2012-08-03 2012-12-12 上海交通大学 基于眼动跟踪的视觉假体图像采集装置
US9679538B2 (en) * 2014-06-26 2017-06-13 Intel IP Corporation Eye display interface for a touch display device
KR20160036166A (ko) * 2014-09-24 2016-04-04 주식회사 이랜텍 의료용 스마트 안경
CN105138119A (zh) * 2015-08-04 2015-12-09 湖南七迪视觉科技有限公司 一种基于人体生物特征控制自动变焦的立体视觉系统
CN107306352A (zh) * 2016-04-25 2017-10-31 张翔宇 头戴式人眼偏转控制拍摄的显示、拍摄设备系统
CN107272224A (zh) * 2017-08-03 2017-10-20 苏州医视医疗科技有限公司 具有双向调节功能的智能眼镜

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112399124A (zh) * 2019-08-14 2021-02-23 大唐移动通信设备有限公司 视频通讯方法及装置
CN113079315A (zh) * 2021-03-25 2021-07-06 联想(北京)有限公司 图像处理方法、装置、电子设备及计算机可读存储介质
CN113079315B (zh) * 2021-03-25 2022-04-22 联想(北京)有限公司 图像处理方法、装置、电子设备及计算机可读存储介质

Also Published As

Publication number Publication date
CN109995986A (zh) 2019-07-09

Similar Documents

Publication Publication Date Title
WO2019128400A1 (zh) 控制智能眼镜拍摄视角移动的方法
US10666856B1 (en) Gaze-directed photography via augmented reality feedback
US11488631B2 (en) Systems and methods for generating time lapse videos
US11647289B2 (en) Systems and methods for stabilizing videos
US9160931B2 (en) Modifying captured image based on user viewpoint
US11102413B2 (en) Camera area locking
JP2018067301A (ja) 拡張現実のシーンを再生する際の自動ズームのための方法、装置及びシステム
JP5230013B2 (ja) 撮像装置
US11962891B2 (en) Image preview method and apparatus, electronic device, and storage medium
WO2019104569A1 (zh) 一种对焦方法、设备及可读存储介质
US11902660B2 (en) Image processing device, image processing method, and program
JP2007251429A (ja) 動画撮影装置及びズーム調整方法
CN112995507A (zh) 提示对象位置的方法及装置
CN114422692B (zh) 视频录制方法、装置及电子设备
JP2021525043A (ja) タイムラプス撮影の制御方法および制御装置、撮像システムならびにプログラム
JP2019022026A (ja) 撮像装置
TW201536050A (zh) 可修正偏移視角的影像擷取方法、電腦程式產品暨其影像擷取裝置
JP2017063276A (ja) 映像表示装置、映像表示方法、及びプログラム
JP2015061234A (ja) 画像処理装置、画像処理方法及びプログラム
WO2022067637A1 (zh) 一种拍摄控制方法、拍摄装置及存储介质
WO2022040936A1 (zh) 变焦视频拍摄方法、拍摄系统、拍摄装置和存储介质
KR101511868B1 (ko) 다중 카메라 디바이스를 활용한 멀티미디어 촬영 방법 및 그 시스템
JP6824061B2 (ja) 撮像装置、撮像装置が実行する制御方法、プログラム、及び記憶媒体

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18894337

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 09/11/2020)

122 Ep: pct application non-entry in european phase

Ref document number: 18894337

Country of ref document: EP

Kind code of ref document: A1