WO2015172630A1 - Dispositif de prise de vue de caméra et son procédé de focalisation - Google Patents

Dispositif de prise de vue de caméra et son procédé de focalisation Download PDF

Info

Publication number
WO2015172630A1
WO2015172630A1 PCT/CN2015/077480 CN2015077480W WO2015172630A1 WO 2015172630 A1 WO2015172630 A1 WO 2015172630A1 CN 2015077480 W CN2015077480 W CN 2015077480W WO 2015172630 A1 WO2015172630 A1 WO 2015172630A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
target
stored
focus
sounds
Prior art date
Application number
PCT/CN2015/077480
Other languages
English (en)
Chinese (zh)
Inventor
孙丽
Original Assignee
努比亚技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 努比亚技术有限公司 filed Critical 努比亚技术有限公司
Publication of WO2015172630A1 publication Critical patent/WO2015172630A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G03PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
    • G03BAPPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
    • G03B13/00Viewfinders; Focusing aids for cameras; Means for focusing for cameras; Autofocus systems for cameras
    • G03B13/32Means for focusing
    • G03B13/34Power focusing
    • G03B13/36Autofocus systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/67Focus control based on electronic image sensor signals

Definitions

  • the present invention relates to the field of imaging technology, and in particular, to an imaging device and a focusing method thereof.
  • the traditional focus method is usually manual focus, where the user selects the focus and then focuses on the focus.
  • Manual focus is cumbersome and inefficient, especially when the target is constantly changing or the target is constantly moving, manual focus is not able to achieve real-time tracking.
  • the existing autofocus has not been able to achieve tracking shooting for a specific target.
  • the main object of the present invention is to provide an image pickup apparatus and a focusing method thereof, which aim to perform autofocusing by sound to achieve tracking shooting of a tracking target.
  • the present invention provides a focusing method of a camera device, including the steps of:
  • the sound is collected; whether the collected sound has a target sound matching the pre-stored sound, and if so, the target sound source that emits the target sound is focused.
  • determining whether the collected sound matches the pre-stored sound includes: if the collected sound is a sound emitted by at least two sound sources, separating the sounds of the sound sources; and extracting the sound sources The acoustic characteristics of the sound are compared with the acoustic characteristics of the pre-stored sound; if the acoustic characteristics of the sound of one of the sound sources match the acoustic characteristics of the pre-stored sound, the sound of the sound source is determined to be the target sound, the sound The source is the target sound source.
  • focusing on the target sound source that emits the target sound comprises: positioning the target sound source that emits the target sound; and focusing on the positioning direction of the target sound source.
  • the method further comprises: if there are at least two pre-stored sounds and at least two matching target sounds, focusing on the target sound source corresponding to the target sound specified by the user.
  • the method further comprises: if there is no target sound in the collected sound that matches the pre-stored sound, Focus on any sound source.
  • the step of determining whether there is a target sound matching the pre-stored sound in the collected sound further comprises: if the collected sound does not have a target sound matching the pre-stored sound, selecting the sound according to the preset rule The sound source is in focus.
  • the preset rule is the principle of proximity or the principle of maximum volume.
  • the pre-stored sound is a sound collected during the shooting process, and the collecting and storing steps are:
  • the invention also provides an image capturing device, comprising a sound collecting module, a processing module and a focusing module, a sound collecting module, configured to collect sound; and a processing module configured to determine whether the collected sound matches the pre-stored sound.
  • the target sound if yes, sends a first focus signal to the focus module; the focus module is configured to focus the target sound source that emits the target sound according to the first focus signal.
  • the processing module is configured to: if the detected sound is detected by at least two sound sources, separate the sounds of the sound sources; extract the acoustic features of the sounds of the sound sources, and respectively separately with the pre-stored sounds The acoustic characteristics are compared; if the acoustic characteristics of the sound of one of the sound sources match the acoustic characteristics of the pre-stored sound, it is determined that the sound of the sound source is the target sound, and the sound source is the target sound source.
  • the focusing module is configured to: position the target sound source that emits the target sound, and control the positioning direction of the camera to the target sound source to perform focusing.
  • the focusing module is configured to: if there are at least two pre-stored sounds and at least two matching target sounds, focus on the target sound source corresponding to the target sound specified by the user.
  • the processing module is configured to: if it is determined that there is no target sound matching the pre-stored sound in the collected sound, send a second focus signal to the focus module; the focus module is configured to: utter any sound according to the second focus signal The sound source is in focus.
  • the processing module is configured to: if it is determined that there is no target sound matching the pre-stored sound, send a second focus signal to the focus module; the focus module is configured to: after receiving the second focus signal, according to The preset rule selects the sound source to focus.
  • the preset rule is the principle of proximity or the principle of maximum volume.
  • the pre-stored sound is a sound collected during the shooting process
  • the sound collecting module is further configured to: determine a tracking target selected by the user on the shooting screen during the shooting process, and collect the sound of the tracking target
  • the processing module is further configured to : The acoustic characteristics of the sound of the tracking target are extracted and stored.
  • the invention also provides a focusing method of a photographing device, comprising:
  • the method further comprises: if there are at least two targets selected in the first shooting picture, and at least two matching target sounds are present, then focusing is performed according to the target sound source corresponding to the target sound specified by the user. .
  • the method further comprises: if there is no corresponding relationship table in the surrounding sound data, selecting the utterance sound source according to the preset rule to perform focusing.
  • the preset rule is the principle of proximity or the principle of maximum volume.
  • the focusing method of the image capturing device collects the sound, separates the sound, and matches the sound, thereby identifying the target sound and the corresponding target sound source, and automatically focusing the target sound source, thereby finally realizing the passing sound. Tracking the tracking target enables real-time tracking shooting even if the tracking target is constantly changing or moving.
  • FIG. 1 is a flow chart showing a first embodiment of a focusing method of an image pickup apparatus of the present invention
  • FIG. 3 is a flow chart showing a second embodiment of a focusing method of the image pickup apparatus of the present invention.
  • Fig. 4 is a block diagram showing the configuration of an image pickup apparatus according to an embodiment of the present invention.
  • the camera device of the present invention includes all devices having a camera function, such as a mobile phone, a tablet computer, a video camera, a surveillance camera, and the like.
  • the focusing method includes the following steps:
  • Step S101 start imaging
  • Step S102 collecting sound during shooting
  • the camera device utilizes at least two microphones, preferably a microphone array consisting of a plurality of microphones to collect sound.
  • Step S103 determining whether there is a target sound in the collected sound that matches the pre-stored sound.
  • a sound segment of a person is pre-recorded or acquired in the camera, and the sound segment is analyzed, and the acoustic features of the sound segment are extracted and stored.
  • the camera device samples the collected sound in real time or timing, and analyzes whether the collected sound has a target sound that matches the pre-stored sound. If there is a target sound, the process proceeds to step S104; if there is no target sound, the image is maintained. The current focus state.
  • the sound matching process is specifically as shown in FIG. 2, and includes the following steps:
  • Step S110 determining whether the collected sound is a sound emitted by a single sound source
  • step S120 If the sound is at least two sound sources, the process proceeds to step S120; if the sound is a single sound source, the process proceeds to step S160.
  • Step S120 separating the sound of each sound source
  • a conventional sound source separation method such as a sound source analysis method based on independent component analysis, can be used to separate the sound of each of the plurality of sound sources, which fully utilizes the source signals of the sound sources between the sound sources to be independent.
  • a linear filter whose number of dimensions is equal to the number of microphones is used according to the number of sound sources, and when the number of sound sources is smaller than the number of microphones, the source signal can be completely recovered.
  • the L1 norm minimization method can be used, which takes advantage of the fact that the probability distribution of the speech power spectrum is close to the Laplacian distribution instead of the Gaussian distribution.
  • the sound source separation is performed by converting analog sound input from at least two sound sources into a digital sound input; converting the digital sound input from the time domain to the frequency domain; generating a first solution set, and the solution set is derived from The estimated error of those sounds that are active in source 1 to N is minimal; A solution set estimates the number of active sound sources to produce an optimal separation solution set that is closest to each sound source of the received analog sound input; the optimal separation solution set is converted to the time domain.
  • the sound of each sound source can be separated.
  • Step S130 extracting acoustic characteristics of the sounds of the respective sound sources, and respectively comparing with the acoustic characteristics of the pre-stored sounds
  • LPC linear predictive cepstral
  • MFCC beautified cepstral parameters
  • the camera device extracts acoustic features from the sounds of the respective sound sources to form a sequence of feature vectors to be identified, such as Performing a matching score (also referred to as a log likelihood score, or a likelihood score, or a score) of each feature vector sequence to be recognized and a feature vector sequence formed by the acoustic features of the pre-stored sound, and making a decision;
  • a matching score also referred to as a log likelihood score, or a likelihood score, or a score
  • the type of pattern recognition method closed voiceprint identification, open voiceprint discrimination, and voiceprint confirmation
  • a rejection judgment is made, and the result is obtained.
  • Step S140 If the acoustic characteristics of the sound of one of the sound sources match the acoustic characteristics of the pre-stored sound, determine that the sound of the sound source is the target sound, and the sound source is the target sound source.
  • the matching score of the sound of one of the sound sources is the highest and exceeds the preset threshold, it is determined that the sound is the target sound, and the sound source is the target sound source.
  • Step S150 Extracting acoustic characteristics of the collected sound
  • the sequence of feature vectors to be identified formed by the acoustic features of the sound is directly extracted.
  • Step S160 determining whether the acoustic characteristics of the collected sound match the acoustic characteristics of the pre-stored sound
  • step S170 Matching the feature vector sequence to be identified with the feature vector sequence formed by the acoustic features of the pre-stored sound, and making a decision; according to the type of the voiceprint recognition method (closed voiceprint discrimination, open voiceprint identification and voiceprint confirmation) ), a rejection judgment is made when needed, and the result is obtained. If they match, the process goes to step S170; otherwise, it is determined that there is no target sound in the collected sound.
  • Step S170 determining that the collected sound is the target sound, and the sound source is the target sound source.
  • the matching score exceeds the preset threshold, it is determined that the collected sound is the target sound, and the sound source is the target sound source.
  • step S104 When the target sound is matched, the process proceeds to step S104.
  • Step S104 Focusing on the target sound source that emits the target sound
  • the target sound source that emits the target sound is positioned by using a conventional positioning method, and then the positioning direction of the camera is aligned with the target sound source, and the target sound source is focused. Therefore, with the focusing method, the tracking target can be tracked in real time by the camera.
  • the sound of at least two people is pre-stored in the camera, and the pre-stored plurality of sounds are prioritized.
  • the camera device is preset according to the preset.
  • the priority order focuses on the target sound source corresponding to the higher priority target sound. That is, the image pickup device stores sound characteristics of a plurality of tracking targets, and when a plurality of tracking targets simultaneously emit sound, the tracking target having a higher priority is aligned.
  • the user may specify a tracking target from the pre-stored tracking targets for tracking shooting.
  • the focusing method includes the following steps:
  • Step S201 start imaging
  • Step S202 collecting sound during shooting
  • Step S203 determining whether there is a target sound in the collected sound that matches the pre-stored sound.
  • step S204 If there is a target sound, the process proceeds to step S204; if there is no target sound, the process proceeds to step S205.
  • Step S204 Focusing on the target sound source that emits the target sound
  • Step S205 Focusing on any sound source
  • the unique sound source is positioned, and the camera is controlled to be in focus with the positioning direction of the sound source; when there are multiple sound sources, the sound sources are separated. And select any sound source for positioning, and control the camera to focus on the positioning direction of the sound source to focus.
  • This embodiment is especially suitable for a conference scene. When an important person in a conference speaks, the focus is on the important person; when an important person does not speak, and others speak, focus on other people.
  • the camera device can also select a focus target according to a preset rule, such as the principle of proximity, the principle of maximum volume, and the like.
  • the camera can pre-store the sound of the tracking target before starting shooting, and then track the tracking target during shooting.
  • the camera can also select a tracking target during shooting and then track the tracking target. For example, during shooting, the user selects a tracking target on the shooting screen, and the camera device flattens the tracking target on the shooting screen according to the existing conversion method.
  • the surface position is converted into a spatial position, and the sound of the tracking target is acquired, and the acoustic characteristics of the sound are extracted and stored by analysis, and then the camera can track the tracking regardless of how the tracking target moves within the imaging range.
  • the focusing method of the image pickup apparatus of the present invention recognizes the target sound and the corresponding target sound source by collecting the sound, separating the sound, and matching the sound, and automatically focusing the target sound source, thereby finally realizing tracking by the sound pair.
  • Target tracking, real-time tracking can be achieved even if the tracking target is constantly changing or moving.
  • the image pickup apparatus includes a sound collection module, a processing module, and a focus module.
  • Sound Acquisition Module Set to collect sound.
  • the sound collection module collects sound through at least two microphones, preferably through a microphone array consisting of multiple microphones.
  • Processing module configured to determine whether there is a target sound matching the pre-stored sound in the collected sound, and if yes, send a first focus signal to the focus module.
  • a sound segment of a person is pre-recorded or acquired in the camera, and the sound segment is analyzed, and the acoustic features of the sound segment are extracted and stored.
  • the processing module samples the collected sound in real time or timing, and analyzes whether the collected sound has a target sound that matches the pre-stored sound. If there is a target sound, the first focus signal is sent to the focusing module.
  • the processing module first determines whether the collected sound is a sound emitted by a single sound source.
  • a conventional sound source separation method such as a sound source analysis method based on independent component analysis, can be used to separate the sound of each of the plurality of sound sources, which fully utilizes the source signals of the sound sources between the sound sources to be independent.
  • a linear filter whose number of dimensions is equal to the number of microphones is used according to the number of sound sources, and when the number of sound sources is smaller than the number of microphones, the source signal can be completely recovered.
  • the L1 norm minimization method can be used, which takes advantage of the fact that the probability distribution of the speech power spectrum is close to the Laplacian distribution instead of the Gaussian distribution.
  • the sound source separation is performed using a method of converting analog sound input from at least two sound sources into a digital sound input; converting the digital sound input from the time domain to the frequency domain; generating a first solution set, and the solution set is derived from The estimated errors of those sounds that are active in sound sources 1 through N are minimal; the number of active sound sources is estimated from the first set of solutions, To produce an optimal separation solution set that is closest to each sound source of the received analog sound input; the optimal separation solution set is converted to the time domain.
  • the optimal separation solution set is converted to the time domain.
  • the acoustic characteristics of the sounds of the respective sound sources are extracted and compared with the acoustic characteristics of the pre-stored sounds, respectively.
  • Current methods for extracting acoustic features commonly used in sound matching methods include linear predictive cepstral (LPCC) parameters, beautified cepstral parameters (MFCC), and the like.
  • the specific processing module extracts acoustic features from the sounds of the respective sound sources to form a sequence of feature vectors to be identified, such as Performing a matching score (also referred to as a log likelihood score, or a likelihood score, or a score) of each feature vector sequence to be recognized and a feature vector sequence formed by the acoustic features of the pre-stored sound, and making a decision;
  • a matching score also referred to as a log likelihood score, or a likelihood score, or a score
  • the type of pattern recognition method closed voiceprint identification, open voiceprint discrimination, and voiceprint confirmation, when necessary, a rejection judgment is made, and the result is obtained.
  • the acoustic characteristics of the sound of one of the sound sources match the acoustic characteristics of the pre-stored sound (eg, the matching score of the sound of one of the sound sources is the highest and exceeds the preset threshold), determining that the sound of the sound source is the target sound, The sound source is the target sound source; otherwise, it is determined that there is no target sound in the collected sound.
  • the acoustic characteristics of the collected sound are directly extracted, and it is judged whether the acoustic characteristics of the collected sound match the acoustic characteristics of the pre-stored sound, and if they match, the collected sound is determined to be The target sound, the sound source is the target sound source, otherwise, it is determined that there is no target sound in the collected sound.
  • the processing module determines that there is no target sound among the collected sounds, then transmitting a second focus signal to the focus module.
  • Focus module Set to focus on the target sound source that emits the target sound according to the first focus signal.
  • the focus module uses a conventional positioning method to locate the target sound source that emits the target sound, and then controls the positioning direction of the camera to the target sound source to focus the target sound source. Therefore, with the focusing method, the tracking target can be tracked in real time by the camera.
  • the focus module is based on the preset priority.
  • the sequence focuses on the target sound source corresponding to the higher priority target sound. That is, the image pickup device stores sound characteristics of a plurality of tracking targets, and when a plurality of tracking targets simultaneously emit sound, the tracking target having a higher priority is aligned. Alternatively, the user may specify a tracking target from the pre-stored tracking targets for tracking shooting.
  • the focus module when the processing module sends the second focus signal to the focus module, the focus module focuses on any of the sound source according to the second focus signal.
  • the sound source is the only sound source, the unique sound source is positioned, and the camera is controlled to focus on the positioning direction of the sound source; when there are multiple sound sources, the processing module performs multiple sound sources.
  • the focus module selects any sound source for positioning, and controls the camera to focus on the positioning direction of the sound source to focus. This embodiment is especially suitable for a conference scene. When an important person in a conference speaks, it focuses on the important task; when an important person does not speak, and another person speaks, it focuses on other people.
  • the focus module can also select the focus target according to the preset rules, such as the principle of proximity, the principle of maximum volume, and the like.
  • the camera can pre-store the sound of the tracking target before starting shooting, and then track the tracking target during shooting.
  • the camera can also select a tracking target during shooting and then track the tracking target. For example, during shooting, the user selects a tracking target on the shooting screen, and the camera device converts the planar position of the tracking target on the shooting image into a spatial position according to an existing conversion method, and acquires the sound of the tracking target, and analyzes The acoustic characteristics of the sound are extracted and stored, and then regardless of how the tracking target moves within the imaging range, the orientation of the target sound source can be determined by the feature matching of the sound, and the camera can perform focus tracking shooting.
  • the image pickup apparatus of the present invention recognizes the target sound and the corresponding target sound source by collecting sound, separating the sound, and matching the sound, and automatically focuses the target sound source, thereby finally realizing the tracking target by the sound. Tracking shots enables real-time tracking even when the tracking target is constantly changing or moving.
  • the image pickup device and the focusing method thereof of the present invention recognize the target sound and the corresponding target sound source by collecting the sound, separating the sound, and matching the sound, and automatically focusing the target sound source, thereby finally realizing the tracking target by the sound.
  • Tracking shots enable real-time tracking even when the tracking target is constantly changing or moving.
  • the technical solution of the present invention is particularly suitable for tracking shooting scenes such as conferences and monitoring, and has industrial applicability.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Studio Devices (AREA)

Abstract

La présente invention concerne un dispositif de prise de vue de caméra et son procédé de focalisation. Le procédé de focalisation comprend les étapes consistant à : lors d'un processus de prise de vue, collecter des sons ; et évaluer si un son cible correspondant à un son prémémorisé est présent dans les sons collectés, et si c'est le cas, focaliser une source sonore cible qui émet le son cible. Une collecte de sons, une séparation des sons et une mise en correspondance des sons permettent par conséquent de reconnaître un son cible et une source de son cible correspondante, et de focaliser automatiquement la source de son cible, ce qui permet enfin la prise de vue avec suivi d'un objet suivi par un son ; et permet également une prise de vue avec suivi en temps réel même dans la mesure où l'objet suivi change continuellement ou se déplace continuellement.
PCT/CN2015/077480 2014-05-15 2015-04-27 Dispositif de prise de vue de caméra et son procédé de focalisation WO2015172630A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410205508.5 2014-05-15
CN201410205508.5A CN103957359B (zh) 2014-05-15 2014-05-15 摄像装置及其对焦方法

Publications (1)

Publication Number Publication Date
WO2015172630A1 true WO2015172630A1 (fr) 2015-11-19

Family

ID=51334574

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/077480 WO2015172630A1 (fr) 2014-05-15 2015-04-27 Dispositif de prise de vue de caméra et son procédé de focalisation

Country Status (2)

Country Link
CN (1) CN103957359B (fr)
WO (1) WO2015172630A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110876036A (zh) * 2018-08-31 2020-03-10 腾讯数码(天津)有限公司 一种视频生成的方法以及相关装置
CN111953894A (zh) * 2016-11-22 2020-11-17 谷歌有限责任公司 用于捕获图像的设备、方法、系统及计算机可读存储介质
CN113284490A (zh) * 2021-04-23 2021-08-20 歌尔股份有限公司 电子设备的控制方法、装置、设备及可读存储介质
CN113573096A (zh) * 2021-07-05 2021-10-29 维沃移动通信(杭州)有限公司 视频处理方法、装置、电子设备及介质
US11388333B2 (en) * 2017-11-30 2022-07-12 SZ DJI Technology Co., Ltd. Audio guided image capture method and device

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103957359B (zh) * 2014-05-15 2016-08-24 努比亚技术有限公司 摄像装置及其对焦方法
CN104092936B (zh) * 2014-06-12 2017-01-04 小米科技有限责任公司 自动对焦方法及装置
CN106303195A (zh) * 2015-05-28 2017-01-04 中兴通讯股份有限公司 拍摄设备及跟踪拍摄方法和系统
CN105208283A (zh) * 2015-10-13 2015-12-30 广东欧珀移动通信有限公司 一种声控拍照的方法及装置
CN105554443B (zh) * 2015-12-04 2018-11-13 浙江宇视科技有限公司 视频图像中异响来源的定位方法及装置
CN105657253B (zh) * 2015-12-28 2019-03-29 联想(北京)有限公司 一种对焦方法及电子设备
CN105872366B (zh) * 2016-03-30 2018-08-24 南昌大学 一种基于fastica算法的盲源分离技术控制聚焦系统
CN107347145A (zh) * 2016-05-06 2017-11-14 杭州萤石网络有限公司 一种视频监控方法及云台网络摄像机
CN105979442B (zh) * 2016-07-22 2019-12-03 北京地平线机器人技术研发有限公司 噪声抑制方法、装置和可移动设备
CN106341601A (zh) * 2016-09-23 2017-01-18 努比亚技术有限公司 移动终端及拍照方法
CN106341665A (zh) * 2016-09-30 2017-01-18 浙江宇视科技有限公司 一种跟踪监控方法及装置
CN106603919A (zh) * 2016-12-21 2017-04-26 捷开通讯(深圳)有限公司 调整拍摄对焦的方法及终端
CN106803886A (zh) * 2017-02-28 2017-06-06 深圳天珑无线科技有限公司 一种拍照的方法及装置
JP6766086B2 (ja) 2017-09-28 2020-10-07 キヤノン株式会社 撮像装置およびその制御方法
JP7292853B2 (ja) 2017-12-26 2023-06-19 キヤノン株式会社 撮像装置及びその制御方法及びプログラム
CN111527446B (zh) * 2017-12-26 2022-05-17 佳能株式会社 摄像设备及其控制方法和记录介质
CN108091091A (zh) * 2017-12-28 2018-05-29 中国电子科技集团公司第五十四研究所 一种低功耗震声图像复合探测系统
CN110875053A (zh) 2018-08-29 2020-03-10 阿里巴巴集团控股有限公司 语音处理的方法、装置、系统、设备和介质
CN109194918B (zh) * 2018-09-17 2022-04-19 东莞市丰展电子科技有限公司 一种基于移动载体的拍摄系统
CN109639961B (zh) * 2018-11-08 2021-05-18 联想(北京)有限公司 采集方法和电子设备
WO2020118503A1 (fr) * 2018-12-11 2020-06-18 华为技术有限公司 Procédé et appareil pour déterminer une région de mise au point d'image
CN111050063A (zh) * 2019-03-29 2020-04-21 苏州浩哥文化传播有限公司 一种基于声源识别的自动化摄像方法及其系统
CN113411487B (zh) * 2020-03-17 2023-08-01 中国电信股份有限公司 设备的控制方法、装置、系统和计算机可读存储介质
CN111783628A (zh) * 2020-06-29 2020-10-16 珠海格力电器股份有限公司 一种位置跟踪方法、装置、电子设备及存储介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20040001707A (ko) * 2002-06-28 2004-01-07 엘지전자 주식회사 화상 통신 기능을 구비한 휴대용 단말기의 촬영 방향 조정방법 및 장치
CN1713717A (zh) * 2004-06-25 2005-12-28 北京中星微电子有限公司 摄像机拍摄方位数字声控定向方法
CN101068308A (zh) * 2007-05-10 2007-11-07 华为技术有限公司 一种控制图像采集装置进行目标定位的系统及方法
CN101770139A (zh) * 2008-12-29 2010-07-07 鸿富锦精密工业(深圳)有限公司 对焦控制系统及方法
CN102413276A (zh) * 2010-09-21 2012-04-11 天津三星光电子有限公司 具有声控聚焦功能的数码摄像机
CN103957359A (zh) * 2014-05-15 2014-07-30 深圳市中兴移动通信有限公司 摄像装置及其对焦方法

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2680616A1 (fr) * 2012-06-25 2014-01-01 LG Electronics Inc. Terminal mobile et son procédé de zoom audio
CN103685905B (zh) * 2012-09-17 2016-12-28 联想(北京)有限公司 一种拍照方法及电子设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20040001707A (ko) * 2002-06-28 2004-01-07 엘지전자 주식회사 화상 통신 기능을 구비한 휴대용 단말기의 촬영 방향 조정방법 및 장치
CN1713717A (zh) * 2004-06-25 2005-12-28 北京中星微电子有限公司 摄像机拍摄方位数字声控定向方法
CN101068308A (zh) * 2007-05-10 2007-11-07 华为技术有限公司 一种控制图像采集装置进行目标定位的系统及方法
CN101770139A (zh) * 2008-12-29 2010-07-07 鸿富锦精密工业(深圳)有限公司 对焦控制系统及方法
CN102413276A (zh) * 2010-09-21 2012-04-11 天津三星光电子有限公司 具有声控聚焦功能的数码摄像机
CN103957359A (zh) * 2014-05-15 2014-07-30 深圳市中兴移动通信有限公司 摄像装置及其对焦方法

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111953894A (zh) * 2016-11-22 2020-11-17 谷歌有限责任公司 用于捕获图像的设备、方法、系统及计算机可读存储介质
US11317018B2 (en) 2016-11-22 2022-04-26 Google Llc Camera operable using natural language commands
US11388333B2 (en) * 2017-11-30 2022-07-12 SZ DJI Technology Co., Ltd. Audio guided image capture method and device
CN110876036A (zh) * 2018-08-31 2020-03-10 腾讯数码(天津)有限公司 一种视频生成的方法以及相关装置
CN110876036B (zh) * 2018-08-31 2022-08-02 腾讯数码(天津)有限公司 一种视频生成的方法以及相关装置
CN113284490A (zh) * 2021-04-23 2021-08-20 歌尔股份有限公司 电子设备的控制方法、装置、设备及可读存储介质
CN113573096A (zh) * 2021-07-05 2021-10-29 维沃移动通信(杭州)有限公司 视频处理方法、装置、电子设备及介质

Also Published As

Publication number Publication date
CN103957359A (zh) 2014-07-30
CN103957359B (zh) 2016-08-24

Similar Documents

Publication Publication Date Title
WO2015172630A1 (fr) Dispositif de prise de vue de caméra et son procédé de focalisation
US11023690B2 (en) Customized output to optimize for user preference in a distributed system
US10743107B1 (en) Synchronization of audio signals from distributed devices
US9595259B2 (en) Sound source-separating device and sound source-separating method
US9330673B2 (en) Method and apparatus for performing microphone beamforming
EP3963576B1 (fr) Génération de transcription attribuée au locuteur
CN102843540B (zh) 用于视频会议的自动摄像机选择
US11138980B2 (en) Processing overlapping speech from distributed devices
WO2019217133A1 (fr) Enrôlement d'identification vocale
US20200349953A1 (en) Audio-visual diarization to identify meeting attendees
KR102230667B1 (ko) 오디오-비주얼 데이터에 기반한 화자 분리 방법 및 장치
KR101508092B1 (ko) 화상 회의를 지원하는 방법 및 시스템
US11468895B2 (en) Distributed device meeting initiation
JP2022062874A (ja) 話者予測方法、話者予測装置、およびコミュニケーションシステム
JP2021197658A (ja) 収音装置、収音システム及び収音方法
CN116866509B (zh) 会议现场画面跟踪方法、装置和存储介质
CN113411487B (zh) 设备的控制方法、装置、系统和计算机可读存储介质
CN112788278B (zh) 视频流的生成方法、装置、设备及存储介质
CN112074901B (zh) 语音识别登入
JP2022135674A (ja) 電子機器、情報処理装置、制御方法、学習方法、プログラム
CN113691753A (zh) 一种处理方法、装置和电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15792926

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 26/04/2017)

122 Ep: pct application non-entry in european phase

Ref document number: 15792926

Country of ref document: EP

Kind code of ref document: A1