CN105578097A - Video recording method and terminal - Google Patents

Video recording method and terminal Download PDF

Info

Publication number
CN105578097A
CN105578097A CN201510407614.6A CN201510407614A CN105578097A CN 105578097 A CN105578097 A CN 105578097A CN 201510407614 A CN201510407614 A CN 201510407614A CN 105578097 A CN105578097 A CN 105578097A
Authority
CN
China
Prior art keywords
distance
sound source
target
video
distance value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510407614.6A
Other languages
Chinese (zh)
Inventor
唐韬
张波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yulong Computer Telecommunication Scientific Shenzhen Co Ltd
Original Assignee
Yulong Computer Telecommunication Scientific Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yulong Computer Telecommunication Scientific Shenzhen Co Ltd filed Critical Yulong Computer Telecommunication Scientific Shenzhen Co Ltd
Priority to CN201510407614.6A priority Critical patent/CN105578097A/en
Publication of CN105578097A publication Critical patent/CN105578097A/en
Pending legal-status Critical Current

Links

Abstract

The embodiment of the invention discloses a video recording method and terminal. The method comprises the steps that a video recording starting instruction is received, and a camera is started to acquire video recording image frames; a target image photographing area in the video recording image frames is acquired; the value of distance to a target photographing object corresponding to the target image photographing area is acquired to act as first distance; a sound signal components generated by each sound source in the video recording environment in the video recording process are acquired, and the value of distance to each sound source is acquired to act as a second distance set; the sound sources corresponding to the distance values of which the difference value with the first distance is less than or equal to a preset value in the second distance set are selected to act as target sound sources; and the sound signal components corresponding to the target sound sources are saved. With application of the video recording method and terminal, sound generated by the photographing object can be saved in video recording and noise can be reduced so that quality of recorded videos can be enhanced and user experience can be enhanced.

Description

A kind of kinescope method and terminal
Technical field
The present invention relates to technical field of information processing, particularly relate to a kind of kinescope method and terminal.
Background technology
Along with the development of the communication technology, except camera, various communication terminal there are also installed camera, as smart mobile phone, palmtop PC and Intelligent worn device etc., photograph and recording function can be provided, user can be made whenever and wherever possible can to record and share the animation of oneself and beautiful scenery at one's side.
But in actual video process; except will recording sound that main reference object occurs; usually also can other sound in the environment of shooting be recorded in video; if automobile is through out-of-date blow a whistle sound and passerby's one's voice in speech; therefore can comprise the sound that a variety of user does not need to record in video, become the noise in video, user even can record a video by these noises sometimes needs the sound of record to flood; affect effect and the quality of video, have a strong impact on Consumer's Experience.
Summary of the invention
Embodiments provide a kind of kinescope method and terminal, sound that reference object sends can be preserved when recording a video and reduce noise, improve the quality of video, improve Consumer's Experience.
Embodiments provide a kind of kinescope method, described method comprises:
Receive video recording enabled instruction, start camera collection video image frame;
Obtain the target photographic images region in described video image frame;
Obtain distance value between the target reference object corresponding with described target photographic images region as the first distance;
Obtain the sound signal components that in video process, in vide environment, each sound source sends, and obtain with the distance value of described each sound source as second distance set;
Choose sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source;
Preserve the sound signal components corresponding with described target sound source.
Accordingly, embodiments provide a kind of terminal, described terminal comprises:
Camera start unit, for receiving video recording enabled instruction, starts camera collection video image frame;
Target area acquiring unit, for obtaining the target photographic images region in described video image frame;
Distance value acquiring unit, for obtaining distance value between the target reference object corresponding with described target photographic images region as the first distance;
Described distance value acquiring unit, also for obtaining the sound signal components that in video process, in vide environment, each sound source sends, and obtains with the distance value of described each sound source as second distance set;
Target sound source determining unit, for choosing sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source;
Voice signal storage unit, for preserving the sound signal components corresponding with described target sound source.
The embodiment of the present invention, when receiving video recording enabled instruction, can start camera collection video image frame, and obtaining the target photographic images region in described video image frame; Also can obtain distance value between the target reference object corresponding with described target photographic images region as the first distance; Also can obtain the sound signal components that in video process, in vide environment, each sound source sends, and obtain with the distance value of described each sound source as second distance set; And then the sound source chosen corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance is as target sound source, and preserve the sound signal components corresponding with described target sound source, sound that reference object sends can be preserved when recording a video and reduce noise, improve the quality of video, improve Consumer's Experience.
Accompanying drawing explanation
In order to be illustrated more clearly in embodiment of the present invention technical scheme, below the accompanying drawing used required in describing embodiment is briefly described, apparently, accompanying drawing in the following describes is some embodiments of the present invention, for those of ordinary skill in the art, under the prerequisite not paying creative work, other accompanying drawing can also be obtained according to these accompanying drawings.
Fig. 1 is a kind of kinescope method schematic flow sheet that the embodiment of the present invention provides;
Fig. 2 is the another kind of kinescope method schematic flow sheet that the embodiment of the present invention provides;
Fig. 3 is a kind of terminal structure schematic diagram that the embodiment of the present invention provides;
Fig. 4 is the structural representation of the target area acquiring unit that the embodiment of the present invention provides;
Fig. 5 is the structural representation of the distance value acquiring unit that the embodiment of the present invention provides;
Fig. 6 is the structural representation of the another kind of distance value acquiring unit that the embodiment of the present invention provides.
Embodiment
Below in conjunction with the accompanying drawing in the embodiment of the present invention, be clearly and completely described the technical scheme in the embodiment of the present invention, obviously, described embodiment is the present invention's part embodiment, instead of whole embodiments.Based on the embodiment in the present invention, those of ordinary skill in the art, not making the every other embodiment obtained under creative work prerequisite, belong to the scope of protection of the invention.
In specific implementation, terminal mentioned in the embodiment of the present invention includes but not limited to: the electronic equipment such as smart mobile phone (as Android phone, IOS mobile phone), panel computer, notebook computer, palmtop PC, wearable intelligent equipment.Be not construed as limiting target reference object in the embodiment of the present invention, target reference object can be any object that can produce voice signal, such as: people, animal, current etc.
A kind of kinescope method provided the embodiment of the present invention below in conjunction with Fig. 1 to Fig. 6 and terminal are specifically described.
See Fig. 1, be a kind of kinescope method schematic flow sheet that the embodiment of the present invention provides, kinescope method as shown in the figure can comprise the following steps:
S101, receives video recording enabled instruction, starts camera collection video image frame.
In specific implementation, when terminal is the video tape recorder only with recording function, when user opens the power button of terminal, can be considered that user inputs video recording enabled instruction; If when user is for using the terminal of such as smart mobile phone class to record a video, first to start the recording function of taking pictures in application in terminal, when starting this recording function, being considered as input video recording enabled instruction.This video recording enabled instruction is generally that user inputted by the input unit (physical keyboard or touch-screen) of terminal.
Camera mentioned by the present embodiment can be the camera of terminal built-in, as carried the mobile phone of camera; This step, when terminal receives video recording enabled instruction, starts camera and obtains the video image frame of camera collection.If terminal does not comprise built-in camera, but be connected with outside camera, if terminal is PC (PersonalComputer, personal computer), in PC, camera is not set, but PC be connected with outside camera; This step, when receiving video recording enabled instruction, starts the camera be connected with PC, and obtains the video image frame of camera collection.
S102, obtains the target photographic images region in described video image frame.
S103, obtains distance value between the target reference object corresponding with described target photographic images region as the first distance.
In specific implementation, display screen in terminal can show the video image frame of camera collection, user is by the target photographic images region in the touch-screen selection video image frame of terminal, also by the facial image region in face recognition technology identification vide environment as target photographic images region, the reference object in the vide environment corresponding to target photographic images region is the target reference object in the present embodiment.In the present embodiment, obtain the distance value between the target reference object corresponding with described target photographic images region, can by focusing to target photographic images region, focus according to the image distance in target photographic images region and the focal length of current camera successfully, calculate the distance value between the target reference object corresponding with target photographic images region; Also by installing distance measuring sensor in terminal, as ultrasonic range finder sensor, laser range sensor, radar range finding transducer and infrared distance measuring transducer etc., go to detect the distance value between target reference object by distance measuring sensor.
S104, obtains the sound signal components that in video process, in vide environment, each sound source sends, and obtains with the distance value of described each sound source as second distance set.
In specific implementation, start voice collection device in video process and gather the sound signal components that in vide environment, each sound source sends, this voice collection device can be specially microphone, and the sound signal components that each sound source sends comprises the sound signal components that target reference object sends.
Aliasing when usual each sound signal components arrives voice collection device, now terminal can cross employing sound isolation technics, go to obtain the corresponding sound signal components of each sound source, such as adopt the aliasing voice signal separation method based on voice fundamental frequencies, first the voice signal of aliasing can be carried out Fourier transform, signal in time domain is transformed into frequency-region signal, then on frequency domain, the dominant frequency signal occurred is separated, namely the frequency spectrum of each dominant frequency signal is just obtained respectively by band pass filter, again by be separated after Spectrum Conversion on time domain to obtain the time-domain signal of each sound signal components.
In actual implementation process, in terminal, microphone array can be installed, adopt the sound localization method of microphone array to carry out auditory localization to each sound source in vide environment.Sound localization method wherein based on microphone array comprises: the steerable beam formation technology based on peak power output, the directional technology based on High-Resolution Spectral Estimation and based on to the difference technology time of advent of voice signal.Because based on differing from the time of advent to voice signal, technique computes amount is little, precision is high, therefore this kind of method is often used in actual applications.Adopting after the sound localization method of microphone array carries out auditory localization to each sound source in vide environment, just the distance value in vide environment between each sound source and terminal can be obtained, also can obtain the orientation of each sound source relative to terminal, terminal obtains distance value between each sound source and terminal as second distance set.
S105, chooses sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source.
In specific implementation, after the distance value got in step S104 in vide environment between each sound source and terminal forms second distance set, just can calculate each distance value in second distance set and the difference between the first distance, choose wherein difference and be less than or equal to the sound source corresponding to distance value of preset value as target sound source.Namely this target sound source is considered as target reference object in the present embodiment, then preserve by step S105 the sound that the corresponding sound signal components of this target sound source sends as target reference object, the sound that in maskable vide environment, other sound sources send, noise in the vide environment gathered during to reduce video recording, make the sound of target reference object better effects if in video more clear, improve Consumer's Experience.
S106, preserves the sound signal components corresponding with described target sound source.
In another embodiment, described method also comprises: receive video recording halt instruction, exports this record a video the video image frame that collects and sound signal components corresponding to described target sound source and forms video.
In specific implementation, when recording a video, terminal detects video recording halt instruction in real time, this video recording halt instruction can timing video recording time arrive timing time terminal system automatically generate, also can be that user inputs at any time.When terminal receives video recording halt instruction, terminal exports the video image frame that collects in this video process and sound signal components corresponding to target sound source and forms video and be saved in the storage area of specifying, and carries out for user operations such as sharing, check.
The embodiment of the present invention, when receiving video recording enabled instruction, can start camera collection video image frame, and obtaining the target photographic images region in described video image frame; Also can obtain distance value between the target reference object corresponding with described target photographic images region as the first distance; Also can obtain the sound signal components that in video process, in vide environment, each sound source sends, and obtain with the distance value of described each sound source as second distance set; And then the sound source chosen corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance is as target sound source, and preserve the sound signal components corresponding with described target sound source, can the sound sent of reference object be preserved when recording a video and reduce noise, improve the quality of video, improve Consumer's Experience.
See Fig. 2, be the another kind of kinescope method schematic flow sheet that the embodiment of the present invention provides, kinescope method as shown in the figure can comprise the following steps:
S201, receives video recording enabled instruction, starts camera collection video image frame.
In specific implementation, when terminal is the video tape recorder only with recording function, when user opens the power button of terminal, can be considered that user inputs video recording enabled instruction; If when user is for using the terminal of such as smart mobile phone class to record a video, first to start the recording function of taking pictures in application in terminal, when starting this recording function, being considered as input video recording enabled instruction.This video recording enabled instruction is generally that user inputted by the input unit (physical keyboard or touch-screen) of terminal.
Camera mentioned by the present embodiment can be the camera of terminal built-in, as carried the mobile phone of camera; This step, when terminal receives video recording enabled instruction, starts camera and obtains the video image frame of camera collection.If terminal does not comprise built-in camera, but be connected with outside camera, if terminal is PC (PersonalComputer, personal computer), in PC, camera is not set, but PC be connected with outside camera; This step, when receiving video recording enabled instruction, starts the camera be connected with PC, and obtains the video image frame of camera collection.
S202, identifies the facial image region in described video image frame, and described facial image region is set to target photographic images region.
In specific implementation, usually in the video having personage, wherein personage is the target reference object of user, in video process, be therefore more concerned about whether the sound that this personage sends can preserve clearly.Be in the use scenes of personage at target reference object, by the facial image region in face recognition technology identification video image frame, this facial image region be set to target photographic images region.
S203, receives video recording focusing instruction, starts described camera and focus to described target photographic images region.
S204, according to the image distance in described target photographic images region and the focal length of current described camera, calculates distance value between the target reference object corresponding with described target photographic images region as the first distance.
In specific implementation, obtain the distance value between the target reference object corresponding with described target photographic images region, by to the focusing of target photographic images region, can focus successfully according to the distance value between the lens imaging formulae discovery target reference object corresponding with target photographic images region.Wherein lens imaging formula is: 1/f=1/u+1/v, wherein f is the real focal length of camera when a certain enlargement ratio, u is object distance (distance value namely between target reference object and camera), and v is image distance (distance value between the target photographic images regional imaging that namely target reference object is corresponding and camera).
S205, gathers the voice signal in vide environment by least three microphones in video process.
S206, is separated the corresponding sound signal components that described voice signal obtains each sound source in described vide environment.
S207, the time of the sound signal components that each sound source is corresponding according to each microphones, calculates with the distance value of described each sound source as second distance set.
In specific implementation, gather by microphone the sound signal components that in vide environment, each sound source sends in video process, and the sound signal components that each sound source sends comprises the sound signal components that target reference object sends.Aliasing when usual each sound signal components arrives microphone, now terminal can cross sound isolation technics, go to obtain the corresponding sound signal components of each sound source, such as adopt the aliasing voice signal separation method based on voice fundamental frequencies, first the voice signal of aliasing can be carried out Fourier transform and the signal in time domain is transformed into frequency-region signal, then on frequency domain, the dominant frequency signal occurred is separated, namely obtain the frequency spectrum of each dominant frequency signal respectively by band pass filter, then by be separated after Spectrum Conversion on time domain to obtain the time-domain signal of each sound signal components.
In actual implementation process, can install at least three microphones in terminal, in same plane, any point is unique to the distance of fixing three points, namely can determine another one point in plane by any 3.Because above-mentioned microphone is fixed in terminal, namely the position of above-mentioned microphone is known, positional distance between such microphone is exactly known, and the speed that voice signal is propagated also is known, the time of reception that these known conditions add each microphone more just can calculate the relative position of each sound source and terminal, such as: the relative position being calculated each sound source and terminal by plane geometry related algorithm.After auditory localization is carried out to each sound source in vide environment, just the distance value in vide environment between each sound source and terminal can be obtained, also can obtain the orientation of each sound source relative to terminal, terminal obtains distance value between each sound source and terminal as second distance set.
S208, chooses sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source.
In specific implementation, after the distance value got in step S207 in vide environment between each sound source and terminal forms second distance set, just each distance value in second distance set and the difference between the first distance can be calculated, choose wherein difference and be less than or equal to the sound source corresponding to distance value of preset value as target sound source, namely this target sound source is considered as target reference object in the present embodiment, then preserve by step S209 the sound that the corresponding sound signal components of this target sound source sends as target reference object, the sound that in maskable vide environment, other sound sources send, noise in the vide environment gathered during to reduce video recording, make the sound of target reference object better effects if in video more clear, improve Consumer's Experience.
S209, preserves the sound signal components corresponding with described target sound source.
The embodiment of the present invention, is set to target photographic images region by face recognition technology by the facial image region in video image frame, and the distance obtained by focusing between the target reference object corresponding with target photographic images region is as the first distance; Also gather the voice signal in vide environment by least three microphones, and be separated by sound isolation technics the sound signal components that in voice signal acquisition vide environment, each sound source is corresponding, and then according to the time difference between each microphones sound signal components, to each sound source position obtain with each sound source between distance value as second distance set; Also can choose sound source corresponding to the distance value being less than or equal to preset value in second distance set with the difference of the first distance as target sound source, and the sound signal components preserved corresponding to target sound source, can the sound sent of reference object be preserved when recording a video and reduce noise, improve the quality of video, improve Consumer's Experience.
See Fig. 3, a kind of terminal structure schematic diagram that the embodiment of the present invention provides, terminal at least can comprise as shown in the figure: camera start unit 301, target area acquiring unit 302, distance value acquiring unit 303, target sound source determining unit 304 and voice signal storage unit 305.
Camera start unit 301, for receiving video recording enabled instruction, starts camera collection video image frame.
In specific implementation, when terminal is the video tape recorder only with recording function, when user opens the power button of terminal, can be considered that user inputs video recording enabled instruction; If when user is for using the terminal of such as smart mobile phone class to record a video, first to start the recording function of taking pictures in application in terminal, when starting this recording function, being considered as input video recording enabled instruction.This video recording enabled instruction is generally that user inputted by the input unit (physical keyboard or touch-screen) of terminal.
Camera mentioned by the present embodiment can be the camera of terminal built-in, as carried the mobile phone of camera; This step, when terminal receives video recording enabled instruction, starts camera and obtains the video image frame of camera collection.If terminal does not comprise built-in camera, but be connected with outside camera, if terminal is PC (PersonalComputer, personal computer), in PC, camera is not set, but PC be connected with outside camera; Camera start unit 301, when receiving video recording enabled instruction, starts the camera be connected with PC, and obtains the video image frame of camera collection.
Target area acquiring unit 302, for obtaining the target photographic images region in described video image frame.
In specific implementation, display screen in terminal can show the video image frame of camera collection, user is by the target photographic images region in the touch-screen selection video image frame of terminal, also by the facial image region in face recognition technology identification vide environment as target photographic images region, the reference object in the vide environment corresponding to target photographic images region is the target reference object in the present embodiment.Be not construed as limiting target reference object in the embodiment of the present invention, target reference object can be any object that can produce voice signal, such as: people, animal, current etc.
Further, see Fig. 4, the structural representation of the target area acquiring unit that the embodiment of the present invention provides, the target area acquiring unit 302 in the present embodiment as shown in the figure can comprise: facial image recognin unit 3201 and target area arrange subelement 3202.
Facial image recognin unit 3201, for identifying the facial image region in described video image frame; Target area arranges subelement 3202, for described facial image region is set to target photographic images region.
In specific implementation, usually in the video having personage, wherein personage is the target reference object of user, in video process, be therefore more concerned about whether the sound that this personage sends can preserve clearly.Be in the use scenes of personage at target reference object, can utilize facial image region in face recognition technology identification video image frame by facial image recognin unit 3201, then target area arranges subelement 3202 and this facial image region is set to target photographic images region.
Distance value acquiring unit 303, for obtaining distance value between the target reference object corresponding with described target photographic images region as the first distance.
In the present embodiment, distance value acquiring unit 303 obtains the distance value between the target reference object corresponding with described target photographic images region, can by focusing to target photographic images region, focus according to the image distance in target photographic images region and the focal length of current camera successfully, calculate the distance value between the target reference object corresponding with target photographic images region; Also by installing distance measuring sensor in terminal, as ultrasonic range finder sensor, laser range sensor, radar range finding transducer and infrared distance measuring transducer etc., go to detect the distance value between target reference object by distance measuring sensor.
Further, see Fig. 5, the structural representation of the distance value acquiring unit that the embodiment of the present invention provides, the distance value acquiring unit 303 in the present embodiment as shown in the figure can comprise: target area focusing subelement 3301 and distance value computation subunit 3302.
Target area focusing subelement 3301, for receiving video recording focusing instruction, starting described camera and focusing to described target photographic images region; Distance value computation subunit 3302, for according to the image distance in described target photographic images region and the focal length of current described camera, calculates distance value between the target reference object corresponding with described target photographic images region as the first distance.
In specific implementation, obtain the distance value between the target reference object corresponding with described target photographic images region, can to be focused the focusing of the subelement 3301 pairs of target photographic images regions by target area, the distance value computation subunit 3302 afterwards of focus is successfully according to the distance value between target reference object corresponding to lens imaging formulae discovery target photographic images region.Wherein lens imaging formula is: 1/f=1/u+1/v, wherein f is the real focal length of camera when a certain enlargement ratio, u is object distance (distance value namely between target reference object and camera), and v is image distance (distance value between the target photographic images regional imaging that namely target reference object is corresponding and camera).
Described distance value acquiring unit 303, also for obtaining the sound signal components that in video process, in vide environment, each sound source sends, and obtains with the distance value of described each sound source as second distance set.
In actual implementation process, in terminal, microphone array can be installed, adopt the sound localization method of microphone array to carry out auditory localization to each sound source in vide environment.Sound localization method wherein based on microphone array comprises: the steerable beam formation technology based on peak power output, the directional technology based on High-Resolution Spectral Estimation and based on to the difference technology time of advent of voice signal, because based on differing from the time of advent to voice signal, technique computes amount is little, precision is high, therefore this kind of method is often used in actual applications.Adopting after the sound localization method of microphone array carries out auditory localization to each sound source in vide environment, just the distance value in vide environment between each sound source and terminal can be obtained, also can obtain the orientation of each sound source relative to terminal, terminal obtains distance value between each sound source and terminal as second distance set.
Further, see Fig. 6, the structural representation of the another kind of distance value acquiring unit that the embodiment of the present invention provides, the distance value acquiring unit in the present embodiment as shown in the figure can comprise: sound signal collecting subelement 3311, voice signal segregant unit 3312 and distance value computation subunit 3313.
Sound signal collecting subelement 3311, for gathering the voice signal in vide environment by least three microphones in video process; Voice signal segregant unit 3312, obtains the corresponding sound signal components of each sound source in described vide environment for separating of described voice signal; Distance value computation subunit 3313, for the time of the corresponding sound signal components of sound source each according to each microphones, calculates the distance value with described each sound source.
In specific implementation, start sound signal collecting subelement 3311 in video process and gather the sound signal components that in vide environment, each sound source sends, this sound signal collecting subelement 3311 can be specially microphone, and the sound signal components that each sound source sends comprises the sound signal components that target reference object sends.
Aliasing when usual each sound signal components arrives voice collection device, now voice signal segregant unit 3312 can cross sound isolation technics, go to obtain the corresponding sound signal components of each sound source, such as adopt the aliasing voice signal separation method based on voice fundamental frequencies, first the voice signal of aliasing can be carried out Fourier transform and the signal in time domain is transformed into frequency-region signal, then on frequency domain, the dominant frequency signal occurred is separated, namely the frequency spectrum of each dominant frequency signal is obtained respectively by band pass filter, again by be separated after Spectrum Conversion on time domain to obtain the time-domain signal of each sound signal components.
In actual implementation process, can install at least three microphones in terminal, in same plane, any point is unique to the distance of fixing three points, namely can determine another one point in plane by any 3.Because above-mentioned microphone is fixed in terminal, the position of namely going up microphone is known, positional distance between such microphone is exactly known, and the speed that voice signal is propagated also is known, distance value computation subunit 3313 just can calculate the relative position of each sound source and terminal according to the time of reception that these known conditions add each microphone again, such as: the relative position being calculated each sound source and terminal by plane geometry related algorithm.
After auditory localization is carried out to each sound source in vide environment, just the distance value in vide environment between each sound source and terminal can be obtained, also can obtain the orientation of each sound source relative to terminal, terminal obtains distance value between each sound source and terminal as second distance set.
Target sound source determining unit 304, for choosing sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source.
In specific implementation, after distance value acquiring unit 303 distance value got in vide environment between each sound source and terminal forms second distance set, target sound source determining unit 304 just can calculate each distance value in second distance set and the difference between the first distance, choose wherein difference and be less than or equal to the sound source corresponding to distance value of preset value as target sound source, namely this target sound source is considered as target reference object in the present embodiment, then preserve by voice signal storage unit 305 sound that the corresponding sound signal components of this target sound source sends as target reference object, the sound that in maskable vide environment, other sound sources send, noise in the vide environment gathered during to reduce video recording, make the sound of target reference object better effects if in video more clear, improve Consumer's Experience.
Voice signal storage unit 305, for preserving the sound signal components corresponding with described target sound source.
In another embodiment, described terminal also comprises: video storage unit 306, for receiving video recording halt instruction, exporting this record a video the video image frame that collects and sound signal components corresponding to described target sound source and forming video.
In specific implementation, when recording a video, terminal detects video recording halt instruction in real time, this video recording halt instruction can timing video recording time arrive timing time terminal system automatically generate, also can be that user inputs at any time.When video storage unit 306 receives video recording halt instruction, export the video image frame that collects in this video process and sound signal components corresponding to target sound source to form video and be saved in the storage area of specifying, carry out for user operations such as sharing, check.
The embodiment of the present invention, by camera start unit 301 in reception video recording enabled instruction, starts camera collection video image frame; The target photographic images region in described video image frame is also obtained by target area acquiring unit 302; Distance value acquiring unit 303 also can be utilized to obtain distance value between the target reference object corresponding with described target photographic images region as the first distance; Also obtain single 303 by distance value and obtain the sound signal components that in video process, in vide environment, each sound source sends, and obtain with the distance value of described each sound source as second distance set; And then choose sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source by target sound source determining unit 304; The sound signal components corresponding with described target sound source is preserved finally by voice signal storage unit 305.Can the sound sent of reference object be preserved when recording a video and reduce noise, improve the quality of video, improve Consumer's Experience.
For each embodiment of the method aforesaid, in order to simple description, therefore it is all expressed as a series of combination of actions, but those skilled in the art should know, the application is not by the restriction of described sequence of movement, because according to the application, certain some step can adopt other orders or carry out simultaneously.Secondly, those skilled in the art also should know, the embodiment described in specification all belongs to preferred embodiment, and involved action and unit might not be that the application is necessary.
One of ordinary skill in the art will appreciate that all or part of flow process realized in above-described embodiment method, that the hardware that can carry out instruction relevant by computer program has come, described program can be stored in a computer read/write memory medium, this program, when performing, can comprise the flow process of the embodiment as above-mentioned each side method.Wherein, described storage medium can be magnetic disc, CD, read-only store-memory body (Read-OnlyMemory, ROM) or random access memory (RandomAccessMemory is called for short RAM) etc.
Above a kind of kinescope method disclosed in the embodiment of the present invention and terminal are described in detail, above disclosedly be only present pre-ferred embodiments, certainly the interest field of the present invention can not be limited with this, therefore according to the equivalent variations that the claims in the present invention are done, the scope that the present invention is contained still is belonged to.

Claims (10)

1. a kinescope method, is characterized in that, comprising:
Receive video recording enabled instruction, start camera collection video image frame;
Obtain the target photographic images region in described video image frame;
Obtain distance value between the target reference object corresponding with described target photographic images region as the first distance;
Obtain the sound signal components that in video process, in vide environment, each sound source sends, and obtain with the distance value of described each sound source as second distance set;
Choose sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source;
Preserve the sound signal components corresponding with described target sound source.
2. method according to claim 1, is characterized in that, the distance value between the target reference object that described acquisition is corresponding with described target photographic images region comprises as the first distance:
Receive video recording focusing instruction, start described camera and focused in described target photographic images region;
According to the image distance in described target photographic images region and the focal length of current described camera, calculate distance value between the target reference object corresponding with described target photographic images region as the first distance.
3. method according to claim 1, is characterized in that, the sound signal components that in described acquisition video process, in vide environment, each sound source sends, and acquisition comprises as second distance set with the distance value of described each sound source:
The voice signal in vide environment is gathered by least three microphones in video process;
Be separated the corresponding sound signal components that described voice signal obtains each sound source in described vide environment;
The time of the sound signal components that each sound source is corresponding according to each microphones, calculate with the distance value of described each sound source as second distance set.
4. method according to claim 1, is characterized in that, described method also comprises:
Receive video recording halt instruction, export this record a video the video image frame that collects and sound signal components corresponding to described target sound source and form video.
5. method according to claim 1, is characterized in that, the target photographic images region in the described video image frame of described acquisition comprises;
Identify the facial image region in described video image frame;
Described facial image region is set to target photographic images region.
6. a terminal, is characterized in that, described terminal comprises:
Camera start unit, for receiving video recording enabled instruction, starts camera collection video image frame;
Target area acquiring unit, for obtaining the target photographic images region in described video image frame;
Distance value acquiring unit, for obtaining distance value between the target reference object corresponding with described target photographic images region as the first distance;
Described distance value acquiring unit, also for obtaining the sound signal components that in video process, in vide environment, each sound source sends, and obtains with the distance value of described each sound source as second distance set;
Target sound source determining unit, for choosing sound source corresponding to the distance value being less than or equal to preset value in described second distance set with the difference of described first distance as target sound source;
Voice signal storage unit, for preserving the sound signal components corresponding with described target sound source.
7. terminal according to claim 6, is characterized in that, described distance value acquiring unit comprises:
Target area focusing subelement, for receiving video recording focusing instruction, starting described camera and focusing to described target photographic images region;
Distance value computation subunit, for according to the image distance in described target photographic images region and the focal length of current described camera, calculates distance value between the target reference object corresponding with described target photographic images region as the first distance.
8. terminal according to claim 6, is characterized in that, described distance value acquiring unit comprises:
Sound signal collecting subelement, for gathering the voice signal in vide environment by least three microphones in video process;
Voice signal segregant unit, obtains the corresponding sound signal components of each sound source in described vide environment for separating of described voice signal;
Distance value computation subunit, for the time of the corresponding sound signal components of sound source each according to each microphones, calculates with the distance value of described each sound source as second distance set.
9. terminal according to claim 6, is characterized in that, described terminal also comprises:
Video storage unit, for receiving video recording halt instruction, exporting this record a video the video image frame that collects and sound signal components corresponding to described target sound source and forming video.
10. terminal according to claim 6, is characterized in that, described target area acquiring unit comprises:
Facial image recognin unit, for identifying the facial image region in described video image frame;
Target area arranges subelement, for described facial image region is set to target photographic images region.
CN201510407614.6A 2015-07-10 2015-07-10 Video recording method and terminal Pending CN105578097A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510407614.6A CN105578097A (en) 2015-07-10 2015-07-10 Video recording method and terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510407614.6A CN105578097A (en) 2015-07-10 2015-07-10 Video recording method and terminal

Publications (1)

Publication Number Publication Date
CN105578097A true CN105578097A (en) 2016-05-11

Family

ID=55887701

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510407614.6A Pending CN105578097A (en) 2015-07-10 2015-07-10 Video recording method and terminal

Country Status (1)

Country Link
CN (1) CN105578097A (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106027933A (en) * 2016-06-21 2016-10-12 维沃移动通信有限公司 Video recording method, video playing method and mobile terminal
CN106791375A (en) * 2016-11-29 2017-05-31 维沃移动通信有限公司 One kind shoots focusing method and mobile terminal
CN107071271A (en) * 2017-02-16 2017-08-18 中国石油化工股份有限公司 The portable explosion-proof camera system of integration
CN107360387A (en) * 2017-07-13 2017-11-17 广东小天才科技有限公司 The method, apparatus and terminal device of a kind of video record
CN108419016A (en) * 2018-04-17 2018-08-17 北京小米移动软件有限公司 Image pickup method, device and terminal
CN108769400A (en) * 2018-05-23 2018-11-06 宇龙计算机通信科技(深圳)有限公司 A kind of method and device of locating recordings
CN110505403A (en) * 2019-08-20 2019-11-26 维沃移动通信有限公司 A kind of video record processing method and device
CN110572600A (en) * 2019-08-20 2019-12-13 维沃移动通信有限公司 video processing method and electronic equipment
CN110740259A (en) * 2019-10-21 2020-01-31 维沃移动通信有限公司 Video processing method and electronic equipment
WO2020062900A1 (en) * 2018-09-29 2020-04-02 华为技术有限公司 Sound processing method, apparatus and device
CN111486537A (en) * 2020-06-15 2020-08-04 江苏新科电器有限公司 Air conditioner with security and monitoring functions
WO2020248829A1 (en) * 2019-06-10 2020-12-17 海信视像科技股份有限公司 Audio and video processing method and display device
CN112578338A (en) * 2019-09-27 2021-03-30 阿里巴巴集团控股有限公司 Sound source positioning method, device, equipment and storage medium
CN113099031A (en) * 2021-02-26 2021-07-09 华为技术有限公司 Sound recording method and related equipment
WO2021175165A1 (en) * 2020-03-06 2021-09-10 华为技术有限公司 Audio processing method and device
RU2798865C1 (en) * 2022-08-29 2023-06-28 Федеральное государственное бюджетное образовательное учреждение высшего образования "Владивостокский государственный университет" (ФГБОУ ВО "ВВГУ") Cell phone

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101246690A (en) * 2007-02-15 2008-08-20 索尼株式会社 Sound processing apparatus, sound processing method and program
US20090141908A1 (en) * 2007-12-03 2009-06-04 Samsung Electronics Co., Ltd. Distance based sound source signal filtering method and apparatus
CN103841357A (en) * 2012-11-21 2014-06-04 中兴通讯股份有限公司 Microphone array sound source positioning method, device and system based on video tracking
CN103916723A (en) * 2013-01-08 2014-07-09 联想(北京)有限公司 Sound acquisition method and electronic equipment
CN104580992A (en) * 2014-12-31 2015-04-29 广东欧珀移动通信有限公司 Control method and mobile terminal

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101246690A (en) * 2007-02-15 2008-08-20 索尼株式会社 Sound processing apparatus, sound processing method and program
US20090141908A1 (en) * 2007-12-03 2009-06-04 Samsung Electronics Co., Ltd. Distance based sound source signal filtering method and apparatus
CN103841357A (en) * 2012-11-21 2014-06-04 中兴通讯股份有限公司 Microphone array sound source positioning method, device and system based on video tracking
CN103916723A (en) * 2013-01-08 2014-07-09 联想(北京)有限公司 Sound acquisition method and electronic equipment
CN104580992A (en) * 2014-12-31 2015-04-29 广东欧珀移动通信有限公司 Control method and mobile terminal

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106027933A (en) * 2016-06-21 2016-10-12 维沃移动通信有限公司 Video recording method, video playing method and mobile terminal
CN106027933B (en) * 2016-06-21 2019-02-15 维沃移动通信有限公司 A kind of recording of video, playback method and mobile terminal
CN106791375A (en) * 2016-11-29 2017-05-31 维沃移动通信有限公司 One kind shoots focusing method and mobile terminal
CN106791375B (en) * 2016-11-29 2019-10-15 维沃移动通信有限公司 A kind of shooting focusing method and mobile terminal
CN107071271A (en) * 2017-02-16 2017-08-18 中国石油化工股份有限公司 The portable explosion-proof camera system of integration
CN107360387A (en) * 2017-07-13 2017-11-17 广东小天才科技有限公司 The method, apparatus and terminal device of a kind of video record
CN108419016A (en) * 2018-04-17 2018-08-17 北京小米移动软件有限公司 Image pickup method, device and terminal
CN108419016B (en) * 2018-04-17 2022-03-11 北京小米移动软件有限公司 Shooting method and device and terminal
CN108769400A (en) * 2018-05-23 2018-11-06 宇龙计算机通信科技(深圳)有限公司 A kind of method and device of locating recordings
CN110970057A (en) * 2018-09-29 2020-04-07 华为技术有限公司 Sound processing method, device and equipment
WO2020062900A1 (en) * 2018-09-29 2020-04-02 华为技术有限公司 Sound processing method, apparatus and device
CN110970057B (en) * 2018-09-29 2022-10-28 华为技术有限公司 Sound processing method, device and equipment
WO2020248829A1 (en) * 2019-06-10 2020-12-17 海信视像科技股份有限公司 Audio and video processing method and display device
CN110505403A (en) * 2019-08-20 2019-11-26 维沃移动通信有限公司 A kind of video record processing method and device
CN110572600A (en) * 2019-08-20 2019-12-13 维沃移动通信有限公司 video processing method and electronic equipment
CN112578338B (en) * 2019-09-27 2024-05-14 阿里巴巴集团控股有限公司 Sound source positioning method, device, equipment and storage medium
CN112578338A (en) * 2019-09-27 2021-03-30 阿里巴巴集团控股有限公司 Sound source positioning method, device, equipment and storage medium
CN110740259A (en) * 2019-10-21 2020-01-31 维沃移动通信有限公司 Video processing method and electronic equipment
WO2021078116A1 (en) * 2019-10-21 2021-04-29 维沃移动通信有限公司 Video processing method and electronic device
CN110740259B (en) * 2019-10-21 2021-06-25 维沃移动通信有限公司 Video processing method and electronic equipment
WO2021175165A1 (en) * 2020-03-06 2021-09-10 华为技术有限公司 Audio processing method and device
CN111486537B (en) * 2020-06-15 2020-10-02 江苏新科电器有限公司 Air conditioner with security and monitoring functions
CN111486537A (en) * 2020-06-15 2020-08-04 江苏新科电器有限公司 Air conditioner with security and monitoring functions
CN113099031B (en) * 2021-02-26 2022-05-17 华为技术有限公司 Sound recording method and related equipment
CN113099031A (en) * 2021-02-26 2021-07-09 华为技术有限公司 Sound recording method and related equipment
RU2798865C1 (en) * 2022-08-29 2023-06-28 Федеральное государственное бюджетное образовательное учреждение высшего образования "Владивостокский государственный университет" (ФГБОУ ВО "ВВГУ") Cell phone

Similar Documents

Publication Publication Date Title
CN105578097A (en) Video recording method and terminal
CN110970057B (en) Sound processing method, device and equipment
CN108156561B (en) Audio signal processing method and device and terminal
KR101703388B1 (en) Audio processing apparatus
KR20150066455A (en) Audio information processing method and apparatus
CN111429517A (en) Relocation method, relocation device, storage medium and electronic device
CN105245811B (en) A kind of kinescope method and device
CN107534725A (en) A kind of audio signal processing method and device
CN110089131A (en) Distributed audio capture and mixing control
CN113099031B (en) Sound recording method and related equipment
CN103051838A (en) Shoot control method and device
CN106851087A (en) Information processor and information processing method
CN114846816B (en) Stereo pickup method, stereo pickup device, terminal device and computer-readable storage medium
CN105578023A (en) Image quick photographing method and device
US20140086551A1 (en) Information processing apparatus and information processing method
CN113192527A (en) Method, apparatus, electronic device and storage medium for cancelling echo
CN105323480A (en) Ultrasonic-based photographing method and device
Zhu et al. Gaussian filter for TDOA based sound source localization in multimedia surveillance
US20170215005A1 (en) Audio refocusing methods and electronic devices utilizing the same
CN106993130A (en) Gather method, device and the mobile device of image
CN106803886A (en) A kind of method and device taken pictures
CN105791666B (en) Automatic focusing mechanism
CN116095254B (en) Audio processing method and device
JP2015104091A (en) Voice processing device with wind noise reduction device, and imaging apparatus with the same
CN117859339A (en) Media device, control method and device thereof, and target tracking method and device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20160511

RJ01 Rejection of invention patent application after publication