CN113676668A - Video shooting method and device, electronic equipment and readable storage medium - Google Patents

Video shooting method and device, electronic equipment and readable storage medium Download PDF

Info

Publication number
CN113676668A
CN113676668A CN202110977501.5A CN202110977501A CN113676668A CN 113676668 A CN113676668 A CN 113676668A CN 202110977501 A CN202110977501 A CN 202110977501A CN 113676668 A CN113676668 A CN 113676668A
Authority
CN
China
Prior art keywords
audio
sound
sound collection
area
prompt information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110977501.5A
Other languages
Chinese (zh)
Inventor
李俊潓
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202110977501.5A priority Critical patent/CN113676668A/en
Publication of CN113676668A publication Critical patent/CN113676668A/en
Priority to PCT/CN2022/113065 priority patent/WO2023025004A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording

Abstract

The application discloses a video shooting method and device, electronic equipment and a readable storage medium, and belongs to the technical field of communication. The video shooting method comprises the following steps: under the condition that a video shooting preview interface comprises at least one sound acquisition area, displaying audio quality prompt information of each sound acquisition area according to audio parameters of audio corresponding to each sound acquisition area; receiving a first input of audio quality prompt information corresponding to a target sound acquisition area by a user; and responding to the first input, adjusting the audio parameters of the audio corresponding to the target sound collection area, and generating a video file.

Description

Video shooting method and device, electronic equipment and readable storage medium
Technical Field
The application belongs to the technical field of communication, and particularly relates to a video shooting method and device, electronic equipment and a readable storage medium.
Background
Currently, more and more users shoot videos through electronic devices to record or communicate. However, in the process of shooting a video by the electronic device, a user cannot know the quality of the sound collected by the electronic device in real time. Only after the video shooting is completed, the user can know the recording quality of the sound by playing the video, and because the user cannot know the quality of the sound recorded by the electronic equipment in the shooting process, the shooting effect of the video is poor, and the shooting experience of the user is influenced.
Disclosure of Invention
The embodiment of the application aims to provide a video shooting method, a video shooting device, electronic equipment and a readable storage medium, and can solve the problems that a user cannot know the quality of sound recorded by the electronic equipment in a shooting process, the shooting effect of a video is poor and the shooting experience of the user is affected easily.
In a first aspect, an embodiment of the present application provides a video shooting method, where the method includes:
under the condition that a video shooting preview interface comprises at least one sound acquisition area, displaying audio quality prompt information of each sound acquisition area according to audio parameters of audio corresponding to each sound acquisition area;
receiving a first input of audio quality prompt information corresponding to a target sound acquisition area by a user;
and responding to the first input, adjusting the audio parameters of the audio corresponding to the target sound collection area, and generating a video file.
In a second aspect, an embodiment of the present application provides a video shooting apparatus, including:
the display module is used for displaying audio quality prompt information of each sound acquisition area according to audio parameters of audio corresponding to each sound acquisition area under the condition that the video shooting preview interface comprises at least one sound acquisition area;
the receiving module is used for receiving first input of audio quality prompt information corresponding to the target sound acquisition area by a user;
and the processing module is used for responding to the first input, adjusting the audio parameters of the audio corresponding to the target sound acquisition area and generating a video file.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored on the memory and executable on the processor, and when executed by the processor, the program or instructions implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the application, the video playing interface comprises sound collection areas, in the video shooting process, the audio parameters of audio corresponding to each sound collection area can be obtained according to each sound collection area, the audio quality prompt information of each sound collection area is displayed, a user can know the quality of the audio corresponding to each sound collection area collected by the electronic equipment in real time, meanwhile, the user can adjust the audio parameters of the audio corresponding to different sound collection areas according to own video shooting requirements by performing first input on the audio quality prompt information, a video file meeting the user shooting requirements is generated, the overall effect of video shooting is improved, and the user shooting experience is improved.
Drawings
Fig. 1 is a schematic flowchart of a video shooting method according to an embodiment of the present application;
fig. 2 is a schematic flowchart of a video shooting method according to an embodiment of the present application;
fig. 3 is a schematic flowchart of another video shooting method provided in an embodiment of the present application;
fig. 4 is a schematic flowchart of another video shooting method provided in the embodiment of the present application;
fig. 5 is a schematic structural diagram of a pop-up video shooting device according to an embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of an electronic device according to an embodiment of the present application;
fig. 7 is a schematic hardware structure diagram of another electronic device provided in an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The embodiment of the application provides a video shooting method, a device, electronic equipment and a readable storage medium, wherein a video playing interface comprises sound collection areas, and in the video shooting process, audio quality prompt information of each sound collection area can be displayed according to audio parameters of audio corresponding to each sound collection area, so that a user can know the quality of the audio corresponding to each sound collection area collected by the electronic equipment in real time conveniently.
The video shooting method, the video shooting device, the electronic device and the readable storage medium provided by the embodiments of the present application are described in detail below with reference to the accompanying drawings by specific embodiments and application scenarios thereof.
Fig. 1 is a schematic flowchart of a video shooting method provided in an embodiment of the present application, and may include the following steps 110 to 130.
And 110, under the condition that the video shooting preview interface comprises at least one sound acquisition area, displaying the audio quality prompt information of each sound acquisition area according to the audio parameters of the audio corresponding to each sound acquisition area.
In the embodiment of the application, in the process of video shooting, an image picture acquired by a camera of the electronic equipment can be displayed on a preview interface of the video shooting, namely the video shooting preview interface. The video shooting preview interface may also be an interface for video recording, an interface for video live broadcasting, and an interface for video call, which is not limited specifically herein.
A user can input a video shooting operation on the electronic equipment to enter a video shooting process in a manual operation mode, for example, clicking a camera application icon on an operation interface of the electronic equipment; or opening video shooting software, entering an interface of the video shooting software, and clicking a video shooting icon/button on the interface; or, the user can input video shooting operation on the electronic equipment in a voice calling mode; the user may also input a video shooting operation on the electronic device by a gesture or by shaking the electronic device, which is not specifically limited herein.
As shown in the schematic view of the video capture preview interface of fig. 2, the video capture preview interface may include characters, trees, cars, animals, musical instruments, and so on. Illustratively, the sound collection area included in the video capture preview interface may be, for example, an area where a person is located, such as the sound collection area 202 shown in fig. 2, or orientation information of the captured audio with respect to the electronic device, such as a position 45 ° behind the electronic device, such as the sound collection area 201 shown in fig. 2.
The 'rear 45 degrees' is only prompt information of a sound collection area in the video shooting process, so that a user can know the source of sound collected by the electronic equipment conveniently. And when the shot video is played, the video is not displayed on a video playing interface.
Taking the area where the person is located as the sound collection area as an example, the image recognition technology can be used for recognizing and obtaining that the video shooting preview interface comprises the person object, and determining that the person included in the video shooting preview interface is in a speaking state, so that the area where the person is located is used as the sound collection area. Objects that can emit sounds, such as animals, musical instruments, and automobiles, can be identified by image recognition technology, but are not specifically described here.
Taking the area where the position information is located as the sound collection area as an example, when the position 45 ° behind the sound collection area is used as the sound collection area, the sound collection area can be obtained by positioning the sound collected by the microphone. For example, as an example, the camera may be used as an origin, and a point microphone array is arranged in the electronic device, so as to determine front-back, up-down, left-right spatial orientation coordinates between the sound source and the electronic device, thereby determining orientation information of the collected sound relative to the electronic device.
In the embodiment of the application, each sound collection area corresponds to one audio, so that a user can know the audio of each sound collection area collected by the electronic equipment in real time conveniently, and one audio quality prompt message is displayed corresponding to each sound collection area.
For example, the audio quality prompt information may be determined according to audio parameters such as decibels and signal-to-noise ratios, and the audio quality prompt information may be obtained by analyzing the audio parameters according to a preset audio recognition model.
As a specific example, the audio quality cue information may be a decibel value of the audio. For example, the audio quality prompt information may be determined according to the decibel size of the audio, the decibel value corresponding to the audio may be directly displayed corresponding to each sound collection area, and the decibel value of the audio displayed on the video shooting preview interface may be directly used as an audio adjustment control to receive the first input of the user. For another example, when the audio quality prompt information is comprehensively determined according to the decibel and the signal-to-noise ratio of the audio, the decibel value can be displayed differently through different colors, specifically, when the decibel of the audio is greater than 80dB, a red decibel value is displayed; when the decibel is greater than 40dB and less than or equal to 80dB, displaying a green decibel value; when the decibel of the audio is less than or equal to 40dB, the decibel value of black and green is displayed.
As yet another specific example, the audio quality hint information may be an indication that different audio qualities may be distinguished by different colors, for example, in order to facilitate a user to quickly understand the audio parameters of the currently captured audio. Meanwhile, the identifier displayed on the video shooting preview interface can be directly used as an audio adjusting control, and can receive the first input of the user.
For example, when the audio quality prompt information is comprehensively determined according to decibels and a SIGNAL-to-NOISE RATIO (SNR), when the decibels are greater than 80dB and the SNR is less than 0dB, a red horn mark is displayed; when the decibel is greater than 40dB or less than or equal to 80dB and the SNR is greater than 10dB, displaying a green horn mark; when the decibel is less than or equal to 40dB, the SNR is greater than 0 or the SNR is less than 10dB, a horn mark of black and green is displayed.
Specifically, fig. 3 shows a schematic diagram of audio quality prompt information, where the audio quality prompt information is displayed by speaker identifiers with different colors, and in conjunction with fig. 3, the video capture preview interface includes a plurality of sound display areas, for example: a display area where a car is located, a display area where a character is located, a display area where a bird is located, a display area where a musical instrument is located, and 45 ° behind an electronic device. Wherein, each sound display area correspondingly displays the loudspeaker identification with different colors. It can be understood that, the user may preset the threshold values corresponding to different colors and different audio parameters represented by different colors according to the needs of the user, which are not specifically limited herein.
In the embodiment of the application, the audio quality prompt information of each sound collection area is displayed according to the audio parameters of the corresponding audio frequency of each sound collection area, so that a user can know the audio quality of each sound collection area collected by the electronic equipment in real time conveniently, and the video shooting experience of the user is improved.
Step 120, receiving a first input of the audio quality prompt information corresponding to the target sound collection area from the user.
In step 120, the sound collection area corresponding to the audio quality prompt message received as the first input is the target sound collection area.
The first input is an input by a user to adjust an audio parameter, and the first input may include a combination of a plurality of operation inputs by the user. For example, the user may click the audio quality prompt, and may display at least one audio parameter adjustment sub-control of the audio quality prompt, for example, a sub-control for adjusting decibel size, an audio deletion sub-control, an audio effect addition sub-control, and the like, which is not limited herein. The user can realize the adjustment of the specific audio parameters through further input of the audio adjustment sub-control.
In the embodiment of the application, the audio quality prompt information of each sound collection area is displayed, so that a user can know the audio quality of each sound collection area collected by the electronic equipment in real time, the user can input the audio quality prompt information in a first mode, the audio parameters of different sound collection areas can be adjusted according to own video shooting requirements, and video shooting experience is improved.
Step 130, in response to the first input, adjusting the audio parameters of the target sound collection area, and generating a video file.
In the embodiment of the application, each sound collection area corresponds to one audio quality prompt message, and a user can adjust the audio parameters corresponding to one or more sound collection areas according to the shooting requirements of the user. After the user finishes adjusting, in the subsequent audio acquisition process, the audio parameters adjusted by the user are used as default acquisition parameters until the video shooting is finished, a video file is generated, and the shooting experience of the user can be effectively improved.
As a specific example, the video file generated by the electronic device includes audio corresponding to the finally adjusted audio parameter of each sound collection area. The electronic equipment can store the generated video file, and when the user needs to edit the video file again, the audio quality prompt information corresponding to each sound collection area can be correspondingly displayed in the editing process of the video file, so that the user can conveniently edit the video file.
In some embodiments, the electronic device generated video file may include a file of audio corresponding to each sound collection area, so that when a user edits the video file, the user can conveniently operate the audio, and the convenience of video editing is improved.
As another specific example, a video file generated by the electronic device may be played in real time by live broadcasting; in the process of video call, the electronic equipment can send the generated video file to a conversation user, so that the watching experience of the video is improved.
In some embodiments, each sound collection area includes a first sound-emitting object, and the following steps may be specifically included in step 110 of the embodiment of the present application:
collecting a first audio; next, according to a preset audio recognition model, acquiring a second audio corresponding to each first sound-emitting object in the first audio; determining audio parameters of the second audio; and then, displaying the audio quality prompt information corresponding to each first sound-emitting object according to the audio parameter of each second audio.
Specifically, the first sound-emitting object may be a physical object in the video capture preview interface. Referring to the video capture preview interface diagram shown in fig. 2, the first sounding object may be, for example, a character, a tree, a car, an animal, a musical instrument, etc., wherein the character may make a conversation sound, a leaf wind sound, a car driving sound, an animal sounding sound, a musical instrument sounding sound, etc., which are not listed herein.
In the video shooting process, the first audio collected by the electronic equipment can comprise sounds emitted by various sound sources in the environment of the electronic equipment. The sound of different first sounding pairs in the first audio can be recognized according to the acoustic characteristics of different entity objects through a preset audio recognition model, and the audio corresponding to each first sounding object is separated through the preset audio recognition model, so that the second audio corresponding to each first sounding object is obtained. Next, the audio quality prompt information of each second audio may be displayed according to audio parameters such as decibels and signal-to-noise ratios of the audio, or the audio parameters may be analyzed according to a preset audio recognition model to obtain and determine the audio quality prompt information of each second audio, which is not specifically limited herein.
In this embodiment of the application, according to the audio quality of each second audio, audio quality prompt information corresponding to each first sound-emitting object is displayed, where the display manner may be as shown in fig. 4, the audio quality prompt information corresponding to each first sound-emitting object is displayed at a preset position of a video shooting preview interface, where the audio quality prompt information of cars and dogs may be displayed in red, the audio quality prompt information of guitars and people may be displayed in green, and the audio quality prompt information of birds may be displayed in black. In this way, the user can quickly know the audio quality of each first display object.
According to the displayed audio quality prompt information of each first sound-emitting object, a user adjusts audio parameters according to own requirements, for example, the decibel value of the audio of the automobile is adjusted, and if the user does not want to shoot a video file including the audio of the automobile, the user can select to delete the audio of the automobile. Illustratively, a user can click, slide and the like on the audio quality prompt information, and then a deletion sub-control can be displayed on a video shooting preview interface, so that the audio of the automobile can be deleted quickly. Therefore, the convenience of the video shooting process of the user can be improved, and the personalized requirements of the user are met.
In some embodiments, the sound collection area corresponds to a third audio, and the step 110 in the embodiment of the present application may further include the following steps:
firstly, receiving a second input of a user selecting a sound collection area; acquiring third audio based on first azimuth information corresponding to the sound acquisition area in response to the second input; determining audio parameters of a third audio; and then, displaying the audio quality prompt information corresponding to the sound acquisition area according to the audio parameter of the third audio.
As a specific example, a microphone array may be disposed in the electronic device, and spatial characteristics of a sound field around the electronic device may be sampled and processed by a certain number of microphones, so as to implement audio processing such as sound source localization and speech enhancement.
In the embodiment of the application, in order to meet the requirement of a user for obtaining the audio of the target direction and improve the video shooting experience of the user, the user can select the focusing position of the sound in the video shooting preview interface through second input, so that a sound collection area can be obtained. The second input may be, for example, a gesture input such as double-click or double-finger sliding, and the second input may further include a combination of a plurality of operation inputs of the user, where the second input for selecting the sound collection area may be preset and is not specifically limited herein.
In some embodiments, in the video capturing process, if the user selects the sound capture area, a preset default area may be used as the sound capture area, for example, a central area of the video capturing preview interface is used as the sound capture area.
In some embodiments, after the sound collection area is selected, the electronic device may focus the first azimuth information corresponding to the sound collection area through the microphone array, and collect a third audio. Therefore, the gain of the audio collected based on the first azimuth information can be improved, and the audio in other azimuths can be suppressed, so that the gain of the audio in the sound collection area can be improved, and the shooting requirement of a user can be met.
In some embodiments, in order to improve user experience and simplify operation steps of a user, step 110 in this embodiment of the present application may further include the following steps:
first, according to a preset image recognition model, second orientation information of each second sounding object in a video shooting preview interface is determined, fourth audio corresponding to each third orientation information is collected based on at least one piece of third orientation information, and audio parameters of each fourth audio are determined.
In some embodiments, when the second orientation information and the third orientation information satisfy a preset matching condition, determining a display area where each second sound-emitting object is located as a sound collection area; next, audio quality cue information for each sound collection region may be displayed according to the audio parameters of the fourth audio.
In some embodiments, in a case that the second orientation information and the third orientation information do not satisfy a preset matching condition, determining a display area corresponding to each fourth audio as a sound collection area; and displaying the audio quality prompt information of each sound acquisition area according to the audio parameters of each fourth audio.
Specifically, the user may select to turn on the image recognition function. In the video shooting process, the video shooting preview interface including the possible sounding objects, such as the characters, animals, musical instruments, automobiles, trees, and the like shown in fig. 2, is identified and obtained based on the preset image recognition model, and the second orientation information of each possible sounding object can be further obtained based on the preset image recognition model.
During video capture, at least one sound source around the electronic device and third orientation information of each sound source can be determined based on a microphone array in the electronic device. The electronic device can acquire and obtain fourth audio corresponding to each third party as information. The audio quality prompt information may be determined according to audio parameters such as decibels and signal-to-noise ratios of audio, and the audio quality prompt information may be obtained by analyzing the audio parameters according to a preset audio recognition model, which is not described herein again.
In the embodiment of the present application, the preset matching condition may be that a difference between the second orientation information and the third orientation information is smaller than or equal to a preset angle range, for example. And matching the second azimuth information of each second sounding object with third azimuth information corresponding to a fourth audio, wherein if the difference between the second azimuth information and the third azimuth information is smaller than or equal to a preset angle range, the display area where the second sounding object is located is a sound acquisition area, and then, the fourth audio corresponding to the third azimuth information can be used as the audio sent by the second sounding object. Thus, the audio quality prompt information of the second sound-emitting object in the sound collection area can be displayed according to the audio parameters of the fourth audio. Therefore, the operation steps of the user can be simplified, and the user experience of video shooting is improved.
In some embodiments, if the difference between the second orientation information and the third orientation information is greater than the preset angle range, it is determined that the second orientation information is not matched with the third orientation information, at this time, a display area corresponding to the fourth audio may be a sound collection area, and then, audio quality prompt information of each sound collection area may be displayed according to an audio parameter of each fourth audio.
According to the video shooting method provided by the embodiment of the application, as shown in fig. 3, the audio parameters of the sound-emitting object in the video shooting preview interface or the audio parameters of the sound source corresponding to the azimuth information can be visually displayed on the video shooting preview interface, so that a user can conveniently know the quality information of the audio collected by the current electronic equipment in real time.
It should be noted that, in the video shooting method provided in the embodiment of the present application, the execution subject may be a video shooting device, or a control module in the video shooting device for executing the video shooting method. The embodiment of the present application takes a method for executing video shooting by a video shooting device as an example, and describes a video shooting device provided by the embodiment of the present application.
Fig. 5 is a schematic structural diagram of a pop-up video camera 500 provided in an embodiment of the present application, and as shown in fig. 5, the video camera 500 may include: a display module 510, a receiving module 520, and a processing module 530.
The display module 510 is configured to display audio quality prompt information of each sound collection area according to an audio parameter of an audio corresponding to each sound collection area when the video shooting preview interface includes at least one sound collection area;
a receiving module 520, configured to receive a first input of audio quality prompt information corresponding to a target sound collection area from a user;
and the processing module 530 is configured to, in response to the first input, adjust an audio parameter of the audio corresponding to the target sound collection area, and generate a video file.
So, in this application embodiment, the video playing interface can include sound collection area, at the in-process of video shooting, can be according to the audio parameter of every sound collection area corresponding audio frequency, show the audio quality prompt message of every sound collection area, thereby convenience of customers knows the quality of the corresponding audio frequency of every sound collection area that electronic equipment gathered in real time, and simultaneously, the user is through carrying out first input to audio quality prompt message, can realize adjusting the audio parameter of the different sound collection area corresponding audio frequencies according to the video shooting demand of oneself, in order to generate the video file that satisfies the user and shoot the demand, improve the whole effect of video shooting, and improve the user and shoot experience.
In some embodiments, each sound collection area comprises a first originating object, the apparatus further comprising:
the acquisition module is used for acquiring a first audio;
the processing module 530 is further configured to obtain, according to a preset audio recognition model, a second audio corresponding to each first sound-emitting object in the first audio;
a processing module 530, further configured to determine audio parameters of the second audio;
the display module 510 is further configured to display, according to the audio parameter of each second audio, audio quality prompt information corresponding to each first sound-emitting object.
In this way, the user can quickly know the audio quality of each first display object.
In some embodiments, the sound collection area corresponds to a third audio, and the apparatus further comprises:
a receiving module 520, configured to receive a second input that a user selects a sound collection area;
the acquisition module is used for responding to the second input and acquiring third audio based on the first azimuth information corresponding to the sound acquisition area; and the number of the first and second groups,
a processing module 530, further configured to determine an audio parameter of a third audio;
the display module 510 is further configured to display audio quality prompt information corresponding to the sound collection area according to the audio parameter of the third audio.
Therefore, the requirement of the user for acquiring the target azimuth audio can be met, and the video shooting experience of the user is improved.
In some embodiments, the apparatus further comprises:
the positioning module is used for determining second azimuth information of each second sounding object in the video shooting preview interface according to a preset image recognition model;
the acquisition module is used for acquiring a fourth audio corresponding to each piece of third orientation information based on at least one piece of third orientation information;
a processing module 530, further configured to determine an audio parameter of each fourth audio;
the processing module 530 is further configured to determine, when the second orientation information and the third orientation information satisfy a preset matching condition, that the display area where each second sound object is located is a sound collection area;
the display module 510 is further configured to display the audio quality prompt information of each sound collection area according to the audio parameter of the fourth audio.
Therefore, the audio quality prompt information of the second sound-producing object in the sound collection area can be displayed, the operation steps of a user are simplified, and the user experience of video shooting is improved.
In some embodiments, the processing module 530 is further configured to determine, when the second orientation information and the third orientation information do not satisfy a preset matching condition, that the display area corresponding to each fourth audio is a sound collection area;
the processing module 530 is further configured to display the audio quality prompt information of each sound collection area according to the audio parameter of each fourth audio.
Therefore, the audio parameters of the sound source corresponding to the azimuth information can be visually displayed on a video shooting preview interface, and a user can conveniently know the quality information of the audio collected by the current electronic equipment in real time.
The video shooting device in the embodiment of the present application may be a device, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiments of the present application are not particularly limited.
The video capture device in the embodiment of the present application may be a device having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The video shooting device provided in the embodiment of the present application can implement each process implemented by the method embodiments of fig. 1 to 4, and is not described here again to avoid repetition.
Optionally, as shown in fig. 6, an electronic device 600 is further provided in this embodiment of the present application, and includes a processor 601, a memory 602, and a program or an instruction stored in the memory 602 and executable on the processor 601, where the program or the instruction is executed by the processor 601 to implement each process of the above-mentioned video shooting method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be noted that the electronic device in the embodiment of the present application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 7 is a schematic diagram of a hardware structure of an electronic device implementing an embodiment of the present application.
The electronic device 700 includes, but is not limited to: a radio frequency unit 701, a network module 702, an audio output unit 703, an input unit 704, a sensor 705, a display unit 706, a user input unit 707, an interface unit 708, a memory 709, and a processor 710.
Those skilled in the art will appreciate that the electronic device 700 may also include a power supply (e.g., a battery) for powering the various components, and the power supply may be logically coupled to the processor 710 via a power management system, such that the functions of managing charging, discharging, and power consumption may be performed via the power management system. The electronic device structure shown in fig. 7 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
The display unit 706 is configured to display audio quality prompt information of each sound collection area according to an audio parameter of an audio corresponding to each sound collection area under the condition that the video shooting preview interface includes at least one sound collection area;
a user input unit 707 configured to receive a first input of audio quality prompt information corresponding to a target sound collection area by a user;
and the processor 710 is configured to adjust an audio parameter of the audio corresponding to the target sound collection area in response to the first input, and generate a video file.
So, in this application embodiment, the video playing interface can include sound collection area, at the in-process of video shooting, can be according to the audio parameter of every sound collection area corresponding audio frequency, show the audio quality prompt message of every sound collection area, thereby convenience of customers knows the quality of the corresponding audio frequency of every sound collection area that electronic equipment gathered in real time, and simultaneously, the user is through carrying out first input to audio quality prompt message, can realize adjusting the audio parameter of the different sound collection area corresponding audio frequencies according to the video shooting demand of oneself, in order to generate the video file that satisfies the user and shoot the demand, improve the whole effect of video shooting, and improve the user and shoot experience.
In some embodiments, each sound collection area comprises a first originating object, the apparatus further comprising:
an input unit 704 for acquiring a first audio;
the processor 710 is further configured to obtain, according to a preset audio recognition model, a second audio corresponding to each first sound-emitting object in the first audio;
a processor 710 further configured to determine audio parameters of the second audio;
the display unit 706 is further configured to display, according to the audio parameter of each second audio, audio quality prompt information corresponding to each first sound-emitting object.
In this way, the user can quickly know the audio quality of each first display object.
In some embodiments, the sound collection area corresponds to a third audio data, and the apparatus further comprises:
a user input unit 707 for receiving a second input that the user selects a sound collection area;
an input unit 704 configured to acquire a third audio based on the first azimuth information corresponding to the sound acquisition region in response to the second input;
a processor 710 further configured to determine an audio parameter of a third audio;
the display unit 706 is further configured to display audio quality prompt information corresponding to the sound collection area according to the audio parameter of the third audio.
Therefore, the requirement of the user for acquiring the target azimuth audio can be met, and the video shooting experience of the user is improved.
In some embodiments, the apparatus further comprises:
the positioning module is used for determining second azimuth information of each second sounding object in the video shooting preview interface according to a preset image recognition model;
an input unit 704, configured to collect, based on at least one third orientation information, a fourth audio corresponding to each third orientation information;
a processor 710 for determining an audio parameter for each fourth audio;
the processor 710 is further configured to determine, when the second orientation information and the third orientation information satisfy a preset matching condition, that the display area where each second sound-generating object is located is a sound collection area;
the display unit 706 is further configured to display audio quality prompt information of each sound collection area according to the audio parameter of the fourth audio.
Therefore, the audio quality prompt information of the second sound-producing object in the sound collection area can be displayed, the operation steps of a user are simplified, and the user experience of video shooting is improved.
In some embodiments, the processor 710 is further configured to determine, when the second orientation information and the third orientation information do not satisfy a preset matching condition, a display area corresponding to each fourth audio as a sound collection area;
the processor 710 is further configured to display an audio quality prompt for each sound collection area according to the audio quality of each fourth audio.
Therefore, the audio parameters of the sound source corresponding to the azimuth information can be visually displayed on a video shooting preview interface, and a user can conveniently know the quality information of the audio collected by the current electronic equipment in real time.
It should be understood that in the embodiment of the present application, the input Unit 704 may include a Graphics Processing Unit (GPU) 7041 and a microphone 7042, and the Graphics Processing Unit 7041 processes image data of still pictures or videos obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The display unit 706 may include a display panel 7061, and the display panel 7061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 707 includes a touch panel 7071 and other input devices 7072. The touch panel 7071 is also referred to as a touch screen. The touch panel 7071 may include two parts of a touch detection device and a touch controller. Other input devices 7072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. Memory 709 may be used to store software programs as well as various data, including but not limited to applications and operating systems. Processor 710 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into processor 710.
The embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the video shooting method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the above video shooting method embodiment, and can achieve the same technical effect, and the details are not repeated here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A video capture method, comprising:
under the condition that a video shooting preview interface comprises at least one sound acquisition area, displaying audio quality prompt information of each sound acquisition area according to audio parameters of audio corresponding to each sound acquisition area;
receiving a first input of audio quality prompt information corresponding to a target sound acquisition area by a user;
and responding to the first input, adjusting the audio parameters of the audio corresponding to the target sound collection area, and generating a video file.
2. The method of claim 1, wherein each sound collection region comprises a first sounding object, and wherein displaying the audio quality prompt information of each sound collection region according to the audio parameter of the audio corresponding to each sound collection region comprises:
collecting a first audio;
acquiring a second audio corresponding to each first sound-emitting object in the first audio according to a preset audio recognition model;
determining audio parameters of the second audio;
and displaying the audio quality prompt information corresponding to each first sound-emitting object according to the audio parameters of each second audio.
3. The method of claim 1, wherein the sound collection region corresponds to a third audio, and the displaying the audio quality prompt information of each sound collection region according to the audio parameter of the audio corresponding to each sound collection region comprises:
receiving a second input of a user selecting a sound collection area;
responding to the second input, and acquiring third audio based on first azimuth information corresponding to the sound acquisition area;
determining audio parameters of the third audio;
and displaying audio quality prompt information corresponding to the sound acquisition area according to the audio parameters of the third audio.
4. The method according to claim 1, wherein the displaying the audio quality prompt information of each sound collection area according to the audio parameter of the audio corresponding to each sound collection area comprises:
determining second azimuth information of each second sound-producing object in the video shooting preview interface according to a preset image recognition model;
acquiring fourth audio corresponding to each piece of third orientation information based on at least one piece of third orientation information;
determining an audio parameter of each of the fourth audios;
under the condition that the second orientation information and the third orientation information meet preset matching conditions, determining a display area where each second sounding object is located as the sound collection area;
and displaying the audio quality prompt information of each sound acquisition area according to the audio parameters of the fourth audio.
5. The method of claim 4,
under the condition that the second direction information and the third direction information do not meet preset matching conditions, determining a display area corresponding to each fourth audio as the sound collection area;
and displaying the audio quality prompt information of each sound acquisition area according to the audio parameters of each fourth audio.
6. A video camera, comprising:
the display module is used for displaying audio quality prompt information of each sound acquisition area according to audio parameters of audio corresponding to each sound acquisition area under the condition that the video shooting preview interface comprises at least one sound acquisition area;
the receiving module is used for receiving first input of audio quality prompt information corresponding to the target sound acquisition area by a user;
and the processing module is used for responding to the first input, adjusting the audio parameters of the audio corresponding to the target sound acquisition area and generating a video file.
7. The apparatus of claim 6, wherein each of the sound collection regions comprises a first originating object, the apparatus further comprising:
the acquisition module is used for acquiring a first audio;
the processing module is further configured to obtain a second audio corresponding to each first sound-emitting object in the first audio according to a preset audio recognition model;
the processing module is further configured to determine an audio parameter of the second audio;
the display module is further configured to display audio quality prompt information corresponding to each first sound-emitting object according to the audio parameter of each second audio.
8. The apparatus of claim 6, wherein the sound collection area corresponds to a third audio, the apparatus further comprising:
the receiving module is used for receiving a second input of selecting a sound acquisition area by a user;
the acquisition module is used for responding to the second input and acquiring third audio based on the first azimuth information corresponding to the sound acquisition area;
the processing module is further configured to determine an audio parameter of the third audio;
the display module is further configured to display audio quality prompt information corresponding to the sound collection area according to the audio parameter of the third audio.
9. The apparatus of claim 6, further comprising:
the positioning module is used for determining second azimuth information of each second sounding object in the video shooting preview interface according to a preset image recognition model;
the acquisition module is used for acquiring a fourth audio corresponding to each piece of third orientation information based on at least one piece of third orientation information;
the processing module is further configured to determine an audio parameter of each of the fourth audios;
the processing module is further configured to determine a display area where each second sound-generating object is located as the sound collection area when the second orientation information and the third orientation information meet a preset matching condition;
the display module is further configured to display audio quality prompt information of each sound collection area according to the audio parameter of the fourth audio.
10. The apparatus of claim 9,
the processing module is further configured to determine, when the second orientation information and the third orientation information do not satisfy a preset matching condition, a display area corresponding to each fourth audio as the sound collection area;
the processing module is further configured to display audio quality prompt information of each sound collection area according to the audio parameter of each fourth audio.
CN202110977501.5A 2021-08-24 2021-08-24 Video shooting method and device, electronic equipment and readable storage medium Pending CN113676668A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202110977501.5A CN113676668A (en) 2021-08-24 2021-08-24 Video shooting method and device, electronic equipment and readable storage medium
PCT/CN2022/113065 WO2023025004A1 (en) 2021-08-24 2022-08-17 Video recording method and apparatus, electronic device, and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110977501.5A CN113676668A (en) 2021-08-24 2021-08-24 Video shooting method and device, electronic equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN113676668A true CN113676668A (en) 2021-11-19

Family

ID=78545815

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110977501.5A Pending CN113676668A (en) 2021-08-24 2021-08-24 Video shooting method and device, electronic equipment and readable storage medium

Country Status (2)

Country Link
CN (1) CN113676668A (en)
WO (1) WO2023025004A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114245156A (en) * 2021-11-30 2022-03-25 广州繁星互娱信息科技有限公司 Live broadcast audio adjusting method and device, storage medium and electronic equipment
CN114390205A (en) * 2022-01-29 2022-04-22 西安维沃软件技术有限公司 Shooting method and device and electronic equipment
WO2023025004A1 (en) * 2021-08-24 2023-03-02 维沃移动通信有限公司 Video recording method and apparatus, electronic device, and readable storage medium
WO2023143171A1 (en) * 2022-01-30 2023-08-03 华为技术有限公司 Audio acquisition method and electronic device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110740259A (en) * 2019-10-21 2020-01-31 维沃移动通信有限公司 Video processing method and electronic equipment
CN112165590A (en) * 2020-09-30 2021-01-01 联想(北京)有限公司 Video recording implementation method and device and electronic equipment
CN112309449A (en) * 2020-10-26 2021-02-02 维沃移动通信(深圳)有限公司 Audio recording method and device
CN112383817A (en) * 2020-11-13 2021-02-19 维沃移动通信有限公司 Volume adjusting method and device
CN113014983A (en) * 2021-03-08 2021-06-22 Oppo广东移动通信有限公司 Video playing method and device, storage medium and electronic equipment

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106331501A (en) * 2016-09-21 2017-01-11 乐视控股(北京)有限公司 Sound acquisition method and device
US9992532B1 (en) * 2017-01-11 2018-06-05 Htc Corporation Hand-held electronic apparatus, audio video broadcasting apparatus and broadcasting method thereof
CN112015365A (en) * 2020-08-27 2020-12-01 维沃移动通信有限公司 Volume adjustment method and device and electronic equipment
CN113676668A (en) * 2021-08-24 2021-11-19 维沃移动通信有限公司 Video shooting method and device, electronic equipment and readable storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110740259A (en) * 2019-10-21 2020-01-31 维沃移动通信有限公司 Video processing method and electronic equipment
CN112165590A (en) * 2020-09-30 2021-01-01 联想(北京)有限公司 Video recording implementation method and device and electronic equipment
CN112309449A (en) * 2020-10-26 2021-02-02 维沃移动通信(深圳)有限公司 Audio recording method and device
CN112383817A (en) * 2020-11-13 2021-02-19 维沃移动通信有限公司 Volume adjusting method and device
CN113014983A (en) * 2021-03-08 2021-06-22 Oppo广东移动通信有限公司 Video playing method and device, storage medium and electronic equipment

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023025004A1 (en) * 2021-08-24 2023-03-02 维沃移动通信有限公司 Video recording method and apparatus, electronic device, and readable storage medium
CN114245156A (en) * 2021-11-30 2022-03-25 广州繁星互娱信息科技有限公司 Live broadcast audio adjusting method and device, storage medium and electronic equipment
CN114390205A (en) * 2022-01-29 2022-04-22 西安维沃软件技术有限公司 Shooting method and device and electronic equipment
CN114390205B (en) * 2022-01-29 2023-09-15 西安维沃软件技术有限公司 Shooting method and device and electronic equipment
WO2023143171A1 (en) * 2022-01-30 2023-08-03 华为技术有限公司 Audio acquisition method and electronic device

Also Published As

Publication number Publication date
WO2023025004A1 (en) 2023-03-02

Similar Documents

Publication Publication Date Title
US11030987B2 (en) Method for selecting background music and capturing video, device, terminal apparatus, and medium
CN113676668A (en) Video shooting method and device, electronic equipment and readable storage medium
CN104991754B (en) The way of recording and device
CN112511750B (en) Video shooting method, device, equipment and medium
US11567729B2 (en) System and method for playing audio data on multiple devices
CN104092936A (en) Automatic focusing method and apparatus
CN111370018B (en) Audio data processing method, electronic device and medium
CN111986689A (en) Audio playing method, audio playing device and electronic equipment
US11863856B2 (en) Method and terminal device for matching photographed objects and preset text information
CN112653902A (en) Speaker recognition method and device and electronic equipment
CN109151565A (en) Play method, apparatus, electronic equipment and the storage medium of voice
CN111445901A (en) Audio data acquisition method and device, electronic equipment and storage medium
CN113727021A (en) Shooting method and device and electronic equipment
CN111416996A (en) Multimedia file detection method, multimedia file playing device, multimedia file equipment and storage medium
CN110798327A (en) Message processing method, device and storage medium
CN112165591B (en) Audio data processing method and device and electronic equipment
CN110767229B (en) Voiceprint-based audio output method, device and equipment and readable storage medium
CN112309449A (en) Audio recording method and device
CN110012359A (en) Answer reminding method and device
CN112788489B (en) Control method and device and electronic equipment
CN113329138A (en) Video shooting method, video playing method and electronic equipment
CN112584225A (en) Video recording processing method, video playing control method and electronic equipment
CN111696565B (en) Voice processing method, device and medium
KR102650763B1 (en) Psychoacoustic enhancement based on audio source directivity
CN111901675B (en) Multimedia data playing method and device, computer equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination