CN116567328A - Audio acquisition method and electronic equipment - Google Patents

Audio acquisition method and electronic equipment Download PDF

Info

Publication number
CN116567328A
CN116567328A CN202210113229.0A CN202210113229A CN116567328A CN 116567328 A CN116567328 A CN 116567328A CN 202210113229 A CN202210113229 A CN 202210113229A CN 116567328 A CN116567328 A CN 116567328A
Authority
CN
China
Prior art keywords
audio
devices
audio acquisition
user
sound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210113229.0A
Other languages
Chinese (zh)
Inventor
卞超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN202210113229.0A priority Critical patent/CN116567328A/en
Priority to PCT/CN2023/072317 priority patent/WO2023143171A1/en
Publication of CN116567328A publication Critical patent/CN116567328A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/436Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44222Analytics of user selections, e.g. selection of programs or purchase activity

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Databases & Information Systems (AREA)
  • Studio Devices (AREA)

Abstract

The application provides a method for collecting audio and electronic equipment, wherein the electronic equipment can be a mobile phone and the like, and the method can support a user to manually select different radio devices on the electronic equipment and also can select radio devices of other distributed equipment in the process of making a sound multimedia file by recording video and the like by the user; in addition, the user is supported to manually adjust gain parameters such as the sound receiving range, the sound intensity and the sound effect of the sound receiving device, so that the user can better select the sound receiving device, switch different sound receiving devices, adjust the sound receiving gain parameters and the like according to different requirements; furthermore, the method can automatically match one or more sound receiving devices according to a plurality of reference indexes such as the sound source position and the position of the sound receiving devices, acquire high-quality and interesting sound-producing multimedia files, meet shooting requirements of different users, improve the quality of collected audio, collect clearer audio for the users, improve the sound receiving effect and facilitate manufacturing of higher-quality sound-producing multimedia files.

Description

Audio acquisition method and electronic equipment
Technical Field
The present disclosure relates to the field of electronic technologies, and in particular, to a method for collecting audio and an electronic device.
Background
With the development and popularization of blowout type multimedia scenes such as short video recording and online live broadcasting, the use of video communication by users using electronic devices is becoming more and more popular.
A user may make a multimedia file full of interest such as a video, a sound photo, a moving picture with an image interchange format (graphics interchange format, gif) through an electronic device such as a mobile phone, and the multimedia file such as a video, a sound photo, a moving picture may be referred to as an "audio-image-combined multimedia file" or as an "audio multimedia file". In the process of producing a sound multimedia file, an image may be generally collected by a camera of an electronic device, and a sound may be collected by a microphone of the electronic device.
Taking the process of shooting video through a mobile phone as an example, a video picture can be generally collected based on a front camera or a rear camera of the mobile phone, and a radio is carried out based on a microphone fixed by the mobile phone, so that a video clip with audio and image combined is shot. In this process, the user can adjust shooting parameters related to the video picture, such as adjusting the focal length, switching cameras, and the like, according to the shooting needs of the individual. For example, the user may zoom in or out on the photographed image, for example, zooming in 50 times on the photographed image may photograph a distant object; or the shot picture is reduced to a macro mode, so that the shot picture is clearer, and the like. Or, the user may switch the front camera and the rear camera for photographing to acquire different framing contents, and the like.
In the process of shooting the audio multimedia file, the optimization of the shooting picture and the diversity of the framing content are mainly focused, and the adjustment or control of the audio multimedia file is not involved. For example, for the audio of the audio multimedia file, the user can only edit the audio after shooting is completed, so that the processing flow is increased; in addition, the user can not control the sound reception of the audio in the shooting process, the processing of the audio is weaker, the interestingness of the audio multimedia file can be reduced, and the shooting experience of the user is affected.
Disclosure of Invention
The method can meet shooting requirements of different users, improves quality of collected audio, collects clearer audio for the users, improves radio effect, and is beneficial to manufacturing higher-quality audio multimedia files.
A first aspect provides a method of capturing audio for use in an electronic device comprising a display screen, the method comprising: displaying an image preview interface for video recording, wherein the image preview interface comprises a first control and a second control, the first control is used for controlling audio in the video recording process, and the second control is used for starting or suspending the video recording process;
Detecting a first operation acting on the first control, and responding to the first operation, and displaying one or more options on the image preview interface, wherein the one or more options correspond to n audio acquisition devices, and n is an integer greater than or equal to 1;
detecting a second operation acting on a first option, and responding to the second operation, determining an audio acquisition device corresponding to the first option as a target audio acquisition device, or automatically matching the target audio acquisition device from the n audio acquisition devices;
detecting a third operation acting on the second control, starting to record video in response to the third operation, and acquiring the audio of the target sound source by using the target audio acquisition device in the process of recording video.
Through the above-mentioned process, to the process of recording the video, the embodiment of the application provides a method for selecting audio acquisition devices or switching audio acquisition devices for users, and users can check available n audio acquisition devices through microphone controls on an interface according to own shooting requirements, and manually click one or more audio acquisition devices selected to be expected to acquire audio, so that different audio acquisition devices are used for acquiring audio, the use requirements of different users are met, the autonomous selectivity of the users is improved, and further the experience of the users for shooting audio and other sound multimedia files is improved.
Optionally, the "image preview interface for video recording" in this embodiment of the present application may be an image display interface corresponding to a shooting mode capable of recording video, where the camera application is not limited to a system application configured by an electronic device factory, and may also include a system application downloaded by a user and installed to the electronic device, where the system application is capable of recording video in any one of a video recording mode, a double-view shooting mode, a picture-in-picture shooting mode, and the likeAnd the like, and is not limited thereto.
Alternatively, the "image preview interface for video recording" may also be user-in-useApplication, & gt>Any application having a function of capturing video, such as an application, for example, the user can use +.>Application of recording small video for 10 seconds, the video recording interface can be included in the scope of the embodiments of the application, and the process of audio acquisition is controlled according to the method provided by the embodiments of the application.
Alternatively, in the embodiment of the present application, a device capable of capturing audio is referred to as "sound capturing device", "audio capturing device", or "sound capturing device", for example, a microphone, an external earphone, or the like, which is not limited in the embodiment of the present application.
By way of example, the "first operation", "second operation", "third operation" may be a click operation, a double click operation, a long press operation, etc., which are not limited in the embodiments of the present application.
By way of example, the "one or more options" may correspond to one or more of an omnidirectional radio option, a front radio option, a rear radio option, a wired radio option, a bluetooth radio option, a more option, etc. on the radio device selection interface, where the "first option" may be any one of these.
With reference to the first aspect, in certain implementation manners of the first aspect, one or more audio collection devices configured on the electronic device are included in the n audio collection devices; and/or one or more audio acquisition devices of other devices which establish a connection relationship with the electronic device are included in the n audio acquisition devices, and the connection relationship includes wired connection or wireless connection.
In this embodiment of the present application, for example, the mobile phone may include one or more microphones installed on the mobile phone itself, and may further include microphones installed on other distributed devices that can be called by the mobile phone. For example, in a home scenario, a mobile phone may use a microphone of a distributed device such as a tablet, a smart screen, a speaker, etc. as an audio capturing device, and each distributed device may also include one or more audio capturing devices such as a microphone, which is not limited in this embodiment of the present application.
Through the process, in the process of manufacturing the audio multimedia file, a user can manually select different audio acquisition devices on one electronic device, and the user can be supported to manually select the audio acquisition devices using different distributed devices to acquire audio. According to the method, a user can manually select or switch different radio equipment according to different shooting requirements, shooting scenes and the like, so that different audio acquisition devices are used for acquiring audio, and the quality of audio acquisition is improved.
In one possible implementation, the electronic device may automatically intelligently match one or more audio capture devices for the user according to different reference indicators (reference options), and the options corresponding to the automatically matched audio capture devices on the audio capture device selection interface may be displayed as selected states.
Optionally, the "different reference indicators" may include one or more of the shooting scene, shooting parameters, user frequency, function of the audio capturing device, distance between the audio capturing device and the target sound source, etc. of the currently recorded video listed in table 3, and the electronic device may automatically match the audio capturing device according to the one or more reference indicators.
With reference to the first aspect and the foregoing implementation manner, in some implementation manners of the first aspect, the automatically matching the target audio capturing device from the n audio capturing devices includes:
according to the position of the target sound source, determining an audio acquisition device closest to the target sound source among the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the audio acquisition range corresponding to each audio acquisition device, determining the audio acquisition device with the largest audio acquisition range among the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the position of the target sound source and the audio acquisition range corresponding to each audio acquisition device, determining the audio acquisition devices of which the audio acquisition ranges cover the target sound source as the target audio acquisition devices; and/or the number of the groups of groups,
according to the use frequency of the user, determining the audio acquisition device with the highest use frequency of the user in the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the use record of the user, the audio acquisition device which is used by the user for the last time in the n audio acquisition devices is automatically determined to be the target audio acquisition device; and/or the number of the groups of groups,
According to the corresponding performance of each audio acquisition device, determining the audio acquisition device with the highest performance in the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the mode that each audio acquisition device is connected with the electronic equipment, the audio acquisition devices which are connected with the electronic equipment in a wired mode in the n audio acquisition devices are determined to be the target audio acquisition devices; and/or the number of the groups of groups,
according to the installation position of a camera used for collecting video pictures in the video recording process, automatically matching the target audio collecting devices from the n audio collecting devices; and/or the number of the groups of groups,
according to the source of a camera used for collecting video pictures in the video recording process, the target audio collecting devices are automatically matched from the n audio collecting devices.
Through the above process, in the process of producing the audio multimedia file, according to the distance between the audio acquisition device and the sound source, multiple reference indexes such as different positions of the audio acquisition device, and the like, one or more audio acquisition devices can be automatically matched for the current scene, or audio acquisition devices such as microphones of different devices can be switched to acquire audio in the process of receiving the audio. The process can improve the sound reception quality, collect clearer audio for users, improve the sound reception effect and facilitate the production of better-quality sound multimedia files.
It should be understood that, according to the implementation manner of the above-mentioned automatic matching audio capturing device, when the audio capturing device is automatically matched with the user on the audio capturing device selection interface, if the automatic matching audio capturing device is not the audio capturing device desired by the user, the user may also manually modify or switch, specifically refer to the above-mentioned manual selection of the operation process of the audio capturing device, and adjust parameters such as gain, so as to improve the user experience of the user.
With reference to the first aspect and the foregoing implementation manner, in some implementation manners of the first aspect, in a video recording process, when one or more cameras used for capturing video frames originate from the electronic device and one or more audio capturing devices configured on the electronic device are included in the n audio capturing devices, a preset correspondence is provided between the one or more cameras of the electronic device and the one or more audio capturing devices configured on the electronic device, and the automatically matching the target audio capturing device from the n audio capturing devices includes: and according to the corresponding relation between one or more cameras used for acquiring video pictures and the preset, automatically matching the target audio acquisition device from one or more audio acquisition devices configured on the electronic equipment.
For example, in a home scenario, when the video frames displayed by the cell phone originate from a distributed tablet camera, the tablet microphone may be synchronously selected as the target audio pickup device.
Or if the mobile phone is only connected with a unique device, whether in wired connection or wireless connection, the device connected with the mobile phone can be preferentially matched automatically as a target audio acquisition device, so that the user can be supported to manually modify or switch, and the manual modification or switching is not repeated here.
With reference to the first aspect and the foregoing implementation manner, in some implementation manners of the first aspect, in a video recording process, when one or more cameras used for capturing video frames originate from other devices that establish a connection with the electronic device, and one or more audio capturing devices of the n audio capturing devices include one or more other devices that establish a connection with the electronic device, the automatically matching, from among the n audio capturing devices, the target audio capturing device includes: and determining an audio acquisition device of other equipment with one or more camera sources used for acquiring video pictures as the target audio acquisition device.
With reference to the first aspect and the foregoing implementation manner, in some implementation manners of the first aspect, when one or more audio capturing devices of the n audio capturing devices include other devices that establish a connection with the electronic device, the method further includes:
detecting a fourth operation acting on the second option, determining one or more devices in connection with the electronic device in response to the fourth operation, and displaying controls including the one or more devices;
detecting a fifth operation acting on a control corresponding to a target device, and responding to the fifth operation, determining one or more audio acquisition devices of the target device as the target audio acquisition device, and acquiring the audio of a target sound source by using the target audio acquisition device in the process of recording video, wherein the connection relation comprises one or more wireless connection relations of Bluetooth connection, wireless local area network Wi-Fi connection, infrared connection and near field communication NFC connection.
With reference to the first aspect and the foregoing implementation manner, in some implementation manners of the first aspect, the one or more options include a preset third option, where the third option is displayed in a selected state, and when the first operation acting on the first control is not detected, an audio capture device corresponding to the third option is determined to be the target audio capture device.
With reference to the first aspect and the foregoing implementation manners, in certain implementation manners of the first aspect, after determining the target audio capturing device, the method further includes:
and displaying prompt information, wherein the prompt information is used for prompting the name or source of the target audio acquisition device used by the electronic equipment currently by a user.
Alternatively, the prompt information may be displayed on an image preview interface of the video recording, which is not limited in the embodiments of the present application.
Alternatively, the prompt may be displayed for a preset duration, for example, 2 seconds; the prompt information may also automatically disappear after the click operation of the user is detected, which is not limited in the embodiment of the present application.
Through the method, the audio acquisition device manually selected by the user or the automatically matched audio acquisition device can be displayed on the video recording image preview interface, so that the user can accurately know which audio acquisition device is receiving the radio and whether the audio acquisition device is the audio acquisition device expected by the user, and the shooting experience of the user is improved.
With reference to the first aspect and the foregoing implementation manners, in certain implementation manners of the first aspect, after determining the target audio capturing device, the method further includes:
Displaying one or more controls, wherein each control of the one or more controls is used for controlling the working parameters of the target audio acquisition device, and the working parameters comprise one or more of an audio acquisition range, audio acquisition intensity and audio playing effect in the working process of the target audio acquisition device;
detecting a sixth operation on the one or more controls, and adjusting an operating parameter of the target audio acquisition device in response to the sixth operation.
By way of example, the "sixth operation" may be a regulatory operation on the radio gain adjustment control and/or the channel control, such as a sliding gain parameter, or the like.
Specifically, "gain adjustment" may include three aspects: angle adjustment, intensity adjustment and effect adjustment. The adjustment process of each aspect may be divided into forward adjustment and reverse adjustment, and the adjustment amplitude and adjustment range of each aspect may be changed according to different electronic devices or different types of radio devices, which are not limited in the embodiments of the present application.
It should be understood that the above embodiments take the process of capturing video by a user as an example, and describe a possible implementation process of capturing, by an electronic device, audio corresponding to a video frame in the process of capturing the video frame. In the process of acquiring the audio, the electronic equipment can acquire the audio of the target sound source by using audio acquisition devices of different types, different numbers and different sources.
Specifically, the process of collecting the audio frequency supports the manual selection of the audio frequency collecting devices with different types, different numbers and different sources by a user, and the electronic equipment can also intelligently match the audio frequency collecting devices with different types, different numbers and different sources for the user according to different scenes and/or reference indexes, and the intelligent switching of the audio frequency collecting devices is realized in the radio receiving process, so that the interesting audio multimedia file is manufactured.
Optionally, the method for collecting audio provided in the embodiment of the present application may be applied to a scene for recording video, a scene for making an audio file, and any scene that needs to be picked up by using an audio collecting device.
For example, a user may use an application installed in an electronic device such as a mobile phone, a tablet, a smart screen, etc. and having a recording function, collect audio through an audio collection device such as a microphone of the electronic device, etc. to make an audio-class multimedia file, where the audio file does not need to be associated with a video frame.
It should be understood that the method of the embodiment of the present application may be used in the process of making an audio-type multimedia file, where a user may select different types, different numbers, and different sources of audio acquisition devices during the process of recording an audio file, and support the user to adjust the audio gain corresponding to the audio acquisition devices, so as to control the audio acquisition process, thereby making an audio multimedia file with good sound receiving effect and full interest.
Or, the method for collecting audio provided in the embodiment of the present application may be further applied to all scenes where sound is received by using the microphone of the electronic device, for example, in a process of answering a phone call, where the microphone of the mobile phone collects the sound of the user, or may be applied to the method in the embodiment of the present application, which is not limited in this embodiment of the present application.
In another possible implementation manner, the application further provides a method for acquiring audio, which is applied to an electronic device including a display screen, and the method includes: displaying a first interface, wherein the first interface comprises one or more options, the one or more options correspond to n audio acquisition devices, and n is an integer greater than or equal to 1; detecting a first operation acting on a first option, determining an audio acquisition device associated with the first option as a target audio acquisition device in response to the first operation, and acquiring the audio of a target sound source by using the target audio acquisition device; or, from the n audio collection devices, automatically matching the target audio collection device, and collecting the audio of the target sound source by using the target audio collection device.
Through the method, the embodiment of the application can also support the user to manually adjust the sound receiving gain parameters and the like of the audio collecting device, for example, the corresponding sound receiving range, sound intensity, sound effect and the like in the sound receiving process are adjusted, so that the user can better select the audio collecting device and adjust the sound receiving gain parameters according to different requirements, and the high-quality interesting sound multimedia file can be obtained more easily.
The embodiment of the application can also support the adjustment of the radio gain parameters by the forward adjustment or the reverse adjustment of the user, has simple operation process, is suitable for different user groups, and meets the use requirements of different users.
A second aspect provides an electronic device comprising: a display screen; one or more audio acquisition devices, one or more processors; one or more memories; a module in which a plurality of application programs are installed; the memory stores one or more programs that, when executed by the processor, cause the electronic device to perform the steps of: displaying an image preview interface for video recording, wherein the image preview interface comprises a first control and a second control, the first control is used for controlling audio in the video recording process, and the second control is used for starting or suspending the video recording process; detecting a first operation acting on the first control, and responding to the first operation, and displaying one or more options on the image preview interface, wherein the one or more options correspond to n audio acquisition devices, and n is an integer greater than or equal to 1; detecting a second operation acting on a first option, and responding to the second operation, determining an audio acquisition device corresponding to the first option as a target audio acquisition device, or automatically matching the target audio acquisition device from the n audio acquisition devices; detecting a third operation acting on the second control, starting to record video in response to the third operation, and acquiring the audio of the target sound source by using the target audio acquisition device in the process of recording video.
With reference to the second aspect, in certain implementations of the second aspect, one or more audio capture devices configured on the electronic device are included in the n audio capture devices; and/or one or more audio acquisition devices of other devices which establish a connection relationship with the electronic device are included in the n audio acquisition devices, and the connection relationship includes wired connection or wireless connection.
With reference to the second aspect and the foregoing implementation manner, in certain implementation manners of the second aspect, the one or more programs, when executed by the processor, cause the electronic device to perform the following steps:
according to the position of the target sound source, determining an audio acquisition device closest to the target sound source among the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the audio acquisition range corresponding to each audio acquisition device, determining the audio acquisition device with the largest audio acquisition range among the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the position of the target sound source and the audio acquisition range corresponding to each audio acquisition device, determining the audio acquisition devices of which the audio acquisition ranges cover the target sound source as the target audio acquisition devices; and/or the number of the groups of groups,
According to the use frequency of the user, determining the audio acquisition device with the highest use frequency of the user in the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the use record of the user, the audio acquisition device which is used by the user for the last time in the n audio acquisition devices is automatically determined to be the target audio acquisition device; and/or the number of the groups of groups,
according to the corresponding performance of each audio acquisition device, determining the audio acquisition device with the highest performance in the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the mode that each audio acquisition device is connected with the electronic equipment, the audio acquisition devices which are connected with the electronic equipment in a wired mode in the n audio acquisition devices are determined to be the target audio acquisition devices; and/or the number of the groups of groups,
according to the installation position of a camera used for collecting video pictures in the video recording process, automatically matching the target audio collecting devices from the n audio collecting devices; and/or the number of the groups of groups,
according to the source of a camera used for collecting video pictures in the video recording process, the target audio collecting devices are automatically matched from the n audio collecting devices.
With reference to the second aspect and the foregoing implementation manner, in some implementation manners of the second aspect, in a video recording process, when one or more cameras used for capturing video frames originate from the electronic device and one or more audio capturing devices configured on the electronic device are included in the n audio capturing devices, a preset correspondence is provided between the one or more cameras of the electronic device and the one or more audio capturing devices configured on the electronic device, and when the one or more programs are executed by the processor, the electronic device is caused to perform the following steps:
and according to the corresponding relation between one or more cameras used for acquiring video pictures and the preset, automatically matching the target audio acquisition device from one or more audio acquisition devices configured on the electronic equipment.
With reference to the second aspect and the foregoing implementation manner, in some implementation manners of the second aspect, in a video recording process, when one or more cameras used for capturing video frames originate from other devices that establish a connection with the electronic device, and one or more audio capturing devices of the n audio capturing devices include one or more other devices that establish a connection with the electronic device, when the one or more programs are executed by the processor, the electronic device is caused to perform the following steps: and determining an audio acquisition device of other equipment with one or more camera sources used for acquiring video pictures as the target audio acquisition device.
With reference to the second aspect and the foregoing implementation manner, in some implementation manners of the second aspect, when one or more audio capturing devices of other devices that establish a connection relationship with the electronic device are included in the n audio capturing devices, the one or more programs when executed by the processor, cause the electronic device to further perform the following steps: detecting a fourth operation acting on the second option, determining one or more devices in connection with the electronic device in response to the fourth operation, and displaying controls including the one or more devices; detecting a fifth operation acting on a control corresponding to a target device, and responding to the fifth operation, determining one or more audio acquisition devices of the target device as the target audio acquisition device, and acquiring the audio of a target sound source by using the target audio acquisition device in the process of recording video, wherein the connection relation comprises one or more wireless connection relations of Bluetooth connection, wireless local area network Wi-Fi connection, infrared connection and near field communication NFC connection.
With reference to the second aspect and the foregoing implementation manner, in some implementation manners of the second aspect, the one or more options include a preset third option, where the third option is displayed in a selected state, and when the first operation acting on the first control is not detected, an audio capture device corresponding to the third option is determined to be the target audio capture device.
With reference to the second aspect and the foregoing implementation manners, in some implementation manners of the second aspect, after determining the target audio capturing device, the electronic device may further perform the following steps: and displaying prompt information, wherein the prompt information is used for prompting the name or source of the target audio acquisition device used by the electronic equipment currently by a user.
With reference to the second aspect and the foregoing implementation manners, in some implementation manners of the second aspect, after determining the target audio capturing device, the electronic device may further perform the following steps: displaying one or more controls, wherein each control of the one or more controls is used for controlling the working parameters of the target audio acquisition device, and the working parameters comprise one or more of an audio acquisition range, audio acquisition intensity and audio playing effect in the working process of the target audio acquisition device; detecting a sixth operation on the one or more controls, and adjusting an operating parameter of the target audio acquisition device in response to the sixth operation.
A third aspect provides a graphical user interface system on an electronic device with a display screen, one or more audio acquisition devices, one or more memories, and one or more processors to execute one or more computer programs stored in the one or more memories, the graphical user interface system comprising a graphical user interface displayed by the electronic device when performing the methods of any of the first and second aspects.
A fourth aspect provides an apparatus for inclusion in an electronic device, the apparatus having functionality to implement the behaviour of the electronic device in the method of any one of the first and second aspects. The functions can be realized by hardware, and can also be realized by executing corresponding software by hardware. The hardware or software includes one or more modules or units corresponding to the functions described above. Such as a display module or unit, a detection module or unit, a processing module or unit, etc.
A fifth aspect provides a computer readable storage medium storing computer instructions that, when run on an electronic device, cause the electronic device to perform the method of any one of the first and first aspects, and the method of any one of the second and second aspects, to collect audio.
A sixth aspect provides a computer program product for causing an electronic device to perform the method of capturing audio of the first aspect or any of the first aspects described above when the computer program product is run on the electronic device.
Drawings
Fig. 1 is a schematic structural diagram of an example of an electronic device according to an embodiment of the present application.
Fig. 2 is a software configuration block diagram of an example of an electronic device according to an embodiment of the present application.
Fig. 3 is a schematic diagram of a user interface for capturing video by a user according to an embodiment of the present application.
Fig. 4 is a schematic diagram of an example of radio equipment available in a home scenario for a mobile phone according to an embodiment of the present application.
Fig. 5 is a schematic view illustrating an embodiment of an audio reception angle adjustment apparatus according to an embodiment of the present application.
Fig. 6 is a schematic diagram of a user interface for capturing video by another user according to an embodiment of the present application.
Fig. 7 is a schematic diagram of an example of a plurality of electronic devices that may be included in a home scenario.
Fig. 8 is a schematic diagram of a user interface for capturing video by a user according to another embodiment of the present application.
Fig. 9 is a schematic diagram of a user interface for capturing video by a user according to another embodiment of the present application.
Fig. 10 is a schematic diagram of a user interface for capturing video by a user according to another embodiment of the present application.
Fig. 11 is a schematic flowchart of an example of a method for capturing audio according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described below with reference to the drawings in the embodiments of the present application.
The method for acquiring audio provided by the embodiment of the application can be applied to any electronic device with shooting function and radio function, such as a mobile phone, a tablet computer, a wearable device, a vehicle-mounted device, an augmented reality (augmented reality, AR)/Virtual Reality (VR) device, a notebook computer, an ultra-mobile personal computer (UMPC), a netbook, a personal digital assistant (personal digital assistant, PDA) and the like, and the specific type of the electronic device is not limited.
Fig. 1 is a schematic structural diagram of an example of an electronic device according to an embodiment of the present application.
It should be understood that the electronic device of the embodiments of the present application may include some or all of the structures of the electronic device 100 shown in fig. 1, or may further include other structures not shown, which are not limited in this embodiment of the present application.
By way of example, the electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a universal serial bus (universal serial bus, USB) interface 130, a charge management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, keys 190, a motor 191, an indicator 192, a camera 193, a display 194, and a subscriber identity module (subscriber identification module, SIM) card interface 195, etc. The sensor module 180 may include a pressure sensor 180A, a gyro sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, a bone conduction sensor 180M, and the like.
It is to be understood that the structure illustrated in the embodiments of the present application does not constitute a specific limitation on the electronic device 100. In other embodiments of the present application, electronic device 100 may include more or fewer components than shown, or certain components may be combined, or certain components may be split, or different arrangements of components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
The processor 110 may include one or more processing units, such as: the processor 110 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a memory, a video codec, a digital signal processor (digital signal processor, DSP), a baseband processor, and/or a neural network processor (neural-network processing unit, NPU), etc. Wherein the different processing units may be separate devices or may be integrated in one or more processors.
The controller may be a neural hub and a command center of the electronic device 100, among others. The controller can generate operation control signals according to the instruction operation codes and the time sequence signals to finish the control of instruction fetching and instruction execution.
A memory may also be provided in the processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that the processor 110 has just used or recycled. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 110 is reduced, thereby improving the efficiency of the system.
In the embodiment of the present application, the processor 110 of the electronic device 100 stores a computer program or instructions corresponding to the method for implementing the audio capturing.
Illustratively, the processor 110 of the electronic device 100 stores a method of controlling audio during the production of an audio multimedia file, in particular: in the process of producing the audio collection of the audio multimedia file, firstly, the embodiment of the application provides a method for manually selecting the audio receiving device for a user, and the user can select different kinds of audio receiving devices with different numbers and different sources to realize the control of audio collection. For example, a user may manually select different sound pickup devices on one electronic device, and may also manually select sound pickup devices using different devices to pick up audio. According to the method, a user can manually select or switch different sound receiving devices according to different shooting requirements, shooting scenes and the like, so that different sound receiving devices are used for collecting audio, and the quality of audio collection is improved.
In addition, in the process of producing the audio-collecting multimedia file, the method for collecting the audio frequency can also support the user to manually adjust the sound-collecting gain parameters of the sound-collecting device, for example, adjust the corresponding sound-collecting range, sound intensity, sound effect and the like in the sound-collecting process, so that the user can better select the sound-collecting device and adjust the sound-collecting gain parameters according to different requirements, and can more easily obtain the high-quality and full-interesting audio multimedia file.
In addition, in the process of collecting audio frequency in the manufacture of the audio-sound multimedia file, the electronic equipment can also automatically match one or more audio-receiving devices for the current scene according to a plurality of reference indexes such as the distance between the audio-receiving device and the sound source, the positions of the audio-receiving devices and the like, or can switch microphones of different equipment and the like in the audio-receiving process to collect audio frequency. The process can improve the sound reception quality, collect clearer audio for users, improve the sound reception effect and facilitate the production of better-quality sound multimedia files.
In some embodiments, the processor 110 may include one or more interfaces. The interfaces may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous receiver transmitter (universal asynchronous receiver/transmitter, UART) interface, a mobile industry processor interface (mobile industry processor interface, MIPI), a general-purpose input/output (GPIO) interface, a subscriber identity module (subscriber identity module, SIM) interface, and/or a universal serial bus (universal serial bus, USB) interface, among others.
The I2C interface is a bi-directional synchronous serial bus comprising a serial data line (SDA) and a serial clock line (derail clock line, SCL). In some embodiments, the processor 110 may contain multiple sets of I2C buses. The processor 110 may be coupled to the touch sensor 180K, charger, flash, camera 193, etc., respectively, through different I2C bus interfaces. For example: the processor 110 may be coupled to the touch sensor 180K through an I2C interface, such that the processor 110 communicates with the touch sensor 180K through an I2C bus interface to implement a touch function of the electronic device 100.
The I2S interface may be used for audio communication. PCM interfaces may also be used for audio communication to sample, quantize and encode analog signals. The UART interface is a universal serial data bus for asynchronous communications. The bus may be a bi-directional communication bus. It converts the data to be transmitted between serial communication and parallel communication. In some embodiments, a UART interface is typically used to connect the processor 110 with the wireless communication module 160. For example: the processor 110 communicates with a bluetooth module in the wireless communication module 160 through a UART interface to implement a bluetooth function. In some embodiments, the audio module 170 may transmit an audio signal to the wireless communication module 160 through a UART interface, to implement a function of playing music through a bluetooth headset.
The MIPI interface may be used to connect the processor 110 to peripheral devices such as a display 194, a camera 193, and the like. The MIPI interfaces include camera serial interfaces (camera serial interface, CSI), display serial interfaces (display serial interface, DSI), and the like. In some embodiments, processor 110 and camera 193 communicate through a CSI interface to implement the photographing functions of electronic device 100. The processor 110 and the display 194 communicate via a DSI interface to implement the display functionality of the electronic device 100.
The GPIO interface may be configured by software. The USB interface 130 is an interface conforming to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, or the like. The USB interface 130 may be used to connect a charger to charge the electronic device 100, and may also be used to transfer data between the electronic device 100 and a peripheral device. And can also be used for connecting with a headset, and playing audio through the headset. The interface may also be used to connect other electronic devices, such as AR devices, etc.
It should be understood that the interfacing relationship between the modules illustrated in the embodiments of the present application is only illustrative, and does not limit the structure of the electronic device 100. In other embodiments of the present application, the electronic device 100 may also use different interfacing manners, or a combination of multiple interfacing manners in the foregoing embodiments.
The charge management module 140 is configured to receive a charge input from a charger. The charger can be a wireless charger or a wired charger. In some wired charging embodiments, the charge management module 140 may receive a charging input of a wired charger through the USB interface 130. In some wireless charging embodiments, the charge management module 140 may receive wireless charging input through a wireless charging coil of the electronic device 100. The charging management module 140 may also supply power to the electronic device through the power management module 141 while charging the battery 142.
The power management module 141 is used for connecting the battery 142, and the charge management module 140 and the processor 110. The power management module 141 receives input from the battery 142 and/or the charge management module 140 and provides power to the processor 110, the internal memory 121, the external memory, the display 194, the camera 193, the wireless communication module 160, and the like. The power management module 141 may also be configured to monitor battery capacity, battery cycle number, battery health (leakage, impedance) and other parameters. In other embodiments, the power management module 141 may also be provided in the processor 110. In other embodiments, the power management module 141 and the charge management module 140 may be disposed in the same device.
The wireless communication function of the electronic device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. The mobile communication module 150 may provide a solution for wireless communication including 2G/3G/4G/5G/6G, etc. applied on the electronic device 100.
The modem processor may include a modulator and a demodulator. The modulator is used for modulating the low-frequency baseband signal to be transmitted into a medium-high frequency signal. The demodulator is used for demodulating the received electromagnetic wave signal into a low-frequency baseband signal. The demodulator then transmits the demodulated low frequency baseband signal to the baseband processor for processing. The low frequency baseband signal is processed by the baseband processor and then transferred to the application processor.
The application processor outputs sound signals through an audio device (not limited to the speaker 170A, the receiver 170B, etc.), or displays images or video through the display screen 194. In some embodiments, the modem processor may be a stand-alone device. In other embodiments, the modem processor may be provided in the same device as the mobile communication module 150 or other functional module, independent of the processor 110.
The wireless communication module 160 may provide solutions for wireless communication including wireless local area network (wireless local area networks, WLAN) (e.g., wireless fidelity (wireless fidelity, wi-Fi) network), bluetooth (BT), global navigation satellite system (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), near field wireless communication technology (near field communication, NFC), infrared technology (IR), etc., as applied to the electronic device 100.
In some embodiments, antenna 1 and mobile communication module 150 of electronic device 100 are coupled, and antenna 2 and wireless communication module 160 are coupled, such that electronic device 100 may communicate with a network and other devices through wireless communication techniques.
The electronic device 100 implements display functions through a GPU, a display screen 194, an application processor, and the like. The GPU is a microprocessor for image processing, and is connected to the display 194 and the application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. Processor 110 may include one or more GPUs that execute program instructions to generate or change display information.
The display screen 194 is used to display images, videos, and the like. The display 194 includes a display panel. In some embodiments, the electronic device 100 may include 1 or N display screens 194, N being a positive integer greater than 1.
For example, in the embodiment of the present application, the electronic device 100 cooperates with modules such as the GPU, the display screen 194, and the application processor, so as to display images, application interfaces, controls, icons, windows, and the like on the display screen of the electronic device 100, thereby implementing the display function of the electronic device 100.
The electronic device 100 may implement photographing functions through an ISP, a camera 193, a video codec, a GPU, a display screen 194, an application processor, and the like.
In this embodiment of the present application, the electronic device 100 may start a shooting function of the camera by running the camera application, and display a shooting preview interface on the camera application interface, where the shooting preview interface may display a picture collected by the camera, and one or more controls such as a shooting shutter control, and a user may click on the shooting shutter control to shoot a photo, record a video clip, and save the photo and the video clip in the local library application, which will not be described herein.
The ISP is used to process data fed back by the camera 193. For example, when photographing, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electric signal, and the camera photosensitive element transmits the electric signal to the ISP for processing and is converted into an image visible to naked eyes. ISP can also optimize the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature, etc. of the photographed scene. In some embodiments, the ISP may be provided in the camera 193.
For example, in the embodiment of the present application, the electronic device 100 may adjust one or more shooting parameters, such as one or more of a beauty parameter, a filter color, an exposure degree, and a sensitivity, through the ISP, which is not limited in the embodiment of the present application.
The camera 193 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image onto the photosensitive element. The photosensitive element converts the optical signal into an electrical signal, which is then transferred to the ISP to be converted into a digital image signal. The ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into an image signal in a standard RGB, YUV, or the like format. In some embodiments, electronic device 100 may include 1 or N cameras 193, N being a positive integer greater than 1.
Illustratively, in the embodiment of the present application, the camera 193 may include a front camera and a rear camera, or include one or more types of cameras such as a wide-angle camera, a macro camera, a depth camera, a portrait camera, etc., and the embodiment of the present application does not limit the number and types of cameras 193 that may be included in the electronic device 100.
The digital signal processor is used for processing digital signals, and can process other digital signals besides digital image signals. For example, when the electronic device 100 selects a frequency bin, the digital signal processor is used to fourier transform the frequency bin energy, or the like.
Video codecs are used to compress or decompress digital video. The electronic device 100 may support one or more video codecs. In this way, the electronic device 100 may play or record video in a variety of encoding formats, such as: dynamic picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4, etc.
The NPU is a neural-network (NN) computing processor, and can rapidly process input information by referencing a biological neural network structure, for example, referencing a transmission mode between human brain neurons, and can also continuously perform self-learning. Applications such as intelligent awareness of the electronic device 100 may be implemented through the NPU, for example: image recognition, face recognition, speech recognition, text understanding, etc.
The external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, to enable expansion of the memory capabilities of the electronic device 100. The external memory card communicates with the processor 110 through an external memory interface 120 to implement data storage functions. For example, files such as music, video, etc. are stored in an external memory card.
The internal memory 121 may be used to store computer executable program code including instructions. The processor 110 executes various functional applications of the electronic device 100 and data processing by executing instructions stored in the internal memory 121. The internal memory 121 may include a storage program area and a storage data area. The storage program area may store an application program (such as a sound playing function, an image playing function, etc.) required for at least one function of the operating system, etc. The storage data area may store data created during use of the electronic device 100 (e.g., audio data, phonebook, etc.), and so on. In addition, the internal memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (universal flash storage, UFS), and the like.
The electronic device 100 may implement audio functions through an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, an application processor, and the like. Such as music playing, recording, etc.
For example, in the embodiment of the present application, the microphone 170C of the electronic device may collect audio, for example, during a process that a user records video using a camera application of the electronic device, the electronic device may collect image data through a camera, collect audio data through the microphone 170C, and correlate the image data with the audio data to make a sound multimedia file such as video.
In addition, the earphone interface 170D may be externally connected to an earphone device, and the earphone device also has a microphone, so that the earphone device may collect audio, which is not described herein.
The pressure sensor 180A is used to sense a pressure signal, and may convert the pressure signal into an electrical signal.
The gyro sensor 180B may be used to determine a motion gesture of the electronic device 100.
The air pressure sensor 180C is used to measure air pressure. In some embodiments, electronic device 100 calculates altitude from barometric pressure values measured by barometric pressure sensor 180C, aiding in positioning and navigation.
The magnetic sensor 180D includes a hall sensor.
The acceleration sensor 180E may detect the magnitude of acceleration of the electronic device 100 in various directions (typically three axes). The magnitude and direction of gravity may be detected when the electronic device 100 is stationary. The electronic equipment gesture recognition method can also be used for recognizing the gesture of the electronic equipment, and is applied to horizontal and vertical screen switching, pedometers and other applications.
A distance sensor 180F for measuring a distance. The electronic device 100 may measure the distance by infrared or laser. In some embodiments, the electronic device 100 may range using the distance sensor 180F to achieve quick focus.
Illustratively, in embodiments of the present application, the electronic device may detect a distance between itself and the other distributed device or devices via the distance sensor 180F; in addition, the electronic device may also detect the position of the sound source object, the distance thereof, and the like through the distance sensor 180F, which will not be described in detail.
The proximity light sensor 180G may include, for example, a Light Emitting Diode (LED) and a light detector, such as a photodiode. The light emitting diode may be an infrared light emitting diode.
The ambient light sensor 180L is used to sense ambient light level.
The fingerprint sensor 180H is used to collect a fingerprint. The electronic device 100 may utilize the collected fingerprint feature to unlock the fingerprint, access the application lock, photograph the fingerprint, answer the incoming call, etc.
The temperature sensor 180J is for detecting temperature. In some embodiments, the electronic device 100 performs a temperature processing strategy using the temperature detected by the temperature sensor 180J.
The touch sensor 180K, also referred to as a "touch panel". The touch sensor 180K may be disposed on the display screen 194, and the touch sensor 180K and the display screen 194 form a touch screen, which is also called a "touch screen". The touch sensor 180K is for detecting a touch operation acting thereon or thereabout. The touch sensor may communicate the detected touch operation to the application processor to determine the touch event type. Visual output related to touch operations may be provided through the display 194. In other embodiments, the touch sensor 180K may also be disposed on the surface of the electronic device 100 at a different location than the display 194.
For example, in the embodiment of the present application, the touch sensor 180K may detect a touch, a click, or the like of a user, generate an operation event, transmit the operation event to the processor 110, and determine, by the processor 110, that the operation event includes the coordinates of the touch point, the touch state, and other relevant parameters, and respond according to the operation event. For example, the user clicks the shooting operation of the shooting shutter control, and the operation event may include information of the shooting shutter control, user touch operation information, and the like, which are transmitted to the processor 110, and the processor 110 completes the shooting operation according to the picture acquired by the camera, and obtains the picture.
The bone conduction sensor 180M may acquire a vibration signal.
The keys 190 include a power-on key, a volume key, etc. The keys 190 may be mechanical keys. Or may be a touch key. The electronic device 100 may receive key inputs, generating key signal inputs related to user settings and function controls of the electronic device 100.
The motor 191 may generate a vibration cue. The motor 191 may be used for incoming call vibration alerting as well as for touch vibration feedback.
The indicator 192 may be an indicator light, may be used to indicate a state of charge, a change in charge, a message indicating a missed call, a notification, etc.
The SIM card interface 195 is used to connect a SIM card. The SIM card may be inserted into the SIM card interface 195, or removed from the SIM card interface 195 to enable contact and separation with the electronic device 100. The electronic device 100 may support 1 or N SIM card interfaces, N being a positive integer greater than 1.
The software system of the electronic device 100 may employ a layered architecture, an event driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture, for exampleSystem architecture, hong Meng system architecture, etc.
Exemplary, embodiments of the present application are in a layered architectureThe system is an example illustrating the software architecture of the electronic device 100.
Fig. 2 is a software configuration block diagram of an example of an electronic device according to an embodiment of the present application.
Alternatively, as shown in FIG. 2The hierarchical architecture of the system may include several layers, each layer having distinct roles and branches, and the layers communicating via software interfaces.
The layered architecture divides the software into several layers, each with distinct roles and branches. The layers communicate with each other through a software interface. In some embodiments, it willThe system is divided into four layers, namely an application program layer, an application program framework layer and a An Zhuoyun line (++)>runtimes) and system libraries, and kernel layers.
The application layer may include a series of application packages.
By way of example, as shown in FIG. 2, the application package may include applications for cameras, photo albums, music, settings, and the like. In the embodiment of the application, a user may take a sound photo, a video segment and other sound multimedia files through a camera application, and the taken sound photo, video segment and other sound multimedia files may be saved into an album (gallery) application.
Optionally, the "camera application" in the embodiment of the present application may be a camera application configured by factory of the electronic device, and may include different shooting modes such as shooting, video recording, and dual-view video recording, which is not limited in the embodiment of the present application.
It should be understood that the camera application in the embodiments of the present application may be associated with the functions of capturing audio and controlling audio during the process of making an audio multimedia file in the embodiments of the present application.
The application framework layer provides an application programming interface (application programming interface, API) and programming framework for application programs of the application layer. The application framework layer includes a number of predefined functions. As shown in fig. 2, the application framework layer may include a window manager, a content provider, a view system, a resource manager, a notification manager, etc., as well as a video service module, an audio service module, etc., according to embodiments of the present application.
Wherein the window manager is used for managing window programs. The window manager can acquire the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like.
The content provider is used to store and retrieve data and make such data accessible to applications. The data may include video, images, audio, calls made and received, browsing history and bookmarks, phonebooks, etc.
The view system includes visual controls, such as controls to display text, controls to display pictures, and the like. The view system may be used to build applications. The display interface may be composed of one or more views. For example, a display including an icon of a camera application on a cell phone interface may include a view displaying text and a view displaying a picture.
The resource manager provides various resources for the application program, such as localization strings, icons, pictures, layout files, video files, and the like.
The notification manager allows the application to display notification information in a status bar, can be used to communicate notification type messages, can automatically disappear after a short dwell, and does not require user interaction. Such as notification manager is used to inform that the download is complete, message alerts, etc. The notification manager may also be a notification in the form of a chart or scroll bar text that appears on the system top status bar, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, a text message is prompted in a status bar, a prompt tone is emitted, the electronic device vibrates, and an indicator light blinks, etc.
The video service module of the embodiment of the application may specifically include an image acquisition module and a synchronization control module, so as to implement the function of recording video of the embodiment of the application.
Specifically, the image acquisition module may be configured to acquire image data corresponding to a video frame acquired by the camera. Alternatively, the video frames may include a single-view frame, a double-view frame, a multi-view frame, and the like, in other words, image data corresponding to the video frames may originate from one camera or a plurality of cameras.
The synchronous control module can be used for guaranteeing the synchronization of the video picture of the video service module and the audio data of the audio service module and guaranteeing the matching of the video picture and the audio data. And when the user carries out forward regulation or reverse regulation on the sound reception gain, the synchronous control module can realize forward synchronization and reverse synchronization of the sound reception gain of the video picture in response to the audio according to the operation of the user.
The audio service module of the embodiment of the application may specifically include an equipment control module, a direction control module and a gain control module, so as to implement the audio acquisition process of the embodiment of the application.
Specifically, the device control module may determine, according to a source of the radio device selected by the user on the interface of the electronic device, whether the radio device collecting the audio is a distributed device or the electronic device itself, and determine that the radio device collects the audio, the bluetooth radio, or the Wi-Fi radio is used.
The direction control module can determine that one or more radio receiving devices originate from the electronic device according to radio receiving options selected by a user on an interface of the electronic device, and control the direction, angle or radio receiving range of one or more radio receiving devices to collect audio.
The gain control module may be configured to determine a current sound reception range (angle), audio intensity, audio playback sound effect, etc. of one or more sound reception devices. After the user adjusts the sound reception gain parameters, the gain control module can record the related parameters such as sound reception angle, strength, sound effect and the like, collect the audio according to the gain parameters and perform corresponding processing on the collected audio.
For example, taking a mobile phone as an example, if a user selects any one of omni-directional radio, front radio, rear radio, wired radio, bluetooth radio on the mobile phone, the device control module of the mobile phone may further determine whether to use its own microphone or the microphones of other distributed devices, and further determine which microphone is used, e.g., one or more of the top microphone, the bottom microphone, or the back microphone; the direction control module of the mobile phone can determine the direction, angle or sound receiving range of the audio collected by the one or more microphones, and then accurately collect the audio. If the sound processing algorithm is overlapped, the collected audio can be correspondingly processed, and the details are not repeated.
Android runtimes include core libraries and virtual machines. Android run time is responsible for scheduling and management of the Android system.
The core library consists of two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. The virtual machine executes java files of the application program layer and the application program framework layer as binary files. The virtual machine is used for executing the functions of object life cycle management, stack management, thread management, security and exception management, garbage collection and the like.
The system library may include a plurality of functional modules. For example: surface manager (surface manager), media library (media library), three-dimensional graphics processing library (e.g., openGL ES), two-dimensional graphics engine (e.g., SGL), image processing library, etc.
The surface manager is used to manage the display subsystem and provides a fusion of the two-dimensional and three-dimensional layers for the plurality of applications.
Media libraries support a variety of commonly used audio, video format playback and recording, still image files, and the like. The media library may support a variety of audio and video encoding formats, such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, etc.
The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like.
A two-dimensional graphics engine is a drawing engine that draws two-dimensional drawings.
The image processing library may provide analysis for various image data, provide various image processing algorithms, etc., for example, may provide processing such as image cutting, image fusion, image blurring, image sharpening, etc., and will not be described in detail herein.
For example, in the embodiment of the present application, when a certain application interface is displayed on the display screen of the electronic device, image rendering, composition, layer processing, and the like of the application interface may be sequentially performed based on the view system, the two-dimensional graphics engine, the three-dimensional graphics processing library, the image processing library, and the like, and finally the application interface, for example, a shooting preview interface of a camera application, and the like, is displayed on the display screen.
The kernel layer is a layer between hardware and software. The inner core layer at least comprises a display driver, a camera driver, an audio driver and a sensor driver.
In the embodiment of the present application, the user uses the camera application to record the video, and through the cooperation of the plurality of software modules and the hardware modules, for example, a shooting preview interface is displayed for the user, the operations of selecting the radio receiving device and adjusting the radio receiving gain parameter performed on the interface by the user are detected, the display of the interface is adjusted or various operations are performed in response to different operations of the user, and the audio is collected by using different radio receiving devices, which is not repeated herein.
For easy understanding, the following embodiments of the present application will take a mobile phone having the structure shown in fig. 1 and fig. 2 as an example, and specifically describe a process of capturing an audio multimedia file provided in the embodiments of the present application in conjunction with the accompanying drawings and application scenarios.
Fig. 3 is a schematic diagram of an example of a graphical user interface (graphical user interface, GUI) for capturing video by a user according to an embodiment of the present application.
By way of example, fig. 3 (a) illustrates a main interface 301 displayed by the handset in an unlocked mode, where the main interface 301 may include a top status bar area that may display one or more of an operator identification, a signal identification of a wireless local area network (wireless local area networks, WLAN) (e.g., a wireless fidelity (wireless fidelity, wi-Fi) network), a power identification of the handset, and a current time, and an interface display area that displays a weather clock component and a plurality of applications (apps) installed on the handset. Alternatively, the application may include a phone, information, gallery, camera, etc. It should be understood that the main interface 301 may further include other more content or display more applications, and the content displayed by the main interface is not limited in this embodiment of the present application.
The user performs an operation as shown in fig. 3 (a), clicks an icon of the camera application, and in response to the user's operation, the mobile phone may display a main interface of the camera application, which main interface 302 of the camera application may also be referred to as a "shooting preview interface", in which different images, pictures, and the presented images, pictures are referred to as a "preview image" or a "preview picture", as shown in fig. 3 (b).
For example, as shown in fig. 3 (b), the photographing preview interface 302 may include a bottom navigation area, a display area of a middle preview screen, and a top menu area. Optionally, on the shooting preview interface 302, the bottom navigation area displays a shooting shutter control 31, a local album control 32, and a camera switch control 33, and further displays a plurality of shooting mode controls such as night scenes, portrait, shooting, professional, video, and the like.
The display area of the middle preview picture is used for displaying the image or picture captured by the current mobile phone camera.
The top menu region may display one or more controls, such as a flash automatic control 34, a shooting effect control 35, and a camera setting control 36, and the number or kind of controls displayed in each region is not limited in the embodiment of the present application. For simplicity, the different functions of each control are not described in detail in the embodiments of the present application.
It should be understood that, in the embodiment of the present application, any interface corresponding to a shooting mode may be referred to as a "shooting preview interface", and different numbers and/or types of controls may be displayed on the shooting preview interfaces corresponding to different shooting modes, which is not limited in the embodiment of the present application.
Assuming that the user has switched the shooting mode of the shooting preview interface 302 to a professional mode, accordingly, a "black small triangle icon" for locating the current shooting mode may point to the "professional" mode control.
Alternatively, when the user switches the current shooting mode to the professional mode, the mobile phone may display one or more controls, icons, parameters, or the like on the shooting preview interface 302 in response to the user's operation.
Exemplary, an international organization for standardization (International Organizationfor Standardization, ISO) sensitivity (abbreviated as ISO) parameter control, a shutter speed (S) control, an exposure compensation (EV) control, an Auto Focus (AF) control, an auto white balance (automatic white balance, AWB) control, and the like, which are sequentially shown in (b) diagram in fig. 3.
The function of the partial control in the professional mode and the shooting parameters are briefly described below in connection with the shooting preview interface 302 illustrated in table 1 and (b) in fig. 3.
TABLE 1
In a possible implementation manner, in an embodiment of the present application, a microphone (micphone controller, MIC) control 37 shown by a dashed box may be added to the user in a control display area of the shooting preview interface 302 corresponding to the professional mode, and the user may control the audio receiving process in the process of shooting a sound multimedia file such as a video through the microphone control 37 (or a microphone icon).
It should be understood that in the embodiments of the present application, a device capable of capturing audio is referred to as "sound capturing device", "audio capturing device", or "sound capturing device", such as a microphone, an external earphone, or the like, which is not limited in the embodiments of the present application.
The user may perform the operation shown in (b) of fig. 3, click on the microphone control 37, and in response to the user's operation, the handset may display the interface 303 shown in (c) of fig. 3.
Alternatively, the interface 303 may be understood as a "sound receiving device selection interface", and the sound receiving device selection interface 303 may display one or more options related to sound receiving devices available on the mobile phone, through which a user may select a sound receiving device when capturing a sound multimedia file, or switch between different sound receiving devices.
Fig. 4 is a schematic diagram of an example of radio equipment available in a home scenario for a mobile phone according to an embodiment of the present application. It should be appreciated that each electronic device, such as the cell phone 100, may associate or use one or more sound receiving devices to collect audio from the external environment during use.
Optionally, in the embodiment of the present application, the one or more radio receiving devices may include a radio receiving device installed on the mobile phone itself, and may further include a radio receiving device installed on another device that can be called by the mobile phone, which is not limited in the embodiment of the present application.
Illustratively, as shown in fig. 4, the sound pickup device of the mobile phone 100 may include a plurality of sound pickup devices such as a top microphone 10, a bottom microphone 20, and a back microphone 30 as part of the structure of the mobile phone itself, each of which may be mounted at a different location of the mobile phone and each of which has a different audio pickup range (otherwise referred to as a "sound pickup range").
In addition, in a home scenario, the mobile phone 100 may establish a distributed connection with the tablet 200, the smart screen 300, the speaker 500, etc., which is not limited to a bluetooth connection or a Wi-Fi connection, etc., which is not limited in the embodiments of the present application. Then, the tablet 200, the smart screen 300, and the speaker 500 are used as the distributed devices of the mobile phone 100, and the mobile phone 100 can call the microphone of the distributed devices such as the tablet 200, the smart screen 300, and the speaker 500 as the sound receiving device.
In addition, the mobile phone 100 may be connected to other hardware devices with an audio collection function through a wired connection, for example, the mobile phone 100 may be connected to an earphone 400, a microphone, or other radio receiving devices with an audio collection function through a data line, and then the earphone 400, the microphone, or the like connected to the data line may also be used as a radio receiving device of the mobile phone 100.
It should be understood that in the process of capturing audio multimedia files such as video, a user may manually select or switch different sound receiving devices according to different capturing requirements, capturing scenes, etc., so as to collect audio using the different sound receiving devices.
In one possible implementation, the user may select the corresponding sound receiving device to collect audio by clicking on any one or more of the one or more options displayed by the sound receiving device selection interface 303.
Illustratively, as shown in fig. 3 (c), the one or more options displayed on the radio selection interface 303 may include: one or more of the omni-directional radio option 37-1, the front radio option 37-2, the rear radio option 37-3, the wired radio option 37-4, the bluetooth radio option 37-5, the more options 37-6, etc., which are not limited in this embodiment of the present application.
Alternatively, the user may view or select more sound receiving devices via more options 37-6. For example, the user may click on the more options 37-6 to view more radio modes or radio devices not displayed on the radio device selection interface 303, for example, after the user clicks on the more options 37-6, the user may see more radio options such as Wi-Fi radio options, which may support the user to select a microphone of a distributed device that establishes Wi-Fi connection with the mobile phone to perform radio.
Alternatively, the user may perform a sliding operation to view more sound receiving modes or sound receiving devices in an area where one or more options are displayed on the sound receiving device selection interface 303, which is not limited in the embodiment of the present application.
Alternatively, the options displayed by the radio device selection interface 303 for the user to select radio devices may be of different numbers and/or categories in different scenarios.
Alternatively, in different scenarios, the option displayed by the radio selection interface 303 for the user to select a radio may present different display states.
Illustratively, the radio device selection interface 303 may include the wired radio option 37-4 when the mobile phone and other electronic devices are wired via the data line, and the radio device selection interface 303 may not include the wired radio option 37-4 when the mobile phone is not wired to other electronic devices.
Alternatively, the radio selection interface 303 may include a bluetooth radio option 37-5 when the mobile phone establishes a bluetooth connection with other electronic devices, and the radio selection interface 303 may not display the bluetooth radio option 37-5 when the mobile phone does not turn on a bluetooth switch or establishes a bluetooth connection with other electronic devices.
Or, when the mobile phone and other electronic devices establish bluetooth connection, the radio device selection interface 303 may include a bluetooth radio option 37-5, where the bluetooth radio option 37-5 is in a working state of normal display, and can receive a click operation of the user. When the mobile phone does not turn on the bluetooth switch or establish bluetooth connection with other electronic devices, the radio device selection interface 303 may also display a bluetooth radio option 37-5, and the bluetooth radio option 37-5 is displayed in a gray inactive state, which is not limited in this embodiment of the present application.
The correspondence between each option and the sound pickup device according to the embodiment of the present application is exemplified in table 2 in connection with the sound pickup device selection interface 303 shown in fig. 3 (c) and the plurality of sound pickup devices available to the mobile phone 100 exemplified in fig. 4.
TABLE 2
It should be noted that, table 2 is only one possible example, and in other possible implementations of the embodiment of the present application, the top microphone 10 or the bottom microphone 20 of the mobile phone 100 in fig. 4 may be used as a front-mounted sound pickup device or a rear-mounted sound pickup device, and the top microphone 10 or the bottom microphone 20 may be determined as a front-mounted sound pickup device or a rear-mounted sound pickup device according to different shooting scenes and different division manners.
Optionally, during the video recording process, a certain sound receiving device may be divided into a front sound receiving device or a rear sound receiving device according to a camera for collecting video pictures, and/or a certain sound receiving device may be divided into a front sound receiving device or a rear sound receiving device according to an installation position of the sound receiving device.
Illustratively, when the mobile phone 100 is recording video, if there is no switching of the camera, the top microphone 10 may be determined to be a front-mounted sound pickup device, which is controlled by the front-mounted sound pickup option 37-2 on the sound pickup device selection interface 303, and the bottom microphone 20 may be determined to be a rear-mounted front-mounted sound pickup device, which is controlled by the rear-mounted sound pickup option 37-3 on the sound pickup device selection interface 303.
If the mobile phone 100 records a video, the user switches the camera for capturing the video picture, for example, switches from the rear camera to the front camera, and then the mounting positions of the top microphone 10 and the bottom microphone 20 can be determined to be closer to one surface of the mobile phone display screen according to the mounting position of each microphone (the top microphone 10, the bottom microphone 20 and the back microphone 30), so that both the top microphone 10 and the bottom microphone 20 can be determined to be front-mounted sound receiving devices, and both the front-mounted sound receiving devices can be controlled through the front-mounted sound receiving options 37-2 on the sound receiving device selection interface 303.
Optionally, when the user clicks the microphone control 37 on the interface 302, and in response to the user's operation, the mobile phone displays the radio selection interface 303, a certain option of the one or more options on the radio selection interface 303 may be in a default selected state, and the user may modify or switch the radio according to his or her own requirement.
Illustratively, when the handset displays the radio selection interface 303 as shown in fig. 3 (c), the omni-directional radio option 37-1 may be selected by default, i.e., if the user does not select any radio option on the radio selection interface 303, the default handset 100 will simultaneously radio the top microphone 10, the bottom microphone 20, and the back microphone 30 controlled by the omni-directional radio option 37-1. If the user desires to modify the sound pickup using only the back microphone 30, then the operation shown in fig. 3 (c) may be performed, clicking on the rear sound pickup option 37-3 on the sound pickup device selection interface 303, the embodiment of the present application is not limited to the sound pickup option selected by default.
Alternatively, the one or more options included in the radio selection interface 303 are all in a non-selected state, the user may perform an operation as shown in (c) of fig. 3, click on the rear radio selection 37-3 on the radio selection interface 303, and in response to the user operation, the mobile phone may collect audio through the back microphone 30 of the mobile phone shown in fig. 4 corresponding to the rear radio selection 37-3.
Optionally, when the mobile phone detects that the user performs the operation shown in fig. 3 (c), and clicks the rear radio option 37-3 on the radio device selection interface 303, the mobile phone may further display a prompt information control 37-7 on the radio device selection interface 303, where a prompt information is displayed: will switch to rear-mounted microphone reception.
Alternatively, the prompt control 37-7 may be displayed on the radio selection interface 303, or on an interface 304 to be further displayed by the mobile phone, which is not limited in this embodiment of the present application.
Alternatively, the prompt control 37-7 may display a preset duration, such as 2 seconds; the prompt control 37-7 may also automatically disappear after detecting the clicking operation of the user, which is not limited in the embodiment of the present application. In addition, in response to the user clicking on the post radio option 37-3, the mobile phone may display an interface 304 as shown in fig. 3 (d), and the interface 304 may be understood as "radio gain adjustment interface".
Optionally, on the interface 304, a radio gain adjustment control 38, shown in dashed boxes, may be further displayed for the user, which radio gain adjustment control 38 may be used to adjust the gain during radio reception.
Specifically, "gain adjustment" may include three aspects: angle adjustment, intensity adjustment and effect adjustment. The adjustment process of each aspect may be divided into forward adjustment and reverse adjustment, and the adjustment amplitude and adjustment range of each aspect may be changed according to different electronic devices or different types of radio devices, which are not limited in the embodiments of the present application.
By way of example, embodiments of the present application list one possible implementation of gain adjustment:
(1) Angle adjustment
The term "angle" is understood to mean the "sound pickup range" of the microphone, and the angle adjustment, i.e. the adjustment of the microphone sound pickup range. It should be understood that, after the installation position of a certain microphone on the electronic device is fixed, the direction of sound pickup of the microphone may be determined according to the installation position of the microphone, or the range of audio that can be collected by the microphone is determined, and the angle between the audio range and the plane of the screen of the electronic device may be a preset angle.
Fig. 5 is a schematic view illustrating an embodiment of an audio reception angle adjustment apparatus according to an embodiment of the present application.
Illustratively, taking the back microphone 30 of the handset 100 of fig. 4 as an example, the back microphone 30 has a fixed mounting position on the handset 100: and the mobile phone is positioned in the central area of the back of the mobile phone. Accordingly, the back microphone 30 may collect audio in the region of the preset range Y shown in fig. 5, and the preset range Y may be: a conical area centered on the back microphone 30 has a fixed angle between the conical surface of the conical area and the plane of the back of the handset 100.
When the user performs a left-right sliding operation on the sound pickup gain adjustment control 38 shown in the dashed box (d) of fig. 3, the sound pickup range of the audio picked up by the back microphone 30 of the mobile phone 100 can be adjusted.
For example, when the gain parameter of the radio gain adjustment control 38 points to the 0 tick mark, the range in which the back microphone 30 of the mobile phone 100 can collect audio is the preset range Y.
When the user performs a sliding operation on the radio gain adjustment control 38 to the left, and performs an inverse adjustment of the angle, the gain parameter is reduced to a certain value, for example, -10, on the left side of the 0 scale line, so that the range in which the back microphone 30 of the mobile phone 100 can collect audio can be enlarged from the range Y to the range Z.
When the user performs a sliding operation to the right on the radio gain adjustment control 38, and performs forward adjustment of the angle, the gain parameter is increased to a certain value, for example 10, on the right side of the 0 scale line, so that the range where the back microphone 30 of the mobile phone 100 can collect audio is reduced from the range Y to the range X.
As can be seen from fig. 5, in the range Z, the back microphone 30 of the mobile phone 100 can collect the audio at the point a, if the user desires to collect the audio at the point B, the angle can be adjusted reversely, and the gain parameter can be reduced continuously, so that the range of the back microphone 30 capable of collecting the audio is enlarged continuously. The gain adjustment shown above can collect the audio at point C, and will not be described here again.
In the embodiment illustrated in fig. 5, the implementation of the user adjusting the sound pickup range through the sound pickup gain adjustment control 38 on the interface 304 is described by taking the example of forward adjustment, narrowing the sound pickup range, reverse adjustment, and expanding the sound pickup range. It will be appreciated that different principles may also be applied to adjust the sound reception range, for example: the specific meanings of the forward regulation and the backward regulation of the gain are not limited in the embodiment of the application.
In one possible implementation, the angle (or range) at which the microphone receives sound may be automatically adjusted based on other operations by the user in addition to supporting manual adjustment by the user.
Alternatively, the "range" of the microphone pickup may be associated with the viewing range of the video frame, with dynamic changes accompanying changes in the range of the video frame. The dynamic change can be a direct proportion change, namely, the video picture is enlarged, the radio reception range is enlarged, the video picture is reduced, and the radio reception range is reduced; alternatively, the "dynamic change" may be an inverse scale change, that is, the video frame is enlarged, the sound reception range is reduced, the video frame is reduced, and the sound reception range is increased, which is not limited in the embodiment of the present application.
For example, when the user manually enlarges the proportion of the video picture in the process of shooting the video, the range of the microphone for receiving sound is also enlarged in response to the operation of enlarging the video picture by the user, and when the proportion of the video picture is manually reduced by the user, the range of the microphone for receiving sound is also reduced in response to the operation of reducing the video picture by the user, which is not repeated herein.
(2) Intensity adjustment
The "intensity" is understood to mean the "sound intensity" of the audio collected by the microphone, which can also be adjusted manually by the user. The intensity adjustment process may also be divided into a forward adjustment, i.e. amplifying the collected audio sound, and a reverse adjustment, i.e. reducing the collected audio sound.
Illustratively, when the gain parameter of the radio gain adjustment control 38 points to the 0 tick mark, the back microphone 30 of the handset 100 may correspond to a preset sound intensity, e.g., the sound intensity may be rated as 0-10, which may be fixed at 5.
When the user performs a rightward sliding operation on the sound pickup gain adjustment control 38, forward adjustment of the sound pickup intensity is performed, and the gain parameter is increased to a certain value on the right side of the 0 scale mark, so that the sound intensity corresponding to the audio collected by the back microphone 30 can be increased. When the user performs a sliding operation on the radio gain adjustment control 38 to the left, and performs a reverse adjustment of the intensity, the gain parameter is increased to a certain value on the right side of the 0 scale line, so that the sound intensity corresponding to the audio collected by the back microphone 30 can be reduced.
In one possible implementation, the "sound intensity" of the microphone for receiving sound may be automatically adjusted according to other operations of the user, in addition to supporting manual adjustment by the user.
Alternatively, the audio "sound intensity" captured by the microphone may be associated with the video frame, with the video frame changing dynamically. The dynamic change can be a proportional change, namely the video picture is amplified, and the intensity of the collected audio sound is increased; alternatively, the "dynamic change" may be an inverse proportion change, that is, the video picture is enlarged, and the intensity of the collected audio sound is reduced, which is not limited in the embodiment of the present application.
For example, when the user manually enlarges the proportion of the video picture in the process of shooting the video, the intensity of the audio sound collected by the microphone is also increased in response to the operation of enlarging the video picture by the user, and when the user manually reduces the proportion of the video picture, the intensity of the audio sound collected by the microphone is also reduced in response to the operation of reducing the video picture by the user.
(3) Effect adjustment
An "effect" may be understood as a sound processing algorithm that adjusts the audio captured by a microphone, thereby modifying the playback effect of the captured audio.
For example, the process of "effect adjustment" may be that the audio collected by the microphone is superimposed by an audio enhancement processing algorithm of Hua-into company, for example, to obtain audio with a Histen audio, or to perform noise reduction processing on the audio collected by the microphone, and to superimpose the audio processing algorithm, so that the audio has an audio with a special play effect such as dolby audio, stereo, etc., which is not limited in this embodiment of the present application.
Optionally, in the above process of "effect adjustment", one or more controls may be added to the interface 304 shown in the (d) diagram in fig. 3 independently, where each control may correspond to a different sound effect, and the user may select, according to personal needs, to superimpose one or more effects on the audio collected by the microphone, where the process of effect adjustment supports dynamic adjustment of the user to meet multiple needs of the user, and this embodiment of the present application is not limited thereto.
Through the implementation process of angle adjustment, intensity adjustment and effect adjustment listed above, different radio strategies can be set by a user according to different scenes and shooting requirements in the process of shooting audio multimedia files such as videos by the user, so that an audio playing effect which better meets the user requirements is generated, the user can conveniently manufacture audio multimedia files such as interesting videos, and the use experience of the user is improved.
Alternatively, the three aspects of radio gain listed above may be manually adjusted independently by the user. Such as angle adjustment, intensity adjustment and effect adjustment, respectively, by different controls, such as one gain adjustment control 38 for each aspect, the gain parameters being individually adjusted or modified by the user as desired by the individual.
Alternatively, the three aspects of the radio gain may be simultaneously adjusted by superimposing any two aspects. For example, if the user modifies the gain parameters of the sound pickup gain adjustment control 38, the user can adjust the angle (range) and the audio intensity of the sound pickup at the same time, and the user can superimpose more sound effects on the audio collected by the microphone through other controls. The mode can independently control the audio acquisition range, the intensity and the sound effect, can meet the use requirements of different users, and does not force the users to overlap part of the sound effect.
Alternatively, the user may simultaneously achieve angle, intensity, and effect adjustments by modifying the gain parameters of the radio gain adjustment control 38. The implementation mode enables the adjusting function to be more centralized, the operation is more convenient, the number of controls on the interface is reduced, the use experience of a user can be improved to a certain extent, and the embodiment of the application is not limited to the above.
In one possible implementation, when the user performs the operation shown in fig. 3 (c), clicking on the post radio option 37-3, the handset may also display a channel control 39 shown in dashed outline on the interface 304 in response to the user's operation.
Optionally, the channel control 39 may correspond to different functions for different scenes, such as for adjusting the sound reception process of the left channel and the right channel, and/or adjusting the play effect of the left channel and the right channel, etc.
For example, when a plurality of microphones are installed on the mobile phone 100, if there is a left-right relative positional relationship between an object (e.g., a person) photographed in a video screen and the plurality of microphones of the mobile phone 100, the "L" control in the channel control 39 may be clicked by the user to select the microphone on the left (left) as the sound receiving device, and the "R" control may be clicked by the user to select the microphone on the right (right) as the sound receiving device, i.e., the user may manually select the microphone closer to the object photographed as the sound receiving device; or the "L" control and the "R" control are selected, namely, the left microphone and the right microphone are selected simultaneously to be used as the sound receiving equipment.
And/or the "L" and "R" controls of the channel control 39 may also be used by a user to adjust the sound intensity of audio captured by one or more sound receiving devices.
For example, as shown in the (d) diagram of fig. 3, the "L" control corresponds to a height greater than the "R" control corresponds to a height, and the sound intensity of the left channel of the audio collected by the one or more radio devices of the mobile phone 100 may be controlled to be greater than the sound intensity of the right channel. Accordingly, in the process of playing the video by the subsequent user, the effect of volume change of the left channel and the right channel can be played for the user along with the switching of video pictures, immersive video watching experience can be provided for the user, and the interestingness of the video is increased.
Illustratively, as shown in fig. 3 (e), the interface 305 is a process for playing video by the user, and if the user adjusts the sound receiving device to collect audio through the channel control 39 in the manner of the aforementioned fig. 3 (d), the sound intensity of the left channel is greater than the sound intensity of the right channel. Then, during the process of playing the video, a channel control may also be displayed on the video playing interface 305, which is used to indicate that the audio effect of "the sound intensity of the left channel is greater than the sound intensity of the right channel" is provided during the process of playing the video.
In the above description of the process shown in fig. 3, in a scene of capturing video, a user may dynamically control one or more sound receiving devices corresponding to audio of a video frame, and the user may dynamically adjust gain parameters such as a sound receiving range, intensity, or sound effect of the one or more sound receiving devices.
Alternatively, the process of dynamically controlling one or more sound receiving devices to collect audio and the process of dynamically adjusting gain parameters may occur before a user starts shooting, for example, the user sets the sound receiving device such as a microphone when shooting video in a professional mode of a camera application in advance, and the corresponding gain parameters, and the adjustment before shooting starts may be referred to as "pre-adjustment".
Alternatively, the process of dynamically controlling one or more radio receiving devices to collect audio and the process of dynamically adjusting gain parameters may occur during the process of capturing the audio by the user, for example, during the process of capturing the video by the user, just before the user starts to use the relevant settings before capturing, and as capturing proceeds, the user may manually modify the relevant settings on the capturing interface, for example, reselect the radio receiving device, readjust the radio receiving gain parameters, etc., and the adjustment during capturing may be referred to as "real-time adjustment", which is not limited in the embodiments of the present application.
Optionally, when the user performs related adjustment of the radio receiving device and the gain parameter before shooting, and performs adjustment during shooting, the latest radio receiving device and gain parameter obtained after adjustment during shooting can be saved as the corresponding shooting parameter in the shooting mode. The next time the user again takes the same mode, the related radio equipment and gain parameters can be used by default, and the user is supported to manually modify or adjust.
Or, the mobile phone only stores the radio receiving device and the gain parameter set by the user before shooting as the corresponding shooting parameters in the shooting mode, and the radio receiving device and the gain parameter set in the shooting process may not be stored.
Optionally, in addition to the professional mode of the camera application, the "microphone control 37" associated with the process of dynamically controlling the audio capturing process of the one or more audio receiving devices and the process of dynamically adjusting the gain parameters provided in the embodiments of the present application may also be set in different capturing modes, such as a video recording mode, a dual-view capturing mode, a sound photo capturing mode, and the like.
Alternatively, the microphone control 37 may be added in a setting interface corresponding to the camera setting control 36 of the shooting preview interface 302, and the user may click on the camera setting control 36 to enter the corresponding setting interface, select a radio receiving device, or the like, which is not limited in the embodiment of the present application.
In yet another possible implementation, after the user clicks the microphone control 37 shown in fig. 3 (b), the mobile phone may automatically and intelligently match one or more sound receiving devices for the user according to different reference indexes, and the option corresponding to the automatically matched sound receiving device on the sound receiving device selection interface 303 is in the selected state.
Optionally, the different reference indexes may include one or more of a current shooting scene, shooting parameters, a user frequency, a function of the sound receiving device, a distance between the sound receiving device and a sound source, and the mobile phone may automatically match the sound receiving device according to the one or more reference indexes.
By way of example, table 3 lists a number of possible reference indicators for automatically matching sound receiving devices, and the mobile phone 100 may detect the current shooting scene, the shooting camera, the frequency of use of the user, the performance of the sound receiving device, the distance between the sound receiving device and the sound source, etc., and automatically match one or more sound receiving devices to collect audio.
TABLE 3 Table 3
In combination with the content of table 3 above, when the camera capturing the video frame is detected to be the rear camera of the mobile phone 100, the rear microphone 30 of the mobile phone 100 can be automatically matched and selected to capture audio, and the rear radio option 37-3 on the radio device selection interface 303 can be displayed as the selected state.
When the camera for acquiring the video picture is a front camera of the mobile phone 100, the user may automatically switch to the top microphone 10 or the bottom microphone 20 of the mobile phone 100 to acquire audio, so as to realize front radio, and facilitate the user to add comments, voice-over and the like for the recorded video. Accordingly, the front radio option 37-2 on the radio selection interface 303 may be displayed in a selected state.
Alternatively, in a home scenario, when the video screen displayed by the mobile phone 100 originates from the camera of the distributed tablet 200, the microphone of the tablet 200 may be synchronously selected as the sound pickup device, and the bluetooth sound pickup option 37-5 on the sound pickup device selection interface 303 may be displayed in a selected state.
Or, when the use effect of the microphone in the mobile phone 100 is significantly limited, for example, when the user uses the mobile phone 100 to photograph a person remotely, any microphone of the mobile phone 100 cannot collect the audio of the person. In this case, a microphone or a sound pickup microphone capable of picking up audio is provided near the person to be photographed by a wired or wireless connection, and sound pickup can be automatically performed by matching the microphone or the sound pickup microphone with the wired connection.
Or, if the mobile phone 100 is only connected with a unique device, whether it is a wired connection or a wireless connection, the device connected with the mobile phone 100 can be preferentially matched automatically to collect audio as the sound receiving device, so as to support the user to modify the sound receiving device manually, and various different use scenarios are not repeated here.
It should be understood that when the radio receiving device is automatically matched with the user on the radio receiving device selection interface 303, if the automatically matched radio receiving device is not the radio receiving device desired by the user, the user may also manually modify or switch, and specifically refer to the above-described operation procedure of manually selecting the radio receiving device, and adjust parameters such as gain, which are not described herein for simplicity.
Fig. 6 is a schematic diagram of a user interface for capturing video by another user according to an embodiment of the present application.
In a possible display manner, on the photographing preview interface 601 shown in fig. 6 (a), the user clicks the microphone control 37, and in response to the user's operation, the mobile phone can display the interface 602 shown in fig. 6 (b).
Optionally, one or more available sound pickup devices may be displayed for the user on the interface 602, with different sound pickup devices being identified in different locations. For example, the "top microphone" option, the "bottom microphone" option, the "back microphone" option, and the like displayed on the interface 602 identify one or more sound receiving devices for the user through the positions of different sound receiving devices, so that the user can accurately determine the shooting requirement of the user, and manually click and select one or more sound receiving devices to collect audio according to the position of the shot object, the position relationship of the mobile phone, and the like.
Illustratively, the user performs the operation shown in fig. 6 (b), clicks the "back microphone" option, and in response to the user's operation, the mobile phone may display an interface 603 shown in fig. 6 (c), and the interface 603 may further display a sound reception gain adjustment control 38 and a sound channel control 39 shown by dashed boxes on the interface 603, where the sound reception gain adjustment control 38 may be used to adjust the gain of the back microphone 30 of the mobile phone 100 during sound reception. The channel control 39 is used to adjust the sound pickup process of the left and right channels during the sound pickup process of the back microphone 30 of the mobile phone 100, and/or to adjust the play effects of the left and right channels, etc.
In another possible display manner, on the photographing preview interface 604 shown in (d) of fig. 6, the user clicks the microphone control 37, and in response to the user's operation, the mobile phone may display an interface 605 shown in (e) of fig. 6.
Optionally, the user may be presented with the number of one or more available sound pickup devices on the interface 605, such as the "microphone A" option, "microphone B" option, and "microphone C" option, etc., presented on the interface 602. The user can check the instruction book of the mobile phone 100, determine the position of each microphone, and manually click and select one or more suitable radio receiving devices to collect audio according to the position of the shot object, the position relationship between the sound source and the mobile phone and the like according to the shooting requirement of the user.
For example, table 4 lists a possible correspondence between the installation position of the microphone of the mobile phone 100 and the number of the microphone, and the user may obtain the correspondence through the specification, the network data, etc. of the mobile phone 100, and view which microphone is specifically corresponding to the options of one or more radio receiving devices displayed on the interface, so that the process of selecting the radio receiving device may be more accurate to match different shooting requirements of individuals.
TABLE 4 Table 4
Position of radio equipment Numbering of radio Actual position corresponding to mobile phone 100 (instruction book)
Top microphone Microphone A Top microphone 10 (XXX 1)
Bottom microphone Microphone B Bottom microphone 20 (XXX 2)
Back microphone Microphone C Back microphone 30 (XXX 3)
…… …… ……
The user performs the operation shown in fig. 6 (e), clicks the "microphone B" option and the "microphone C" option, and in response to the user's operation, the mobile phone may display an interface 606 shown in fig. 6 (f), and the interface 606 may further display a sound receiving gain adjustment control 38 and a sound channel control 39 shown by dotted frames for the user, where the sound receiving gain adjustment control 38 may be used to simultaneously adjust gains of the back microphone 30 and the bottom microphone 20 of the mobile phone 100 during sound receiving, and the sound channel control 39 is used to adjust sound receiving processes of the left channel and the right channel during sound receiving of the back microphone 30 and the bottom microphone 20 of the mobile phone 100, and/or adjust play effects of the left channel and the right channel, etc.
Alternatively, the angles and intensities corresponding to the gains of the back microphone 30 and the bottom microphone 20 of the mobile phone 100 during the sound pickup process may be increased simultaneously, or decreased simultaneously, which is not limited in the embodiment of the present application.
The selection process of two different sound receiving devices is described above through (a) diagram- (b) diagram- (c) diagram in fig. 6 and (d) diagram- (e) diagram- (f) diagram in fig. 6, where multiple sound receiving devices can be displayed for the user in different manners on any interface of the process, and the user can select one or multiple sound receiving devices to collect audio; the user can also control the process of collecting the audio by the audio receiving device through the audio receiving gain adjusting control 38, the channel control 39 and the like, and independently adjust or intensively adjust the angle (range), the intensity, the sound effect and the like of the audio collected by the audio receiving device.
The process can provide the user with a plurality of different operation modes, the user can manually click and select one or more suitable radio equipment to collect the audio according to the shooting requirement of the user, the position of the shot object, the position relation between the sound source and the mobile phone and the like, the use requirement of different users is met, the autonomous selectivity of the user is improved, and the experience of the user for shooting the audio multimedia files such as videos is further improved.
Fig. 7 is a schematic diagram of a different electronic device included in a home scenario.
For example, in the home scenario shown in fig. 7, a plurality of electronic devices may be included, for example, the tablet 200, the smart screen 300, and the sound box 500 shown in fig. 4, and the mobile phone 100 used daily by the user, etc. are included in the home scenario. In this scenario, the tablet 200 is located in room 1, the smart screen 300 is located in room 2, the speaker 500 is located in room 3, and the user holds the mobile phone 100 to capture audio multimedia files such as video in a different room of the home scenario.
In the process of recording video by the mobile phone 100, the microphone of the mobile phone 100 may be used to collect audio, such as the top microphone 10, the bottom microphone 20, and the back microphone 30 shown in fig. 4, and the process may refer to the description of the foregoing embodiments, which is not repeated herein.
In addition, the tablet 200, the smart screen 300, and the speaker 500 are used as the distributed devices of the mobile phone 100, and the mobile phone 100 may also use the microphones of the distributed devices such as the tablet 200, the smart screen 300, and the speaker 500 to collect audio.
Assuming that the mobile phone 100 and the tablet 200, the smart screen 300, and the speaker 500 have established a distributed connection, such as a Wi-Fi connection, the embodiment of the present application does not limit the manner in which the distributed connection is established.
Several possible audio acquisition processes when capturing video are described below for the context of multiple distributed devices.
Fig. 8 is a schematic diagram of a user interface for capturing video by a user according to another embodiment of the present application.
In a possible display manner, on the photographing preview interface 801 shown in fig. 8 (a), the user clicks the microphone control 37, and in response to the user's operation, the mobile phone can display the interface 802 shown in fig. 8 (b). The user clicks on more options 37-6 of the interface 802, and in response to the user's operation, the mobile phone may display a window 802-1, where more radio receiving devices may be displayed for the user in the window 802-1, for example, wi-Fi radio receiving options that are not displayed by the interface 802, which is not limited in this embodiment of the present application.
The user performs the operation shown in fig. 8 (b), clicks on the Wi-Fi radio option in the window 802-1, and in response to the user's operation, the mobile phone displays an interface 803 shown in fig. 8 (c).
Optionally, a window 803-1 may be displayed on the interface 803, where one or more options are displayed in the window 803-1 for the user, where each option corresponds to a plurality of device microphones that establish a distributed connection with the mobile phone 100, for example, a "tablet microphone" option, an "intelligent screen microphone" option, and a "sound box microphone" option, and one or more sound receiving devices are identified for the user through different devices, so that the user can accurately determine his or her shooting requirement, and manually click to select one or more suitable sound receiving devices to collect audio according to the positional relationship between the plurality of device microphones and the mobile phone, and so on.
Illustratively, the user performs the operation shown in fig. 8 (c), clicks on the "smart-screen microphone" option, and in response to the user's operation, the mobile phone may display an interface 804 shown in fig. 8 (d), on which interface 804 the radio gain adjustment control 38 and the channel control 39 shown in dashed boxes may be further displayed for the user, and on which interface 804 the "smart-screen microphone" option is also displayed for prompting the user for the current radio gain adjustment control 38 and channel control 39 for adjusting microphone parameters of the smart-screen 300. Specifically, the sound pickup gain adjustment control 38 may be used to adjust the gain of the microphone of the smart screen 300 during sound pickup. The channel control 39 is used to adjust the sound reception process of the left and right channels of the microphone of the smart screen 300, and/or to adjust the play effects of the left and right channels, etc.
It should be noted that, in the home scenario, the plurality of distributed devices are mutually trusted devices, so when the microphone of the smart screen 300 is trusted by the mobile phone 100 to collect audio, that is, the mobile phone 100 has the authority to control the trusted distributed devices, the radio gain adjustment control 38 and the channel control 39 displayed on the interface 804 of the mobile phone 100 can be used to adjust the microphone parameters of the smart screen 300.
Through the process shown in fig. 8, it is introduced that the user may manually choose to use the microphones of other devices to collect audio, and likewise, the user may also choose any two distributed device microphones, or more devices' microphones to collect audio from different rooms on the interface 803.
In addition, the video frames of the user when capturing video on the cell phone 100 may also originate from the camera of the distributed device. Alternatively, when the video frame of the mobile phone 100 is derived from a distributed device, the radio device of the mobile phone 100 may preferably use a device having the same source as the video frame to keep the synchronization of the audio and video, which is not limited in the embodiment of the present application.
For the home scenario shown in fig. 7, the handset 100 may acquire audio and/or video from multiple home devices, and any one or more of the rooms may acquire audio and/or video regardless of which room in the home scenario the user holds the handset 100. The scene can be applied to video calls, audio calls and the like among a plurality of family members, and the user can acquire the audio and/or video of other rooms without holding a mobile phone to replace the room, so that the use scene is enriched, and the use experience of the user is improved.
Fig. 9 is a schematic diagram of a user interface for capturing video by a user according to another embodiment of the present application.
Alternatively, when the interface 901 shown in (a) of fig. 9 displays microphone options of a plurality of devices, in the implementation process described in fig. 8, if the user clicks the microphone option of the plurality of devices, in response to the operation of the user, the microphones of the plurality of devices selected by the user may collect audio at the same time and return the collected audio to the mobile phone 100.
Alternatively, in fig. 9, a different implementation may be provided, specifically, if the microphone of the plurality of devices is selected as the sound receiving device by the user according to the operation shown in the diagram (a) in fig. 9, in response to the operation of the user, the microphone of the plurality of devices selected by the user may be used as a "candidate plurality of sound receiving devices", and during the process of capturing the video by the user, according to different reference indexes, for example, the reference indexes and the automatic matching principles shown in table 3, audio may be automatically acquired for the user from the "candidate plurality of sound receiving devices", which is not limited in this embodiment of the present application.
Illustratively, on the interface 901 shown in (a) of fig. 9, assuming that the user clicks to select the "tablet microphone" option, the "smart-screen microphone" option and the "speaker microphone" option, in response to the user's operation, the mobile phone may display the interface 902 shown in (b) of fig. 9, automatically pop up the prompt window 902-1 on the interface 902, and prompt information may be displayed in the prompt window 902-1: is it automatically switched between the following radio microphones? And the prompt window 902-1 may also display a "flat microphone" option, a "smart-screen microphone" option, and a "sound box microphone" option in a selected state, and if the user desires to cancel a microphone of a certain device, the user may click on the corresponding option to cancel.
When the user performs the operation shown in the (b) diagram in fig. 9, the "flat microphone" option, the "smart screen microphone" option and the "sound box microphone" option are selected, and the "determine" control is clicked, and in response to the user's operation, the mobile phone determines that the "flat microphone, the smart screen microphone and the sound box microphone" are all used as "candidate multiple sound receiving devices" in the process of recording video, and the mobile phone 100 can further automatically match one or more sound receiving devices from the "candidate multiple sound receiving devices" to collect audio for the user according to different reference indexes.
Illustratively, in the home scenario as shown in (c) of fig. 9, the cellular phone 100 and the tablet 200, the smart screen 300 and the sound box 500 establish a distributed connection and the user selects "tablet microphone, smart screen microphone and sound box microphone" as "candidate plural sound receiving devices" according to the procedure shown in (a) and (b) of fig. 9, and the sound receiving devices may be automatically switched among the microphones of the plural devices. Assume that a user takes a video using the mobile phone 100, the object to be taken is a child, and the sound source from which the video is taken is also a child. In the shooting process, the moving track of the girl is a position A, a position B and a position C.
Alternatively, during the process of capturing video by the mobile phone 100, the distance d between the sound source (girl) and the microphone of each of the distributed devices may be detected, and the sound receiving device may be automatically matched and switched according to the distance d between the sound source (girl) and the microphone of each of the distributed devices.
In one possible implementation, the mobile phone 100 and the tablet 200, the smart screen 300 and the speaker 500 may all include positioning devices, so that the distance between any two devices, or the distance between the detection and sound source, etc. can be measured in real time.
Alternatively, the positioning means may be a sensor with a positioning function or a structure with a positioning function or the like. A sensor capable of measuring distance, such as a laser sensor, an infrared sensor, or the like; the structure with the positioning function may be a bluetooth positioning structure, a positioning chip based on Ultra Wide Band (UWB) wireless sensing capability, a positioning structure based on a global satellite positioning system (global positioning system, GPS), a positioning structure based on a wireless fidelity (wireless fidelity, WIFI) module, etc., which is not limited in this embodiment of the present application.
Illustratively, as shown in (c) of fig. 9, when the child (sound source) is at the position a, the distance d1 between the child and the sound box 500 is smaller than the distance d2 between the child and the mobile phone 100, and the distance d1 between the child and the sound box 500 is smaller than the distance d3 between the child and the smart screen 300. Thus, at location a, the mobile phone 100 may automatically match the audio picked up by the microphone of the speaker 500, i.e., the audio corresponding to the video frame captured by the mobile phone 100 at location a originates from the microphone of the speaker 500.
When the girl (sound source) is at the position B, the distance d4 between the girl and the mobile phone 100 is smaller than the distance d5 between the girl and the smart screen 300, and d4 is smaller than the distance d6 between the girl and the tablet 200. Therefore, at the position B, the mobile phone 100 can automatically match the audio received through its microphone, that is, the audio corresponding to the video recorded by the mobile phone 100 originates from the audio receiving device of the mobile phone 100 itself.
In addition, if the video picture photographed at the position B is acquired by the rear camera of the mobile phone 100, and accordingly, in combination with the reference indexes listed in table 3, the mobile phone 100 may preferentially acquire the audio through the back microphone 30, which is not limited in the embodiment of the present application.
When the child (sound source) moves to the position C, the distance d7 between the child and the tablet 200 is smaller than the distance d8 between the child and the cell phone 100, and smaller than the distance d9 between the child and the smart screen 300. Thus, at location C, the handset 100 can automatically match the audio received through the microphone of the tablet 200, i.e., the audio corresponding to the video recorded by the handset 100 originates from the microphone … … of the tablet 200.
In summary, in the process of moving the child (sound source) from the position a to the position B, the microphone of the sound box 500 may collect the audio first, then gradually switch to the microphone of the back microphone 30 of the mobile phone 100 itself to collect the audio, and then gradually switch to the microphone of the tablet 200 to collect the audio.
It should be noted that, in the above process in this embodiment of the present application, whether the user manually selects the microphone of the distributed device such as the tablet 200, the smart screen 300, the sound box 500, etc. to receive sound, or intelligently matches the microphone of one of the distributed devices to receive sound, for example, the smart screen 300 may also include n microphones, where n is greater than or equal to 1.
In this scenario, the n microphones included in the smart screen 300 may be considered as part of the self-structure of the mobile phone 100, and according to the foregoing possible manners described in fig. 3 and table 2, according to the relative position, distance, and other reference options of the sound source and each of the n microphones, the sound reception using one or more of the n microphones of the smart screen 300 is automatically determined, which is not limited in the embodiment of the present application.
Alternatively, when the user selects to use the smart-screen microphone for receiving sound according to the window 803-1 shown in fig. 8 (c), n microphones included in the smart-screen 300 and/or the installation position of each microphone on the smart-screen 300 may be further displayed in the interface 803 shown in fig. 8 (c), and the user may further select one or more of the n microphones, which may refer to the foregoing process of selecting the distributed device, and will not be repeated here for brevity.
In combination with the above-described process, if the sound source is in a motion state during the process of capturing audio multimedia files such as video, the audio receiving device for collecting audio can be automatically switched. In the switching process, the distance between the mobile phone for shooting the video and the sound source can be detected in real time, or the radio receiving equipment can be automatically matched according to different reference indexes such as the positions, shooting environments, shooting parameters and the like of other distributed equipment, or the switching between microphones of different equipment can be realized, so that the radio receiving effect in the video shooting process can be improved.
Specifically, the mobile phone can detect the distance between the sound source moving process and a plurality of sound receiving devices in real time, intelligently switch and use different sound receiving devices according to the continuously changing distance, for example, the sound receiving device closer to the sound source is preferentially used for collecting audio, the process can improve sound receiving quality, collect clearer audio for users, and the method is beneficial to manufacturing sound multimedia files such as higher quality video.
It should be understood that the above embodiments take the process of capturing video by a user as an example, and describe a possible implementation process of capturing, by an electronic device, audio corresponding to a video frame in the process of capturing the video frame. In the process of collecting the audio, different kinds, different numbers and different sources of audio receiving equipment can be selected according to different strategies, and a user can be supported to adjust the audio gain corresponding to the audio receiving equipment, so that the audio collection is controlled.
Optionally, the process of collecting the audio frequency supports the manual selection of radio receiving equipment with different types, different numbers and different sources by a user, and the electronic equipment can also intelligently match the radio receiving equipment with different types, different numbers and different sources for the user according to different scenes and different reference indexes, and the intelligent switching of the radio receiving equipment is realized in the radio receiving process, so that the interesting audio multimedia file is manufactured.
Optionally, the method for collecting audio provided in the embodiment of the present application may be applied to a scene for making an audio file, in addition to a scene for recording video.
For example, a user may use an application installed in an electronic device such as a mobile phone, a tablet, a smart screen, etc. and having a recording function, collect audio through a radio device such as a microphone of the electronic device, etc. to make an audio-type multimedia file, where the audio file does not need to be associated with a video frame.
It should be understood that the method of the embodiment of the present application may be used in the process of making an audio-type multimedia file, and in the process of recording an audio file, a user may select different types, different numbers, and different sources of audio receiving devices, and support the user to adjust the audio gain corresponding to the audio receiving devices, so as to control the audio collecting process, thereby making an audio-type multimedia file with good audio receiving effect and full of interests.
Or, the method for collecting audio provided in the embodiment of the present application may be further applied to all scenes where sound is received by using the microphone of the electronic device, for example, in a process of answering a phone call, where the microphone of the mobile phone collects the sound of the user, or may be applied to the method in the embodiment of the present application, which is not limited in this embodiment of the present application.
Optionally, the capability of collecting audio provided in the embodiment of the present application may be a system capability of an electronic device such as a mobile phone, and accordingly, the system capability may be called by a camera application, a recording software, or other various application programs having a video recording function and/or an audio collecting function, and may be downloaded and installed by a userOr the like, or can be applied to +.>Application, & gt>Applications, etc. any application having audio and/or video acquisition capabilities, embodiments of the present application are not limited in this regard.
Alternatively, the capability of collecting audio provided in the embodiment of the present application may be a preset function configured by a camera application of an electronic device such as a mobile phone, or a function manually opened by a user in an application such as a related camera application, which is not limited in the embodiment of the present application.
Optionally, in addition to the above-described professional mode shooting scene, the method provided in the embodiment of the present application may also be used to make a sound photo, a double-shot, a multi-shot, a video recording, or other different shooting modes or shooting scenes, which is not limited in this embodiment of the present application.
Each shooting mode has individual specificity, but as long as the shooting process or the process of making the audio multimedia file involves the collection of the audio, the method provided by the embodiment of the application can be used for realizing the control of the audio collection process, such as the control of the sound receiving equipment, the adjustment of the sound receiving gain parameter and the like.
Taking a scene of recording a video clip in a dual-view shooting mode as an example, fig. 10 is a schematic diagram of a user interface of a user shooting video according to another embodiment of the present application.
In a possible display manner, the shooting preview interface 1001 of the mobile phone may be as shown in fig. 10 (a), where the user switches the current shooting mode to the double-shot mode, and accordingly, the "black small triangle icon" for positioning the current shooting mode may point to the "double-shot" mode control. Alternatively, on the photographing preview interface 1001 of the double-shot mode, two preview windows may be displayed: preview window 1001-1 and preview window 1001-2.
In another possible display manner, the shooting preview interface 1001 of the mobile phone may be as shown in (b) of fig. 10, and in the dual-view shooting mode, two preview windows on the shooting preview interface 1002 may be displayed in a form of "picture-in-picture: the preview window 1002-1 may be displayed in a floating manner in a display area of a preview image of the photographing preview interface 1002, which is not limited in the embodiment of the present application.
Alternatively, the pictures in the different preview windows may originate from different mobile phone cameras, or the pictures in the different preview windows may also originate from cameras of different devices, such as pictures acquired by the cameras of the smart screen 300 and the tablet 200 in the home scene.
For example, for the interface 1001 shown in fig. 10 (a), the screen in the preview window 1001-1 may be derived from a rear wide-angle camera of the mobile phone, and used for capturing a remote scene, which is called a "rear Jing Liu" screen; the screen in the preview window 1001-2 may originate from a front camera of the handset for capturing a close-up of the face of the person, and is referred to as the "front Jing Liu" screen.
Optionally, in the video capturing process of the dual-view screen illustrated in (a) of fig. 10, the relevant control process of the radio receiving device may include the following different scenarios:
Scene 1: the different preview windows are completely independent, and the user can select the radio receiving device for each preview window, adjust the radio receiving gain parameter, etc. for the window.
For example, as shown in fig. 10 (a), controls such as microphone control and ISO sensitivity may be displayed in the preview window 1001-1 and the preview window 1001-2, and the user may click on the respective microphone control in the preview window 1001-1 and the preview window 1001-2, set the audio receiving device corresponding to the video frame in each preview window, and further adjust the gain parameter corresponding to the audio receiving device, etc., which are described in the foregoing embodiments, and will not be repeated here.
Illustratively, the preview window 1001-1 and the preview window 1001-2 may correspond to different channel parameters, for example, a channel control may be displayed in each preview window, and the user may individually adjust the sound effect of each preview window; as shown in fig. 10 (a), the preview window 1001-1 may correspond to different left channel parameters, and the preview window 1001-2 may correspond to different right channel parameters, which is not limited in the embodiment of the present application.
Scene 2: one preview window is configured with fixed radio equipment and radio gain parameters, and the other preview window can support the user to select the radio equipment, adjust the radio gain parameters and the like.
For example, as shown in fig. 10 (b), the preview window 1002-1 displayed in floating in the form of "picture-in-picture" may correspond to a preset sound pickup device, and the display area of the preview image of the photographing preview interface 1002 may include a microphone control, and the user may select the sound pickup device, adjust the sound pickup gain parameter, and the like according to the procedure described in the foregoing embodiment.
It should be understood that the video frame of the preview window 1002-1, which is typically displayed in floating in a "picture-in-picture" manner, is preferably from the front camera of the mobile phone, and accordingly, the preview window 1002-1 may correspond to the fixed top microphone 10 and/or the fixed bottom microphone 20, which is not limited in this embodiment.
Scene 3: each preview window in the double-shot mode is configured with fixed radio equipment, radio gain parameters.
As shown in fig. 10 (a), as long as no microphone control may be displayed in the preview window 1001-1 and the preview window 1001-2 in the dual-view shooting mode, the preview window 1001-1 is configured to correspond to the back microphone 30 of the mobile phone 100, and the preview window 1001-2 corresponds to the top microphone 10 or the bottom microphone 20 of the mobile phone 100, i.e. each preview window is configured to collect audio using a fixed audio receiving device, which is not limited in the embodiment of the present application.
Scene 4: in the double-shot mode, the user only needs to set the audio control mode, such as the sound receiving device and the sound receiving gain parameter, and the mode can be suitable for the video of each preview window.
Illustratively, as shown in fig. 10 (a), a sound pickup apparatus, an adjusted sound pickup gain parameter, etc. set by the user in the single view screen may be applied to each preview window of the double view photographing mode. In other words, if the user modifies the sound receiving device, the sound receiving devices for capturing audio corresponding to the two preview windows are changed to the user modified sound receiving device accordingly.
It should be noted that, in the process of shooting a single-view video, the process of selecting the radio equipment by the user and the process of adjusting the gain parameters of different radio equipment are described in fig. 3 to 9; fig. 10 illustrates a possible implementation in the process of capturing a dual-view video. It should be understood that, for the multi-view shooting mode, reference may also be made to the implementation process of the above-mentioned dual-view shooting mode, which is not described herein.
In summary, in the process of producing the audio multimedia file, first, the embodiment of the present application may provide a method for manually selecting the sound receiving device for the user, where the user may manually select different sound receiving devices on one electronic device, and may manually select the sound receiving device using different devices to collect the audio. According to the method, a user can manually select or switch different sound receiving devices according to different shooting requirements, shooting scenes and the like, so that different sound receiving devices are used for collecting audio, and the quality of audio collection is improved.
Secondly, the embodiment of the application also provides a method for controlling the sound receiving equipment to collect the audio for the user, the user can manually adjust the sound receiving gain parameters of the sound receiving device, for example, adjust the corresponding sound receiving range, sound intensity, sound effect and the like in the sound receiving process, so that the user can better select the sound receiving device and adjust the sound receiving gain parameters according to different requirements, and the user can more easily obtain high-quality and interesting sound multimedia files.
Furthermore, the embodiment of the application also provides a convenient and simple adjusting mode for the user, the user can realize the adjustment of the radio gain parameter through the forward adjusting or reverse adjusting mode, the operation process is simple, the method is applicable to different user groups, and the use requirements of different users are met.
In addition, the embodiment of the application also provides an intelligent method for matching the sound receiving devices, which can automatically match one or more sound receiving devices for the current scene according to different reference indexes such as the distance between the sound receiving devices and the sound source, the positions of the sound receiving devices and the like, or switch the sound receiving devices such as microphones of different equipment to collect audio in the sound receiving process. The process can improve the sound reception quality, collect clearer audio for users, improve the sound reception effect and facilitate the production of better-quality sound multimedia files.
The above embodiment describes an audio collection process for producing a sound multimedia file from a user interaction level in conjunction with fig. 3 to 10, and a method for controlling audio collection in the process will be described from a software implementation policy level in conjunction with fig. 11.
It should be appreciated that the method may be implemented in an electronic device (e.g., a cell phone, tablet, etc.) having a structure such as a display screen, one or more cameras, etc., as shown in fig. 1, 2.
Fig. 11 is a schematic flowchart of an example of a method for capturing audio according to an embodiment of the present application, as shown in fig. 11, the method may include the following steps:
1101, displaying an image preview interface for video recording, wherein the image preview interface comprises a first control and a second control, the first control is used for controlling audio in the video recording process, and the second control is used for starting or suspending the video recording process.
Optionally, the "image preview interface for video recording" in this embodiment of the present application may be an image display interface corresponding to a shooting mode capable of recording video, where the camera application is not limited to a system application configured by an electronic device factory, and may also include a system application downloaded by a user and installed to the electronic device, where the system application is capable of recording video in any one of a video recording mode, a double-view shooting mode, a picture-in-picture shooting mode, and the like And the like, and is not limited thereto.
Alternatively, the "image preview interface for video recording" may also be user-in-useApplication, & gt>Any application having a function of capturing video, such as an application, for example, the user can use +.>Application of recording small video for 10 seconds, the video recording interface can be included in the scope of the embodiments of the application, and the process of audio acquisition is controlled according to the method provided by the embodiments of the application.
Alternatively, in the embodiment of the present application, a device capable of capturing audio is referred to as "sound capturing device", "audio capturing device", or "sound capturing device", for example, a microphone, an external earphone, or the like, which is not limited in the embodiment of the present application.
For example, the "first control" may be the microphone control 37 on the interface 302 shown in the diagram (b) in fig. 3 and the "second control" may be the shooting shutter control 31 on the interface 302, where the user clicks the shooting shutter control 31 once to start recording video, and clicks the shooting shutter control 31 once again to pause or end the recording process, which will not be described herein.
1102, detecting a first operation acting on the first control, and responding to the first operation, and displaying one or more options on the image preview interface, wherein the one or more options correspond to n audio acquisition devices, and n is an integer greater than or equal to 1.
1103, detecting a second operation acting on the first option, and in response to the second operation, determining the audio capturing device corresponding to the first option as a target audio capturing device, or automatically matching the target audio capturing device from the n audio capturing devices.
1104, detecting a third operation acting on the second control, starting to record video in response to the third operation, and acquiring the audio of the target sound source by using the target audio acquisition device in the process of recording video.
By way of example, the "first operation", "second operation", "third operation" may be a click operation, a double click operation, a long press operation, etc., which are not limited in the embodiments of the present application.
By way of example, the "one or more options" may correspond to one or more of the omnidirectional radio option 37-1, the front radio option 37-2, the rear radio option 37-3, the wired radio option 37-4, the bluetooth radio option 37-5, the more options 37-6, etc. on the radio device selection interface 303 as shown in fig. 3 (c) described in the foregoing embodiments, and the "first option" may be any one of them.
Through the process, aiming at the process of recording video, a method for selecting the audio acquisition device or switching the audio acquisition devices is provided for a user, the user can check available n audio acquisition devices through microphone controls on an interface according to own shooting requirements, and manually click and select one or more audio acquisition devices to acquire audio, different audio acquisition devices are used for acquiring audio, so that the use requirements of different users are met, the autonomous selectivity of the user is improved, and further the experience of the user for shooting audio and other sound multimedia files is improved.
Optionally, the n audio acquisition devices include one or more audio acquisition devices configured on the electronic device; and/or one or more audio acquisition devices of other devices which establish a connection relationship with the electronic device are included in the n audio acquisition devices, and the connection relationship includes wired connection or wireless connection.
In this embodiment of the present application, for example, the mobile phone may include one or more microphones installed on the mobile phone itself, and may further include microphones installed on other distributed devices that can be called by the mobile phone. For example, in a home scenario, a mobile phone may use a microphone of a distributed device such as a tablet, a smart screen, a speaker, etc. as an audio capturing device, and each distributed device may also include one or more audio capturing devices such as a microphone, which is not limited in this embodiment of the present application.
In one possible implementation, the electronic device may automatically intelligently match one or more audio capture devices for the user according to different reference indicators (reference options), and the options corresponding to the automatically matched audio capture devices on the audio capture device selection interface may be displayed as selected states.
Optionally, the "different reference indicators" may include one or more of the shooting scene, shooting parameters, user frequency, function of the audio capturing device, distance between the audio capturing device and the target sound source, etc. of the currently recorded video listed in table 3, and the electronic device may automatically match the audio capturing device according to the one or more reference indicators.
For example, the electronic device may determine, as the target audio capturing device, an audio capturing device closest to the target sound source among the n audio capturing devices according to the position of the target sound source.
And/or the electronic equipment can determine the audio acquisition device with the largest audio acquisition range among the n audio acquisition devices as the target audio acquisition device according to the audio acquisition range corresponding to each audio acquisition device.
And/or the electronic equipment can determine the audio acquisition device of which the audio acquisition range covers the target sound source as the target audio acquisition device according to the position of the target sound source and the audio acquisition range corresponding to each audio acquisition device.
And/or the electronic equipment can determine the audio acquisition device with the highest use frequency of the user in the n audio acquisition devices as the target audio acquisition device according to the use frequency of the user.
And/or the electronic equipment can automatically determine the audio acquisition device which is used by the user for the last time in the n audio acquisition devices as the target audio acquisition device according to the use record of the user.
And/or the electronic equipment can determine the audio acquisition device with the highest performance in the n audio acquisition devices as the target audio acquisition device according to the corresponding performance of each audio acquisition device.
And/or the electronic equipment can determine the audio acquisition devices which are connected with the electronic equipment in a wired way from the n audio acquisition devices as the target audio acquisition devices according to the mode that each audio acquisition device is connected with the electronic equipment.
And/or the electronic equipment can automatically match the target audio acquisition device from the n audio acquisition devices according to the installation position of the camera used for acquiring the video picture in the video recording process.
And/or the electronic device may automatically match the target audio acquisition device from the n audio acquisition devices according to a source of a camera used for acquiring the video image in the video recording process, which is not limited in the embodiment of the present application.
Optionally, when the user clicks the "first control (e.g. the microphone control 37)" to view one or more available audio capturing devices, a certain option of the one or more options on the audio capturing device selection interface 303 may be in a default selected state, e.g. a "third option", that is, the audio capturing device corresponding to the option is a default audio capturing device, and the user may modify or switch the audio capturing device according to his or her own requirement.
It should be understood that, according to the implementation manner of the above-mentioned automatic matching audio capturing device, when the audio capturing device is automatically matched with the user on the audio capturing device selection interface, if the automatic matching audio capturing device is not the audio capturing device desired by the user, the user may also manually modify or switch, specifically refer to the above-mentioned manual selection of the operation process of the audio capturing device, and adjust parameters such as gain, which will not be repeated herein for simplicity.
In another possible implementation manner, in the video recording process, when one or more cameras used for capturing video frames originate from the electronic device and the n audio capturing devices include one or more audio capturing devices configured on the electronic device, a preset corresponding relationship is provided between the one or more cameras of the electronic device and the one or more audio capturing devices configured on the electronic device, and the electronic device may automatically match the target audio capturing device from the one or more audio capturing devices configured on the electronic device according to the one or more cameras used for capturing video frames and the preset corresponding relationship.
In still another possible implementation manner, during the video recording process, when one or more cameras used for capturing video frames originate from other devices that establish a connection with the electronic device, and one or more audio capturing devices of the n audio capturing devices include one or more other devices that establish a connection with the electronic device, the electronic device may determine, as the target audio capturing device, an audio capturing device that the other devices that have one or more cameras used for capturing video frames originate from.
For example, in a home scenario, when the video frames displayed by the cell phone originate from a distributed tablet camera, the tablet microphone may be synchronously selected as the target audio pickup device.
Or if the mobile phone is only connected with a unique device, whether in wired connection or wireless connection, the device connected with the mobile phone can be preferentially matched automatically as a target audio acquisition device, so that the user can be supported to manually modify or switch, and the manual modification or switching is not repeated here.
In a possible scenario, when one or more audio capture devices of the n audio capture devices include other devices that establish a connection with the electronic device, the electronic device may detect a fourth operation acting on the second option, determine, in response to the fourth operation, one or more devices that establish a connection with the electronic device, and display a control that includes the one or more devices; and detecting a fifth operation acting on a control corresponding to the target equipment, and responding to the fifth operation, determining one or more audio acquisition devices of the target equipment as the target audio acquisition devices, and acquiring the audio of a target sound source by using the target audio acquisition devices in the process of recording video. The connection relation comprises one or more wireless connection relations among Bluetooth connection, wireless local area network Wi-Fi connection, infrared connection and near field communication NFC connection.
In yet another possible implementation, after determining the target audio capture device, the electronic device may display a prompt for prompting a user of a name or source of the target audio capture device currently being used by the electronic device.
Alternatively, the prompt information may be displayed on an image preview interface of the video recording, which is not limited in the embodiments of the present application.
Alternatively, the prompt may be displayed for a preset duration, for example, 2 seconds; the prompt information may also automatically disappear after the click operation of the user is detected, which is not limited in the embodiment of the present application.
In yet another possible implementation manner, after determining the target audio acquisition device, the electronic device may display one or more controls on the image preview interface of the video recording, where each control of the one or more controls is used to control an operating parameter of the target audio acquisition device, where the operating parameter includes one or more of an audio acquisition range, an audio acquisition intensity, and an audio playing effect during operation of the target audio acquisition device; detecting a sixth operation on the one or more controls, and adjusting an operating parameter of the target audio acquisition device in response to the sixth operation.
By way of example, the "sixth operation" may be a manipulation of the radio gain adjustment control 38 and/or the channel control 39 as in the (d) diagram of fig. 3, such as a sliding gain parameter, or the like.
Specifically, "gain adjustment" may include three aspects: angle adjustment, intensity adjustment and effect adjustment. The adjustment process of each aspect may be divided into forward adjustment and reverse adjustment, and the adjustment amplitude and adjustment range of each aspect may be changed according to different electronic devices or different types of radio devices, which are not limited in the embodiments of the present application.
It should be understood that the above embodiments take the process of capturing video by a user as an example, and describe a possible implementation process of capturing, by an electronic device, audio corresponding to a video frame in the process of capturing the video frame. In the process of acquiring the audio, the electronic equipment can acquire the audio of the target sound source by using audio acquisition devices of different types, different numbers and different sources.
Specifically, the process of collecting the audio frequency supports the manual selection of the audio frequency collecting devices with different types, different numbers and different sources by a user, and the electronic equipment can also intelligently match the audio frequency collecting devices with different types, different numbers and different sources for the user according to different scenes and/or reference indexes, and the intelligent switching of the audio frequency collecting devices is realized in the radio receiving process, so that the interesting audio multimedia file is manufactured.
Optionally, the method for collecting audio provided in the embodiment of the present application may be applied to a scene for recording video, a scene for making an audio file, and any scene that needs to be picked up by using an audio collecting device.
For example, a user may use an application installed in an electronic device such as a mobile phone, a tablet, a smart screen, etc. and having a recording function, collect audio through an audio collection device such as a microphone of the electronic device, etc. to make an audio-class multimedia file, where the audio file does not need to be associated with a video frame.
It should be understood that the method of the embodiment of the present application may be used in the process of making an audio-type multimedia file, where a user may select different types, different numbers, and different sources of audio acquisition devices during the process of recording an audio file, and support the user to adjust the audio gain corresponding to the audio acquisition devices, so as to control the audio acquisition process, thereby making an audio multimedia file with good sound receiving effect and full interest.
Or, the method for collecting audio provided in the embodiment of the present application may be further applied to all scenes where sound is received by using the microphone of the electronic device, for example, in a process of answering a phone call, where the microphone of the mobile phone collects the sound of the user, or may be applied to the method in the embodiment of the present application, which is not limited in this embodiment of the present application.
In summary, in the process of producing the audio multimedia file, first, the embodiment of the application can support the user to manually select different audio collection devices on one electronic device, and also can support the user to manually select the audio collection devices using different devices to collect audio. According to the method, a user can manually select or switch different radio equipment according to different shooting requirements, shooting scenes and the like, so that different audio acquisition devices are used for acquiring audio, and the quality of audio acquisition is improved.
Secondly, the embodiment of the application can also support the manual adjustment of the sound receiving gain parameters of the audio collecting device by a user, such as adjusting the corresponding sound receiving range, sound intensity, sound effect and the like in the sound receiving process, so that the user can better select the audio collecting device and adjust the sound receiving gain parameters according to different requirements, and can obtain high-quality interesting sound multimedia files more easily.
Furthermore, the embodiment of the application can also support the adjustment of the radio gain parameters by the forward adjustment or the reverse adjustment of the user, has a simple operation process, is suitable for different user groups, and meets the use requirements of different users.
In addition, according to the embodiment of the application, one or more audio acquisition devices can be automatically matched for the current scene according to different multiple reference indexes such as the distance between the audio acquisition devices and the sound source, the positions of the audio acquisition devices and the like, or the audio acquisition devices such as microphones of different equipment are switched to acquire audio in the sound reception process. The process can improve the sound reception quality, collect clearer audio for users, improve the sound reception effect and facilitate the production of better-quality sound multimedia files.
It should be noted that, in the description of the embodiments of the present application, unless otherwise indicated, "/" means or, for example, a/B may represent a or B; "and/or" herein is merely an association relationship describing an association object, and means that three relationships may exist, for example, a and/or B may mean: a exists alone, A and B exist together, and B exists alone. In addition, in the description of the embodiments of the present application, "plurality" means two or more than two.
It should be understood that the manner, the case, the category, and the division of the embodiments in the embodiments of the present application are for convenience of description only, and should not be construed as a particular limitation, and the features in the various manners, the categories, the cases, and the embodiments may be combined without contradiction.
It should also be understood that the terms "first," "second," and "third" in the embodiments of the present application are merely for distinction and should not be construed as limiting the present application in any way. For example, "first operation", "second operation", and the like in the embodiments of the present application are meant to include different user operations.
It should also be understood that, in various embodiments of the present application, the size of the sequence number of each process does not mean that the execution sequence of each process should be determined by its functions and internal logic, and should not constitute any limitation on the implementation process of the embodiments of the present application.
It should be further noted that, in the embodiment of the present application, the "preset", "fixed value", and the like may be implemented by pre-storing corresponding codes, tables, or other manners that may be used to indicate related information in the electronic device, and the specific implementation manner of the present application is not limited, for example, the "preset audio acquisition device", "preset correspondence", and the like in the embodiment of the present application.
It will be appreciated that the electronic device, in order to achieve the above-described functions, includes corresponding hardware and/or software modules that perform the respective functions. The steps of an algorithm for each example described in connection with the embodiments disclosed herein may be embodied in hardware or a combination of hardware and computer software. Whether a function is implemented as hardware or computer software driven hardware depends upon the particular application and design constraints imposed on the solution. Those skilled in the art may implement the described functionality using different approaches for each particular application in conjunction with the embodiments, but such implementation is not to be considered as outside the scope of this application.
The present embodiment may divide the functional modules of the electronic device according to the above method example, for example, each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module. The integrated modules described above may be implemented in hardware. It should be noted that, in this embodiment, the division of the modules is schematic, only one logic function is divided, and another division manner may be implemented in actual implementation.
In the case of dividing the respective functional modules with the respective functions, one possible composition diagram of the electronic device involved in the above-described embodiment may include: a display unit, a detection unit and a processing unit. Wherein the display unit, the detection unit and the processing unit cooperate with each other, may be used to support the electronic device to perform the above-described steps, etc., and/or for other processes of the techniques described herein.
It should be noted that, all relevant contents of each step related to the above method embodiment may be cited to the functional description of the corresponding functional module, which is not described herein.
The electronic device provided in this embodiment is configured to perform the above method for collecting audio, so that the same effect as the above implementation method can be achieved.
In case an integrated unit is employed, the electronic device may comprise a processing module, a storage module and a communication module. The processing module may be configured to control and manage actions of the electronic device, for example, may be configured to support the electronic device to execute the steps executed by the display unit, the detection unit, and the processing unit. The memory module may be used to support the electronic device to execute stored program code, data, etc. And the communication module can be used for supporting the communication between the electronic device and other devices.
Wherein the processing module may be a processor or a controller. Which may implement or perform the various exemplary logic blocks, modules, and circuits described in connection with this disclosure. A processor may also be a combination that performs computing functions, e.g., including one or more microprocessors, digital signal processing (digital signal processing, DSP) and microprocessor combinations, and the like. The memory module may be a memory. The communication module can be a radio frequency circuit, a Bluetooth chip, a Wi-Fi chip and other equipment which interact with other electronic equipment.
In one embodiment, when the processing module is a processor and the storage module is a memory, the electronic device according to this embodiment may be a device having the structure shown in fig. 1.
The present embodiment also provides a computer-readable storage medium having stored therein computer instructions that, when executed on an electronic device, cause the electronic device to perform the related method steps described above to implement the method for capturing audio in the above embodiment.
The present embodiment also provides a computer program product which, when run on a computer, causes the computer to perform the above-mentioned related steps to implement the method of capturing audio in the above-mentioned embodiments.
In addition, embodiments of the present application also provide an apparatus, which may be specifically a chip, a component, or a module, and may include a processor and a memory connected to each other; the memory is configured to store computer-executable instructions, and when the device is running, the processor may execute the computer-executable instructions stored in the memory, so that the chip performs the method for capturing audio in the above method embodiments.
The electronic device, the computer readable storage medium, the computer program product or the chip provided in this embodiment are used to execute the corresponding method provided above, so that the beneficial effects thereof can be referred to the beneficial effects in the corresponding method provided above, and will not be described herein.
It will be appreciated by those skilled in the art that, for convenience and brevity of description, only the above-described division of the functional modules is illustrated, and in practical application, the above-described functional allocation may be performed by different functional modules according to needs, i.e. the internal structure of the apparatus is divided into different functional modules to perform all or part of the functions described above.
In the several embodiments provided in this application, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the apparatus embodiments described above are merely illustrative, e.g., the division of modules or units is merely a logical function division, and there may be additional divisions when actually implemented, e.g., multiple units or components may be combined or integrated into another apparatus, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, which may be in electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and the parts shown as units may be one physical unit or a plurality of physical units, may be located in one place, or may be distributed in a plurality of different places. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a readable storage medium. Based on such understanding, the technical solution of the embodiments of the present application may be essentially or a part contributing to the prior art or all or part of the technical solution may be embodied in the form of a software product stored in a storage medium, including several instructions to cause a device (may be a single-chip microcomputer, a chip or the like) or a processor (processor) to perform all or part of the steps of the methods of the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read Only Memory (ROM), a random access memory (random access memory, RAM), a magnetic disk, or an optical disk, or other various media capable of storing program codes.
The foregoing is merely specific embodiments of the present application, but the scope of the present application is not limited thereto, and any person skilled in the art can easily think about changes or substitutions within the technical scope of the present application, and the changes and substitutions are intended to be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (11)

1. A method of capturing audio for application to an electronic device including a display screen, the method comprising:
displaying an image preview interface for video recording, wherein the image preview interface comprises a first control and a second control, the first control is used for controlling audio in the video recording process, and the second control is used for starting or suspending the video recording process;
detecting a first operation acting on the first control, and responding to the first operation, and displaying one or more options on the image preview interface, wherein the one or more options correspond to n audio acquisition devices, and n is an integer greater than or equal to 1;
detecting a second operation acting on a first option, and responding to the second operation, determining an audio acquisition device corresponding to the first option as a target audio acquisition device, or automatically matching the target audio acquisition device from the n audio acquisition devices;
Detecting a third operation acting on the second control, starting to record video in response to the third operation, and acquiring the audio of the target sound source by using the target audio acquisition device in the process of recording video.
2. The method of claim 1, wherein the step of determining the position of the substrate comprises,
the n audio acquisition devices comprise one or more audio acquisition devices configured on the electronic equipment; and/or the number of the groups of groups,
the n audio acquisition devices comprise one or more audio acquisition devices of other devices which establish a connection relationship with the electronic device, and the connection relationship comprises wired connection or wireless connection.
3. The method according to claim 1 or 2, wherein said automatically matching said target audio capture device from said n audio capture devices comprises:
according to the position of the target sound source, determining an audio acquisition device closest to the target sound source among the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the audio acquisition range corresponding to each audio acquisition device, determining the audio acquisition device with the largest audio acquisition range among the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
According to the position of the target sound source and the audio acquisition range corresponding to each audio acquisition device, determining the audio acquisition devices of which the audio acquisition ranges cover the target sound source as the target audio acquisition devices; and/or the number of the groups of groups,
according to the use frequency of the user, determining the audio acquisition device with the highest use frequency of the user in the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the use record of the user, the audio acquisition device which is used by the user for the last time in the n audio acquisition devices is automatically determined to be the target audio acquisition device; and/or the number of the groups of groups,
according to the corresponding performance of each audio acquisition device, determining the audio acquisition device with the highest performance in the n audio acquisition devices as the target audio acquisition device; and/or the number of the groups of groups,
according to the mode that each audio acquisition device is connected with the electronic equipment, the audio acquisition devices which are connected with the electronic equipment in a wired mode in the n audio acquisition devices are determined to be the target audio acquisition devices; and/or the number of the groups of groups,
according to the installation position of a camera used for collecting video pictures in the video recording process, automatically matching the target audio collecting devices from the n audio collecting devices; and/or the number of the groups of groups,
According to the source of a camera used for collecting video pictures in the video recording process, the target audio collecting devices are automatically matched from the n audio collecting devices.
4. The method according to claim 1 to 3, wherein, during video recording, when one or more cameras used for capturing video pictures originate from the electronic device and one or more audio capturing devices configured on the electronic device are included in the n audio capturing devices, there is a preset correspondence between the one or more cameras of the electronic device and the one or more audio capturing devices configured on the electronic device,
and said automatically matching said target audio capture device from said n audio capture devices, comprising:
and according to the corresponding relation between one or more cameras used for acquiring video pictures and the preset, automatically matching the target audio acquisition device from one or more audio acquisition devices configured on the electronic equipment.
5. A method according to any one of claims 1 to 3, wherein, during video recording, when one or more cameras used for capturing video pictures originate from other devices that are in connection with the electronic device, and one or more audio capturing devices of the n audio capturing devices that include other devices that are in connection with the electronic device, the automatically matching the target audio capturing device from the n audio capturing devices includes:
And determining an audio acquisition device of other equipment with one or more camera sources used for acquiring video pictures as the target audio acquisition device.
6. The method of any one of claims 1 to 5, wherein when one or more of the n audio capture devices includes other devices that establish a connection with the electronic device, the method further comprises:
detecting a fourth operation acting on the second option, determining one or more devices in connection with the electronic device in response to the fourth operation, and displaying controls including the one or more devices;
detecting a fifth operation acting on a control corresponding to a target device, and responding to the fifth operation, determining one or more audio acquisition devices of the target device as the target audio acquisition device, and acquiring the audio of a target sound source by using the target audio acquisition device in the process of recording video, wherein the connection relation comprises one or more wireless connection relations of Bluetooth connection, wireless local area network Wi-Fi connection, infrared connection and near field communication NFC connection.
7. The method of any one of claims 1 to 6, wherein the one or more options include a preset third option, the third option is displayed in a selected state, and when the first operation on the first control is not detected, an audio capture device corresponding to the third option is determined to be the target audio capture device.
8. The method of any one of claims 1 to 7, wherein after determining the target audio acquisition device, the method further comprises:
and displaying prompt information, wherein the prompt information is used for prompting the name or source of the target audio acquisition device used by the electronic equipment currently by a user.
9. The method of any one of claims 1 to 8, wherein after determining the target audio acquisition device, the method further comprises:
displaying one or more controls, wherein each control of the one or more controls is used for controlling the working parameters of the target audio acquisition device, and the working parameters comprise one or more of an audio acquisition range, audio acquisition intensity and audio playing effect in the working process of the target audio acquisition device;
Detecting a sixth operation on the one or more controls, and adjusting an operating parameter of the target audio acquisition device in response to the sixth operation.
10. An electronic device, comprising:
a display screen;
one or more audio acquisition devices;
one or more processors;
one or more memories;
a module in which a plurality of application programs are installed;
the memory stores one or more programs that, when executed by the processor, cause the method of any of claims 1-9 to be performed.
11. A computer readable storage medium storing computer instructions which, when run on an electronic device, cause the electronic device to perform the method of any one of claims 1 to 9.
CN202210113229.0A 2022-01-30 2022-01-30 Audio acquisition method and electronic equipment Pending CN116567328A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202210113229.0A CN116567328A (en) 2022-01-30 2022-01-30 Audio acquisition method and electronic equipment
PCT/CN2023/072317 WO2023143171A1 (en) 2022-01-30 2023-01-16 Audio acquisition method and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210113229.0A CN116567328A (en) 2022-01-30 2022-01-30 Audio acquisition method and electronic equipment

Publications (1)

Publication Number Publication Date
CN116567328A true CN116567328A (en) 2023-08-08

Family

ID=87470507

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210113229.0A Pending CN116567328A (en) 2022-01-30 2022-01-30 Audio acquisition method and electronic equipment

Country Status (2)

Country Link
CN (1) CN116567328A (en)
WO (1) WO2023143171A1 (en)

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103888703B (en) * 2014-03-28 2015-11-25 努比亚技术有限公司 Strengthen image pickup method and the camera head of recording
CN104378570A (en) * 2014-09-28 2015-02-25 小米科技有限责任公司 Sound recording method and device
CN113099031B (en) * 2021-02-26 2022-05-17 华为技术有限公司 Sound recording method and related equipment
CN113676592B (en) * 2021-08-02 2023-02-17 Oppo广东移动通信有限公司 Recording method, recording device, electronic equipment and computer readable medium
CN113676668A (en) * 2021-08-24 2021-11-19 维沃移动通信有限公司 Video shooting method and device, electronic equipment and readable storage medium
CN113676687A (en) * 2021-08-30 2021-11-19 联想(北京)有限公司 Information processing method and electronic equipment

Also Published As

Publication number Publication date
WO2023143171A1 (en) 2023-08-03

Similar Documents

Publication Publication Date Title
CN109951633B (en) Method for shooting moon and electronic equipment
CN115473957B (en) Image processing method and electronic equipment
US20220417416A1 (en) Photographing method in telephoto scenario and mobile terminal
CN111628916B (en) Method for cooperation of intelligent sound box and electronic equipment
CN112532892B (en) Image processing method and electronic device
CN114168235B (en) Method for determining function switching entrance and electronic equipment
CN115514883B (en) Cross-equipment collaborative shooting method, related device and system
CN110138999B (en) Certificate scanning method and device for mobile terminal
US20230350629A1 (en) Double-Channel Screen Mirroring Method and Electronic Device
CN114064160A (en) Application icon layout method and related device
CN116708751B (en) Method and device for determining photographing duration and electronic equipment
CN116723415B (en) Thumbnail generation method and terminal equipment
CN113950045B (en) Subscription data downloading method and electronic equipment
CN116561085A (en) Picture sharing method and electronic equipment
CN116567328A (en) Audio acquisition method and electronic equipment
CN115002333B (en) Image processing method and related device
CN115460343B (en) Image processing method, device and storage medium
CN116723382B (en) Shooting method and related equipment
CN115119214B (en) Stereo networking method, system and related device
CN116095512B (en) Photographing method of terminal equipment and related device
CN116567394A (en) Method for sharing multimedia file, transmitting terminal equipment and receiving terminal equipment
CN117336571A (en) Video processing method and related equipment
CN116962564A (en) Directional pickup method and device
CN114827439A (en) Panoramic image shooting method and electronic equipment
CN116777740A (en) Screen capturing method, electronic equipment and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination