WO2022042151A1 - 一种录音方法和设备 - Google Patents

一种录音方法和设备 Download PDF

Info

Publication number
WO2022042151A1
WO2022042151A1 PCT/CN2021/107868 CN2021107868W WO2022042151A1 WO 2022042151 A1 WO2022042151 A1 WO 2022042151A1 CN 2021107868 W CN2021107868 W CN 2021107868W WO 2022042151 A1 WO2022042151 A1 WO 2022042151A1
Authority
WO
WIPO (PCT)
Prior art keywords
recording
audio data
distributed
distributed recording
further configured
Prior art date
Application number
PCT/CN2021/107868
Other languages
English (en)
French (fr)
Inventor
杨枭
朱洲
郑浩
张逸楠
卢梁飞
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to US18/042,808 priority Critical patent/US20230319217A1/en
Priority to EP21859991.8A priority patent/EP4187535A4/en
Publication of WO2022042151A1 publication Critical patent/WO2022042151A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/72412User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/033Voice editing, e.g. manipulating the voice of the synthesiser
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11CSTATIC STORES
    • G11C7/00Arrangements for writing information into, or reading information out from, a digital store
    • G11C7/16Storage of analogue signals in digital stores using an arrangement comprising analogue/digital [A/D] converters, digital memories and digital/analogue [D/A] converters 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/4223Cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/436Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
    • H04N21/4363Adapting the video stream to a specific local network, e.g. a Bluetooth® network
    • H04N21/43637Adapting the video stream to a specific local network, e.g. a Bluetooth® network involving a wireless protocol, e.g. Bluetooth, RF or wireless LAN [IEEE 802.11]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44227Monitoring of local network, e.g. connection or bandwidth variations; Detecting new devices in the local network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4788Supplemental services, e.g. displaying phone caller identification, shopping application communicating with other users, e.g. chatting
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/765Interface circuits between an apparatus for recording and another apparatus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W84/00Network topologies
    • H04W84/02Hierarchically pre-organised networks, e.g. paging networks, cellular networks, WLAN [Wireless Local Area Network] or WLL [Wireless Local Loop]
    • H04W84/10Small scale networks; Flat hierarchical networks
    • H04W84/12WLAN [Wireless Local Area Networks]
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/0017Lossless audio signal coding; Perfect reconstruction of coded audio signal by transmission of coding error
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • G11B2020/10537Audio or video recording
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • G11B2020/10537Audio or video recording
    • G11B2020/10546Audio or video recording specifically adapted for audio data
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • G11B2020/1062Data buffering arrangements, e.g. recording or playback buffers
    • G11B2020/1075Data buffering arrangements, e.g. recording or playback buffers the usage of the buffer being restricted to a specific kind of data
    • G11B2020/10777Data buffering arrangements, e.g. recording or playback buffers the usage of the buffer being restricted to a specific kind of data instructions or commands
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/60Substation equipment, e.g. for use by subscribers including speech amplifiers
    • H04M1/6008Substation equipment, e.g. for use by subscribers including speech amplifiers in the transmitter circuit
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/52Details of telephonic subscriber devices including functional features of a camera
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • H04N9/8211Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a sound signal

Definitions

  • the present application relates to the audio field, and in particular, to a distributed recording method and device.
  • the wireless MIC device may include a transmitter and a receiver.
  • the extra purchase of external recording accessories or wireless MIC equipment is expensive and the size of the equipment is also larger.
  • it needs to be wired or fixed with the mobile phone.
  • the embodiment of the present application provides a recording method.
  • a distributed recording method is adopted, and the receiver sends a distributed recording instruction to one or more transmitters, so that one or more The recording instruction performs distributed recording, and feeds back the audio data obtained from the recording to the receiver.
  • the receiver time-aligns, de-noises and/or decompresses the received one or more audio data, and sends it to the upper layer recording application. It can ensure that users can clearly record the subject's voice, narration and commentary anytime and anywhere without purchasing any external equipment.
  • a distributed recording system includes: at least two terminal devices, and the at least two terminal devices include a first device and at least one second device.
  • the first device and the second device may be mobile phones, tablet computers, wearable devices, and the like.
  • the first device is configured to receive the user's input information, and determine the recording mode according to the input information; the first device is further configured to, when the recording mode is the distributed recording mode, send a distributed recording start instruction to at least one second device;
  • the second device is used to receive an instruction to start distributed recording, and periodically collect first audio data according to the instruction to start distributed recording, wherein the first audio data is local audio data collected by the second device; the second device is also used to: The periodically collected first audio data is subjected to vocal enhancement and/or noise reduction processing, and then sent to the first device in the same period; the first device is also used to receive at least one first audio data, so that the camera application calls at least one first audio data.
  • the receiver controls one or more transmitters to perform recording, and sends the audio data obtained from the recording to the upper-layer recording application. It ensures that users can clearly record the subject's voice, narration and commentary anytime and anywhere without purchasing any external equipment.
  • the second device is further configured to: determine recording parameters according to the instruction for starting the distributed recording; and collect the first audio data according to the recording parameters.
  • the second device before the second device sends the periodically collected first audio data to the first device in the same period, the second device is further configured to: perform the periodically collected first audio data Compression processing is performed to obtain compressed first audio data.
  • the present application compresses the collected audio data, which can effectively reduce the data size in the transmission process and ensure the security of the data in the transmission process.
  • the first device is further configured to: when the first audio data is compressed data, perform audio decompression on the compressed first audio data to obtain decompressed first audio data.
  • the present application can also decompress the received compressed audio data, and at the same time reduce the amount of data in the transmission data, and ensure that it can be restored without loss after receiving.
  • the distributed recording mode includes: a multi-camera cooperative recording mode; when the recording mode is a multi-camera cooperative recording mode, before the camera application calls the at least one first audio data, the first device is further used for: To collect the second audio data, the camera application calls at least one of the first audio data and the second audio data, where the second audio data is local audio data collected by the first device.
  • the present application can also be combined with the local audio data collected by the receiver, which ensures that the voice of the shooting object, as well as the narration and commentary can be clearly recorded in the process of shooting the video.
  • the first device before the camera application calls the at least one piece of first audio data, the first device is further configured to: when there are multiple pieces of first audio data, perform a time alignment operation on the multiple pieces of first audio data; or When the recording mode is the multi-machine cooperative recording mode, a time alignment operation is performed on at least one of the first audio data and the second audio data.
  • time alignment is performed on the multiple audio data, so as to avoid the time error caused by transmission between different audios.
  • the first device before the camera application calls the at least one second audio data, the first device is further configured to: perform noise reduction and/or vocal enhancement processing on the second audio data.
  • the application can also perform preprocessing on audio such as noise reduction and vocal enhancement, so that the audio uploaded to the camera application is clearer.
  • the distributed recording mode includes: a first sub-distributed recording mode and a second sub-distributed recording mode; when the recording mode is the first sub-distributed recording mode, the first device is also used to send Distributed recording instructions to a second device, and receiving the first audio data sent by the second device; or, when the recording mode is the second sub-distributed recording mode, the first device is also used to send distributed recording instructions to multiple devices. a second device, and receive first audio data sent by a plurality of second devices.
  • the system further includes: the first device is further configured to send a distributed recording request to at least one second device; The second device is further configured to receive a distributed recording request sent by the first device; the second device is further configured to display prompt information on the display screen of the second device, where the prompt information is used to prompt whether to allow the second device to perform distributed recording recording; the second device is further configured to send a distributed recording response message to the first device, where the distributed recording response message is used to indicate whether the second device is allowed to perform distributed recording; the first device is further configured to receive at least one first device The distributed recording response message sent by the second device; the first device is further configured to send an instruction to start distributed recording to at least one second device that allows distributed recording.
  • the recording mode further includes a local recording mode; the first device is further configured to, according to the input information input by the user, record in the local recording mode, the first sub-distributed recording mode, and the second sub-distributed recording mode Switch to and from the multi-camera cooperative recording mode.
  • the first device is further configured to: display a distributed recording icon on a display screen of the first device.
  • the second device is further configured to: display a distributed recording prompt icon on the display screen of the second device.
  • a recording terminal device is provided, where the recording terminal device is the first device or the second device; the recording terminal device includes: a display, a processor, a memory, a transmitter and a receiver; when the recording terminal device is the first device When the recording mode is the distributed recording mode, the receiver is used to receive the input information of the user and determine the recording mode according to the input information; the processor is used to control the transmitter to send an instruction to start the distributed recording to at least one second when the recording mode is the distributed recording mode.
  • the receiver when the recording terminal device is the second device, the receiver is used to receive an instruction to start distributed recording, and periodically collect first audio data according to the instruction to start distributed recording, wherein the first audio data is collected by the second device The received local audio data; the transmitter is used for periodically collecting the first audio data, performing vocal enhancement and/or noise reduction processing on the first audio data, and then sending it to the first device in the same period; when When the recording terminal device is the first device, the receiver is further configured to receive at least one first audio data; the processor is further configured to control the camera application to call the at least one first audio data. In this application, the receiver controls one or more transmitters to perform recording, and sends the audio data obtained from the recording to the upper-layer recording application. It ensures that users can clearly record the subject's voice, narration and commentary anytime and anywhere without purchasing any external equipment.
  • the processor when the recording terminal device is the second device, the processor is configured to: determine the recording parameters according to the instruction to start the distributed recording; and control the microphone to collect the first audio data according to the recording parameters.
  • the processor when the recording terminal device is the second device, the processor is further configured to: compress the periodically collected first audio data to obtain compressed first audio data.
  • the present application compresses the collected audio data, which can effectively reduce the data size in the transmission process and ensure the security of the data in the transmission process.
  • the processor when the recording terminal device is the second device, the processor is further configured to: when the first audio data is compressed data, perform audio decompression on the compressed first audio data to obtain decompression After the first audio data.
  • the present application can also decompress the received compressed audio data, and at the same time reduce the amount of data in the transmission data, and ensure that it can be restored without loss after receiving.
  • the distributed recording mode includes: a multi-machine collaborative recording mode; when the recording terminal device is the first device, the processor is further configured to: when the recording mode is the multi-machine collaborative recording mode, collect the second For audio data, the camera application calls at least one of first audio data and second audio data, where the second audio data is local audio data collected by the first device.
  • the present application can also be combined with the local audio data collected by the receiver, which ensures that the voice of the shooting object, as well as the narration and commentary can be clearly recorded in the process of shooting the video.
  • the processor when the recording terminal device is the first device, the processor is further configured to: when there are multiple first audio data, perform a time alignment operation on the multiple first audio data; or when the recording mode In the multi-machine collaborative recording mode, a time alignment operation is performed on at least one of the first audio data and the second audio data.
  • time alignment is performed on the multiple audio data, so as to avoid the time error caused by transmission between different audios.
  • the processor when the recording terminal device is the first device, the processor is further configured to: perform noise reduction and/or vocal enhancement processing on the second audio data.
  • the application can also perform preprocessing on audio such as noise reduction and vocal enhancement, so that the audio uploaded to the camera application is clearer.
  • the distributed recording mode includes: a first sub-distributed recording mode and a second sub-distributed recording mode; when the recording terminal device is the first device, the transmitter is also used for, when the recording mode is In the first sub-distributed recording mode, send distributed recording instructions to a second device; or, the transmitter is also used to, when the recording mode is the second sub-distributed recording mode, send distributed recording instructions to a plurality of second devices.
  • the receiver is also used to, when the recording mode is the first sub-distributed recording mode, receive the first audio data sent by the second device; or, when the recording mode is the second sub-distributed recording mode, receive multiple The first audio data sent by the second device.
  • the transmitter when the recording terminal device is the first device, the transmitter is further configured to send a distributed recording request to at least one second device; when the recording terminal device is the second device, the receiver is further configured to use in, receiving a distributed recording request sent by the first device; the display is also used to display prompt information on the display screen of the second device, and the prompt information is used to prompt whether to allow the second device to perform distributed recording;
  • the transmitter is further configured to send a distributed recording response message to the first device, where the distributed recording response message is used to indicate whether the second device is allowed to perform distributed recording;
  • the receiver is further configured to receive a distributed recording response message sent by at least one second device;
  • the transmitter is further configured to send an instruction to start distributed recording to at least one second device that allows distributed recording.
  • the recording mode further includes a local recording mode; when the recording terminal device is the first device, the processor is further configured to, according to the input information input by the user, record in the local recording mode, the first sub-distributed recording mode, the second sub-distributed recording mode and the multi-machine collaborative recording mode.
  • the display is configured to: display the distributed recording icon on the display screen of the first device.
  • the display is configured to: display a distributed recording prompt icon on the display screen of the second device.
  • a computer-readable storage medium where instructions are stored in the computer-readable storage medium, and when the instructions are executed on a terminal, the terminal is made to execute any one of the methods of the first aspect.
  • a computer program device comprising instructions which, when run on a terminal, cause the terminal to perform the method of any one of the first aspects.
  • the present application discloses a distributed recording method and device, in which a distributed recording instruction is sent through a receiver to control one or more transmitters to perform recording, and the audio data obtained from the recording is sent to an upper-layer recording application. It ensures that users can clearly record the subject's voice, narration and commentary anytime and anywhere without purchasing any external equipment.
  • FIG. 1 is a schematic diagram of a distributed recording scene provided by an embodiment of the present application.
  • Figure 2a is a schematic diagram of a professional recording device
  • Figure 2b is a schematic diagram of the use of the recording device shown in Figure 2a;
  • Figure 3 is a schematic diagram of another professional recording equipment
  • FIG. 4 is a schematic diagram of a one-to-one distributed recording scenario provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a receiver architecture provided by an embodiment of the present application.
  • FIG. 6 is a schematic diagram of a camera interface of a terminal device provided by an embodiment of the present application.
  • FIG. 7 is a schematic diagram of a transmitter architecture provided by an embodiment of the present application.
  • FIG. 8 is a schematic diagram of an interface for displaying prompt information provided by an embodiment of the present application.
  • FIG. 9 is a schematic diagram of a distributed recording interface provided by an embodiment of the present application.
  • FIG. 10 is a schematic diagram of a one-to-many distributed recording scenario provided by an embodiment of the present application.
  • FIG. 11 is a schematic diagram of another receiver architecture provided by an embodiment of the present application.
  • FIG. 12 is a schematic diagram of a multi-machine collaborative distributed recording scenario provided by an embodiment of the present application.
  • FIG. 13 is a schematic diagram of another receiver architecture provided by an embodiment of the present application.
  • 16 is a flowchart of another distributed recording method provided by an embodiment of the present application.
  • FIG. 17 is a schematic diagram of a distributed recording device according to an embodiment of the present application.
  • This application is mainly applied to the scene of using a terminal device to shoot video.
  • a terminal device to shoot video.
  • the photographer is shooting a video of the subject, and the distance between the subject and the photographer is relatively far. Therefore, the camera equipment used by the photographer cannot clearly The voice of the subject is captured and recorded.
  • the camera equipment used by the photographer cannot clearly capture and record the voice of the narration and commentary.
  • some external professional recording accessories are usually purchased, such as shown in Figure 2a.
  • Such professional recording accessories are connected to the terminal device through a wired connection, and send audio data streams to the terminal device.
  • such professional recording accessories can be controlled by the terminal device.
  • the connection mode is shown in Fig. 2b, for example, and it can be seen that the professional recording device shown in Fig. 2a is connected to the terminal device below through a wired connection.
  • an application application, APP
  • information such as waveform and spectrum can also be displayed on the terminal device.
  • the size of the MIC device used in such professional recording accessories is closely related to the performance index. It focuses on performance indicators such as sensitivity, bandwidth, frequency response flatness, acoustic overload point and so on.
  • Such professional recording accessories usually use larger MIC devices or directional MIC devices, and have strong signal processing capabilities.
  • Such professional recording accessories process the audio data collected by the MIC first inside the accessory, and then transmit the audio data to the connected terminal through a universal serial bus (USB) interface or other equivalent interface. on the device.
  • USB universal serial bus
  • a signal processing algorithm can also be used to perform secondary processing on the processed audio data, thereby realizing diversified recording functions.
  • the portability is poor.
  • the device needs to be connected to the terminal device, and the MIC parameters such as direction and gain are manually adjusted through the terminal device. After use, it needs to be removed, which makes the operation process too complicated.
  • the shooting distance is long, or the environment is relatively noisy, the recording effect is not very good, so this type of equipment is only suitable for professional users, not for ordinary users.
  • a wireless MIC scheme which includes a transmitter and a receiver.
  • the subject can wear a transmitter, such as a lavalier MIC with wireless function, at this time, the wireless lavalier MIC can be fixed on the collar of the subject in order to pick up the human voice, such as the left half of Figure 3 shown.
  • the receiver can be wired to the terminal device.
  • the solution picks up audio data through the wireless lavalier MIC and transmits it to the receiver wirelessly, and the receiver transmits the received audio data to the connected terminal device through the USB interface or other equivalent interface.
  • a receiver can be connected to 1 or 2 transmitters.
  • the wireless lavalier MIC and receiver have a single function and are bulky, and users need to purchase them separately.
  • Good performance is usually more than 500 yuan, while more professional models can be as high as several thousand yuan.
  • it needs to be connected to the terminal equipment when used in combination with the collection, and the receiver and transmitter should be disassembled after use. This makes it only suitable for professional users, not suitable for ordinary users to apply anytime, anywhere.
  • the present application proposes a method for turning a terminal device into a wireless recording MIC, including turning some terminal devices into transmitters, and turning a certain terminal device into a receiver.
  • a user shoots through a terminal device, there is no need to purchase other devices separately, but another terminal device can be used for peripheral sound pickup.
  • the receiver sends distributed recording instructions, controls one or more transmitters to record, and sends the audio data obtained from the recording to the upper-layer recording application. It ensures that users can clearly record the subject's voice, narration and commentary anytime and anywhere without purchasing any external equipment, so as to achieve a professional recording effect.
  • FIG. 4 is a schematic diagram of a one-to-one distributed recording scenario provided by an embodiment of the present application.
  • one receiver is used to match one transmitter.
  • the photographer uses the receiver to shoot video.
  • the transmitter is located around the subject, so that the voice of the subject can be clearly picked up.
  • the receiver may be the first device, and the transmitter may be the second device.
  • “first” and “second” do not limit their order, but are only used to distinguish different terminal devices.
  • the transmitter and receiver in this application need to be pre-connected in the same wireless network environment, for example, through any wireless fidelity (wireless fidelity, WiFi), Bluetooth (bluetooth, BT), ZigBee (ZigBee), etc.
  • the wireless transmission method is connected to the same wireless network environment. In other words, the transmitter and receiver need to be in the same wireless local area network.
  • the present application provides a schematic diagram of a receiver architecture shown in FIG. 5 .
  • the receiver architecture can be applied to terminal equipment.
  • the terminal equipment involved in this application may be, but is not limited to, mobile phones, televisions, stereos, wearable devices, tablet computers, personal digital assistants (PDAs), laptop computers (laptops), mobile computers, augmented reality (augmented reality) , AR) equipment, virtual reality (virtual reality, VR) equipment, artificial intelligence (artificial intelligence (AI) equipment and other terminal equipment or portable terminal equipment.
  • PDAs personal digital assistants
  • AR augmented reality
  • VR virtual reality
  • AI artificial intelligence
  • the receiver architecture includes an APP layer, a framework layer and a hardware layer.
  • the APP layer includes a camera APP, and of course any other possible APP, which is not limited in this application.
  • the solution is described from the perspective of user usage.
  • the user first selects a recording mode through the receiver, such as a distributed recording mode, a multi-machine cooperative recording mode or a local recording mode.
  • the receiver can send a distributed recording request to the wireless network where its own device is located, for example, it can send it in the form of broadcasting.
  • the transmitter in the same wireless network receives the distributed recording request, it can confirm whether to allow distributed recording or multi-machine collaborative recording according to the user of the transmitter.
  • the transmitter determines that the transmitter is allowed to perform distributed recording or multi-machine cooperative recording, the transmitter can send a response message to the wireless network where it is located.
  • the receiver can identify one or more transmitters that allow recording.
  • the receiver can send an instruction to start distributed recording or multi-machine cooperative recording, so that one or more confirmed transmitters can record, and periodically send the recorded audio data to the receiver.
  • the transmitter sends audio data in real time
  • the receiver also receives audio data in real time.
  • the receiver receives the audio data sent by the transmitter, it can be called by the upper-layer camera APP of the receiver.
  • the receiver can also upload the received audio data in real time, such as uploading to a server, so as to realize functions such as online live broadcast.
  • the terminal device used by the photographer at this time can be called a receiver.
  • the photographer can first open the camera APP on the receiver, and then click the distributed recording icon displayed on the interface according to the interface displayed by the receiver to select the corresponding recording mode.
  • the recording mode can include distributed recording and local recording.
  • the distributed recording icon may also be displayed in some options of the setting menu, or the distributed recording icon may be displayed on some other possible user interfaces (user interface, UI).
  • the recording modes can include distributed recording mode, multi-machine collaborative recording mode and local recording mode. In one example, if the photographer does not click the distributed recording icon to select the recording mode, the camera APP can select the local recording mode as the default recording mode.
  • FIG. 6 shows a schematic diagram of a camera interface of a terminal device.
  • the interface is a display interface for the user to use the camera APP on the receiver to perform video recording or live broadcast.
  • the photographer can select the recording mode through the distributed recording icon 601 at the lower left of the interface.
  • the action of selecting the recording mode through the distributed recording icon 601 may also occur before the video recording or live broadcast. That is to say, the photographer can select the corresponding recording mode through the distributed recording icon 601 in advance, and then perform video recording or live broadcast. The photographer can switch between different recording modes by clicking the distributed recording icon 601 .
  • the photographer can select the distributed recording mode by clicking the distributed recording icon 601 . Since this scenario only includes one transmitter and one receiver, in this scenario, the distributed recording mode can be the first sub-distributed recording mode, which is used to indicate that one transmitter and one receiver are wirelessly connected. Connect, and make distributed recordings.
  • the camera interface shown in FIG. 6 may also include a button 602 for starting/stopping recording, a button 603 for pausing recording, a button 604 for shooting a certain frame in the video, and a button for adjusting the size of the picture.
  • the interface may also include video recording time, and a taskbar located at the top of the interface.
  • the task bar may include information such as operator, WiFi, power, and time.
  • other possible icons or information may also be included in the task bar, which is not limited in this application again.
  • the camera interface may also include any other icons with certain specific functions other than the icons shown in the figure. It can be understood that FIG. 6 only illustrates a possible form of the camera interface, and the styles of the distributed recording icons and other icons can be arbitrarily designed according to the actual situation, which is not limited in this application.
  • the camera APP sends a mode selection instruction to the distributed recording system configuration module in the framework layer, wherein the mode selection instruction can be used for.
  • the distributed recording system configuration module may determine that the recording mode selected by the photographer is the distributed recording mode. After that, the distributed recording system configuration module generates a distributed recording request according to the first instruction, and sends the distributed recording request to the WiFi interface. After the WiFi interface receives a distributed recording request, it forwards the request to the WiFi subsystem at the hardware layer. And send the request to one or more transmitters around the subject through the WiFi subsystem.
  • the WiFi interface is a software interface abstracted from the framework layer.
  • the WiFi subsystem may include a hardware interface, a WiFi protocol stack and an antenna.
  • the WiFi protocol stack involved in the present application may be a hardware-substantiated WiFi protocol stack, which is used to perform corresponding processing on data, so that the processed data satisfies the WiFi protocol format. More specifically, the WiFi interface forwards the distributed recording request to the hardware interface of the hardware layer. After receiving the distributed recording request, the hardware interface forwards the distributed recording request to the WiFi protocol stack, and performs corresponding processing on the distributed recording request in the WiFi protocol stack, and sends it to the transmitter through the antenna. It can be understood that, for the specific manner of correspondingly processing the distributed recording request in the WiFi protocol stack, reference may be made to the current known technology, which will not be repeated here for the convenience of description.
  • the distributed recording system configuration module After the distributed recording system configuration module receives the first instruction, it can also generate a distributed recording instruction, and send the distributed recording instruction to the multi-machine recording algorithm module.
  • the multi-machine recording algorithm module is used to receive audio data from different terminal devices, perform audio decompression, recording processing, time alignment and other operations on one or more received audio data, and send the processed audio data. to the camera APP of the application layer.
  • the multi-camera recording algorithm module can be configured according to the instructions sent by the distributed recording system configuration module. For example, after receiving the distributed recording instruction, the multi-camera recording algorithm module will configure according to the distributed recording instruction , select the output of the WiFi interface as the input of the audio data stream, but not the output of the local recording algorithm module as the input.
  • the WiFi subsystem of the receiver may also receive distributed recording response messages sent from one or more transmitters.
  • the distributed recording response message sent from only one transmitter may be received.
  • the distributed recording scene at this time is a distributed recording scene of one transmitter and one receiver. More specifically, the distributed recording response message sent by a certain transmitter can be received through the antenna. After the antenna receives the distributed recording response message, it can process the received distributed recording response message through the WiFi protocol stack, and then transmit the processed distributed recording response message to the WiFi in the framework layer through the hardware interface. interface.
  • the WiFi interface sends the distributed recording response message to the distributed recording system configuration module, so that the distributed recording system configuration module can determine which transmitters allow distributed recording according to the distributed recording response message.
  • the distributed recording system configuration module After the distributed recording system configuration module determines that a certain transmitter is allowed to perform distributed recording according to the distributed recording response message, the distributed recording system configuration module generates a distributed recording start command, and sends the distributed recording start command to the WiFi interface .
  • the above-mentioned instruction for generating and starting distributed recording may be that the receiver sends out prompt information through the display screen, so that the user can determine whether to start distributed recording immediately according to the prompt information. If the receiver receives an operation instruction that the user chooses to start the distributed recording immediately, it can control the configuration module of the distributed recording system to generate an instruction to start the distributed recording.
  • the distributed recording system configuration module of the receiver may actively generate an instruction to start distributed recording after determining that a certain transmitter is allowed to perform distributed recording according to the distributed recording response message.
  • the configuration module of the distributed recording system of the receiver may receive the operation of immediately starting the distributed recording that is actively triggered by the user after determining that a certain transmitter is allowed to perform the distributed recording according to the distributed recording response message. instruction.
  • the receiver may not prompt the user whether to start distributed recording immediately, but the user actively provides an operation instruction to start distributed recording through physical or virtual buttons on the touch screen.
  • the WiFi interface After receiving the instruction to start distributed recording, the WiFi interface forwards the instruction to start distributed recording to the WiFi subsystem of the hardware layer. And send the start distributed recording instruction to the transmitter that allows distributed recording through the WiFi subsystem. It can be understood that the process of sending an instruction to start distributed recording is the same as sending a distributed recording request, which is not repeated here for convenience of description.
  • the WiFi subsystem of the receiver After the WiFi subsystem of the receiver sends the instruction to start the distributed recording, it can also receive the audio data sent from the transmitter that allows the distributed recording. More specifically, the audio data sent by the transmitter can be received through the antenna. Then, after receiving the audio data, the antenna can process the received audio data correspondingly through the WiFi protocol stack, and then transmit the processed audio data to the WiFi interface in the framework layer through the hardware interface. It can be understood that, for the specific manner of correspondingly processing the received audio data in the WiFi protocol stack, reference may be made to the current known technology, which will not be repeated here for the convenience of description.
  • the WiFi interface in the framework layer After receiving the processed audio data, the WiFi interface in the framework layer directly forwards the audio data to the multi-camera recording algorithm module. So that one or more sub-modules in the multi-camera recording algorithm module can process the received audio data.
  • the audio decoding sub-module in the multi-camera recording algorithm module can first decompress the compressed audio data to obtain the original audio data, and then the obtained original audio data is processed by the recording processing sub-module for recording processing.
  • the process of compressing audio data can also be regarded as a process of encoding audio data
  • the process of decompressing the compressed audio data can also be regarded as a process of decoding.
  • the recording processing may include noise reduction, vocal enhancement and other possible operations on the audio data.
  • the audio data processed by the recording is directly sent to the camera APP of the application layer.
  • the multi-camera recording algorithm module can send the processed audio data to the camera APP at the application layer through the recording record source (audio record.source) interface, so that the camera APP can use it.
  • the audio data For example, the audio data and the camera image are combined into a video file, or the audio data and the camera image are combined into a video file and then forwarded to other terminal devices through existing methods, so as to realize live video.
  • the receiver and the transmitter are connected in the same wireless local area network through WiFi, the receiver can transmit the synthesized video files to a server or directly to other terminal devices through other network connection methods. For example, it can be via cellular network, Bluetooth or ZigBee, etc.
  • WiFi direct technology assuming that the receiver and the transmitter use the 2.4GHz frequency band for WiFi interconnection
  • the data exchange with the WAN can also be realized through the 5GHz frequency band.
  • the receiver and the transmitter are connected through other wireless methods such as Bluetooth or ZigBee, the receiver can be directly connected to the WAN through WiFi in the 2.4GHz frequency band or the 5GHz frequency band.
  • the audio decoding sub-module is an optional sub-module. If the audio data sent by the transmitter has not been compressed, the audio data can be directly processed by the recording processing sub-module for recording processing.
  • the distributed recording system configuration module receives the mode selection instruction sent by the camera APP, and may also be a second instruction for informing the distributed recording system configuration module that the user has selected the local recording mode.
  • the distributed recording system configuration module may determine that the recording mode selected by the photographer is the local recording mode. After that, the distributed recording system configuration module may also generate a local recording instruction according to the second instruction, and send the local recording instruction to the multi-machine recording algorithm module. For example, after receiving the local recording instruction, the multi-camera recording algorithm module will configure according to the local recording instruction, and select the output of the local recording algorithm module as the input of the audio data stream.
  • the camera APP can directly send the local recording control command to the local recording algorithm module through the set parameter interface, so that the local algorithm module can control the acquisition of one or more MICs at the hardware layer. analog signal to audio data.
  • the local recording control instruction can be one byte in length.
  • the local recording algorithm module receives the local recording control command, it can control all the MICs on the receiver to collect audio data.
  • the number of MICs is 2-4, such as 3 MICs in one example.
  • the local recording algorithm module can also control the adjustment of the pickup range, noise intensity, etc. of the MIC on the receiver according to the local recording control instructions.
  • the pickup range can be greater than 3 meters and so on.
  • the MIC converts the collected analog signal into a digital signal through an analog-to-digital converter, and then transmits the digital signal of the audio data to the local recording algorithm module in the framework layer for corresponding processing.
  • the local recording algorithm module processes the digital signal of the audio data, for example, it may be any possible processing manner of one or several combinations of pre-emphasis, filtering, reverberation, and the like.
  • the local recording algorithm module can transmit audio data to the multi-camera recording algorithm module through the audio path established with the multi-camera recording algorithm module. At this time, the multi-camera recording algorithm module has no function and is only used to forward the audio data to the APP.
  • the analog-to-digital converter may use a codec (coder-decoder, Codec) to convert the collected analog signal into a digital signal.
  • codec coder-decoder, Codec
  • the multi-camera recording algorithm module will select the output of the local recording algorithm module as the default audio data stream input.
  • the present application also provides a schematic diagram of a transmitter architecture shown in FIG. 7 .
  • the transmitter architecture can be applied to terminal equipment.
  • the terminal device can be, but is not limited to, any terminal device or portable terminal device such as mobile phone, wearable device, tablet computer, PDA, laptop computer, mobile computer, AR device, VR device, AI device, etc.
  • the transmitter architecture includes an APP layer, a framework layer and a hardware layer.
  • the WiFi subsystem at the hardware layer can receive the distributed recording request sent by the receiver. More specifically, for example, the transmitter can receive the distributed recording request sent by the receiver through the antenna. and send the distributed recording request to the WiFi protocol stack. The WiFi protocol stack then forwards the distributed recording request to the WiFi interface in the framework layer through the hardware interface. The WiFi interface can forward the received distributed recording request to the distributed recording transmitter configuration module. After receiving the distributed recording request, the distributed recording transmitter configuration module can send prompt information to the APP layer, and the prompt information is displayed on the display screen to prompt the user whether to allow the terminal device to perform distributed recording as a transmitter. For example, as shown in FIG.
  • a pop-up window for prompting appears on the display screen interface, and the pop-up window may include text, graphics, patterns and other information to prompt the user whether to perform distributed recording.
  • the pop-up window may include text, graphics, patterns and other information to prompt the user whether to perform distributed recording.
  • the APP layer will send the first information to the distributed recording transmitter configuration module in the framework layer according to the user's operation on the display screen.
  • the distributed recording transmitter configuration module may determine the terminal device as a transmitter according to the first information, and generate a distributed recording response message. Then, the distributed recording transmitter configuration module sends the distributed recording response message to the WiFi subsystem of the hardware layer through the WiFi interface, and sends it to the receiver through the WiFi subsystem, so that the receiver can send the distributed recording response message to the transmitter according to the distributed recording response message.
  • the computer sends an instruction to start distributed recording.
  • the APP layer will send the second information to the distributed recording transmitter configuration module in the framework layer according to the user's operation on the display screen.
  • the distributed recording transmitter configuration module may determine that the terminal device is not a transmitter according to the second information, and generate a distributed recording response message.
  • the transmitter's WiFi subsystem can also receive commands sent by the receiver to initiate distributed recording. More specifically, for example, the transmitter may receive an instruction to start distributed recording sent by the receiver through an antenna. And send the start distributed recording instruction to the WiFi protocol stack. The WiFi protocol stack then forwards the start distributed recording instruction to the WiFi interface in the framework layer through the hardware interface. The WiFi interface can forward the received instruction to start the distributed recording to the configuration module of the distributed recording transmitter.
  • the distributed recording transmitter configuration module after receiving the instruction to start the distributed recording, can send the instruction to start the distributed recording to the local recording algorithm module through the setparameter interface.
  • the local recording algorithm module After receiving the instruction to start the distributed recording, the local recording algorithm module , the algorithm parameters can be adjusted according to the instruction to start distributed recording to meet the needs of distributed recording.
  • the command to start distributed recording can be 1 byte in length.
  • the local recording algorithm module can control some MICs on the receiver to collect analog signals of audio data.
  • one or two MICs can be controlled, wherein if one MIC is controlled, it can be the MIC located on the top of the terminal device, and if two MICs are controlled, it can be the MIC located on the top and the bottom MIC of the terminal device.
  • the local recording algorithm module can also control the adjustment of the pickup range, noise intensity, etc. of the MIC on the receiver according to the instruction to start the distributed recording. For example, single MIC noise reduction or dual MIC noise reduction, the pickup range can be within 1 meter and so on.
  • the MIC After the MIC collects the analog signal of the audio data, it can send the analog signal of the audio data to the analog-to-digital converter, convert the collected analog signal into a digital signal, and then send the digital signal of the audio data of the machine to the machine Recording algorithm module.
  • the local recording algorithm module can also process the digital signal of the audio data, for example, it can be any possible processing method of one or several combinations of pre-emphasis, filtering, reverberation, etc.
  • the digital signal of the audio data is sent to the distributed recording transmitter configuration module.
  • the local recording algorithm module can also compress the audio data collected by the local machine. The compressed audio data is then sent to the distributed recording transmitter configuration module.
  • the transmitter can collect audio data periodically, and when the period is short, such as several milliseconds, several microseconds or even several nanoseconds, it can be regarded as a real-time acquisition of audio signals. It can be understood that the transmitter may transmit the audio signals collected periodically in the same period, that is, in one period, the transmitter collects the audio signals and sends them to the receiver. For the receiver, the audio signal can then be received in the same period.
  • the transmitter For the receiver, if the transmitter is wired or wirelessly connected with headphones. At this time, you can use the MIC of the headset to pick up the sound. In the case of the subject, the transmitter can be placed in a pocket or other location such as a clothing pocket. At this point the transmitter can be used as a "wireless lavalier microphone".
  • the transmitter configuration module can send the data related to the distributed recording prompt icon to the APP layer after receiving the instruction to start the distributed recording, and display the distributed recording prompt icon on the APP layer.
  • the task bar above the display screen of the terminal device shown in FIG. 9 may also have a distributed recording prompt icon 801 .
  • the distributed recording prompt icon 801 is used to indicate that the terminal device as a transmitter has entered the distributed recording mode and is performing distributed recording. It can be understood that FIG. 9 only illustrates a possible distributed recording prompt icon 801 , and the specific style and display position of the distributed recording prompt icon 801 can be arbitrarily designed according to the actual situation, which is not limited in this application.
  • the distributed recording transmitter configuration module can also send the received audio data to the WiFi subsystem of the hardware layer through the WiFi interface.
  • the WiFi subsystem can also be used to send audio data to the receiver. More specifically, for example, the WiFi interface sends the audio data to the hardware interface located at the hardware layer, and then transmits the audio data to the WiFi protocol stack through the hardware interface, and performs corresponding processing on the audio data in the WiFi protocol stack, and then the WiFi protocol The stack sends the processed audio data to the receiver via the antenna. It can be understood that, for the specific manner of correspondingly processing the audio data in the WiFi protocol stack, reference may be made to the current known technology, which will not be repeated here for the convenience of description.
  • the receiver can also receive distributed recording response messages sent by multiple transmitters, and control the multiple transmitters to pick up sounds.
  • the distributed recording scene at this time is the distributed recording scene of multiple transmitters to one receiver.
  • the photographer shoots video through the receiver.
  • the transmitter can also be located in other positions, such as transmitter 3, which can be used for recording narration or commentary.
  • the transmitter may also be allocated to multiple photographed objects, and the position of the transmitter is not specifically limited in this application.
  • the distributed recording mode can be the second sub-distributed recording mode, which is used to indicate that multiple transmitters and one receiver are wirelessly connected. Connect, and make distributed recordings.
  • the transmitter is the same as the transmitter shown in FIG. 7 and FIG. 9 , for details, reference may be made to the corresponding description, which will not be repeated in this application.
  • the receiver architecture compared to the receiver architecture shown in Figure 5, the difference is only in the multi-camera recording algorithm module.
  • FIG. 11 is a schematic diagram of another receiver architecture provided by an embodiment of the present application. Since the scene in Fig. 10 is a distributed recording scene of one receiver and multiple transmitters, since the local recording mode is not involved, Fig. 11 does not show the local recording algorithm module, the corresponding analog-to-digital converter and multiple MICs . For a receiver in a distributed recording scenario with one receiver and multiple transmitters, after the distributed recording system configuration module receives the first instruction sent by the camera APP, it can transmit the distributed recording request to the The WiFi sub-module at the hardware layer sends distributed recording instructions to multiple transmitters through the WiFi sub-module.
  • the WiFi sub-module of the receiver can also receive audio data sent from multiple transmitters. And transmit the received multiple audio data to the WiFi interface of the framework layer, and transmit the received multiple audio data to the multi-machine recording algorithm module through the WiFi interface. Since the transmitters may be located at different positions, the transmission time of the audio data over the wireless will result in different transmission times, and therefore the reception times will also be different. Therefore, for the audio data received at different times, the time alignment sub-module in the multi-camera recording algorithm module can be used to time-align multiple audio data, so that the time sequence of the output audio data is correct and avoid mutual interference.
  • the time alignment of a plurality of audio data can be realized through corresponding detection and delay buffers, and the specific implementation method can adopt the existing method, which will not be repeated here.
  • the recording processing sub-module in the multi-machine recording algorithm module can be enabled to perform recording processing on the received multiple audio data. Possible operations such as noise reduction, reverberation reduction, and vocal enhancement are possible.
  • Possible operations such as noise reduction, reverberation reduction, and vocal enhancement are possible.
  • the audio decoding sub-module in the multi-camera recording algorithm module can also be enabled to decompress the compressed audio data, and Get raw audio data.
  • the order of time alignment and recording processing for multiple audio data can be adjusted according to the actual situation, for example, time alignment of multiple audio data is performed before recording processing, or multiple audio data is recorded first. The processing is then time aligned.
  • time alignment of multiple audio data is selected before the recording processing is performed.
  • the camera APP in FIG. 11 can also display a distributed recording icon on the display screen, which is exactly the same as the display mode of the camera APP in FIG. 5 .
  • a distributed recording icon on the display screen, which is exactly the same as the display mode of the camera APP in FIG. 5 .
  • FIG. 12 is a schematic diagram of a multi-machine collaborative distributed recording scenario provided by an embodiment of the present application.
  • Fig. 12 is compared with Fig. 10.
  • the receiver that is used as a video capture is also recording at the same time.
  • the receiver as the video shooting also includes multiple transmitters to record at the same time. Obviously, in the scene shown in Figure 12, the receiver and multiple transmitters cooperate to record, so as to ensure that the voice of the subject, as well as the narration and commentary can be clearly recorded.
  • the present application provides yet another schematic diagram of a receiver architecture.
  • the camera APP can display a distributed recording icon on the display screen, so that the user can select different recording modes through the distributed recording icon when using it. Recording modes can include distributed recording, multi-machine collaborative recording and local recording.
  • the specific display mode is exactly the same as the display mode of the camera APP in FIG. 5 . For details, please refer to the corresponding description, which will not be repeated here.
  • the distributed recording system configuration module receives the mode selection instruction sent by the camera APP, and may also be a third instruction for informing the distributed recording system configuration module that the user has selected the multi-machine collaborative recording mode.
  • the distributed recording system configuration module of the framework layer receives the third instruction sent by the camera APP
  • the distributed recording system configuration module can determine that the recording mode selected by the photographer is the multi-camera cooperative recording mode.
  • the distributed recording system configuration module generates a distributed recording request according to the third instruction, and sends the distributed recording request to the WiFi interface.
  • the WiFi interface receives the distributed recording request, it forwards the distributed recording request to the WiFi subsystem at the hardware layer. And send the request to one or more transmitters around the subject through the WiFi subsystem.
  • the distributed recording system configuration module may generate a multi-machine collaborative recording request according to the third instruction. It can be understood that the function of the multi-machine collaborative recording request is the same as that of the above-mentioned distributed recording request.
  • the distributed recording system configuration module after the distributed recording system configuration module receives the third instruction, it can also generate a multi-machine cooperative recording instruction, and send the multi-machine cooperative recording instruction to the multi-machine recording algorithm module.
  • the multi-camera recording algorithm module Before processing the audio data, the multi-camera recording algorithm module can be configured according to the instructions sent by the distributed recording system configuration module. For example, after receiving the multi-camera collaborative recording instruction, the multi-camera recording algorithm module will Configure, select the output of the WiFi interface and the output of the local recording algorithm module as the input of the audio data stream at the same time.
  • the camera APP can directly send the local recording control command to the local recording algorithm module through the setparameter interface, so that the local algorithm module controls one or more MICs of the hardware layer to collect audio data the analog signal.
  • the specific mode of the operation of the local algorithm module is the same as that of the local recording mode, you can refer to the corresponding description of the local recording mode part, and will not be repeated here.
  • the WiFi subsystem of the receiver may also receive distributed recording response messages sent from one or more transmitters.
  • the receiver sends can receive a multi-machine cooperative recording response message.
  • the function and function of the multi-machine collaborative recording response message is the same as that of the distributed recording response message.
  • the following descriptions will take the distributed recording response message as an example, but it is conceivable that the distributed recording response message can be replaced by a multi-machine response message. Collaborative recording of response messages. For the scenario shown in Figure 12, it may be that only the distributed recording response message sent from one transmitter is received.
  • the distributed recording scene at this time is a distributed recording scene of one transmitter and one receiver.
  • distributed recording response messages may be received from multiple transmitters.
  • the distributed recording scene at this time is a distributed recording scene of multiple transmitters and one receiver. More specifically, the distributed recording response message sent by one or more transmitters may be received through the antenna. After the antenna receives the distributed recording response message, it can process the received distributed recording response message through the WiFi protocol stack, and then transmit the processed distributed recording response message to the WiFi in the framework layer through the hardware interface. interface. It can be understood that, for the specific manner of correspondingly processing the received audio data in the WiFi protocol stack, reference may be made to the current known technology, which will not be repeated here for the convenience of description.
  • the WiFi interface sends the distributed recording response message to the distributed recording system configuration module, so that the distributed recording system configuration module can determine which transmitters allow distributed recording according to the distributed recording response message.
  • the distributed recording system configuration module After the distributed recording system configuration module determines that one or more transmitters are allowed to perform distributed recording according to the distributed recording response message, the distributed recording system configuration module generates an instruction to start distributed recording, and sends the start distributed recording to the WiFi interface. Recording instructions. After receiving the instruction to start distributed recording, the WiFi interface forwards the instruction to start distributed recording to the WiFi subsystem of the hardware layer. And send the start distributed recording instruction to the transmitter that allows distributed recording through the WiFi subsystem. It can be understood that the process of sending an instruction to start distributed recording is the same as sending a distributed recording request, which is not repeated here for convenience of description.
  • the distributed recording system configuration module may further generate an instruction to start the multi-machine cooperative recording . It can be understood that the functions and functions of starting a multi-machine cooperative recording instruction are exactly the same as that of starting a distributed recording instruction.
  • the WiFi subsystem of the receiver After the WiFi subsystem of the receiver sends the instruction to start distributed recording (or the instruction to start multi-machine cooperative recording), it can also receive audio data sent from the transmitter that allows distributed recording.
  • the WiFi subsystem of the receiver After the WiFi subsystem of the receiver sends the instruction to start distributed recording (or the instruction to start multi-machine cooperative recording), it can also receive audio data sent from the transmitter that allows distributed recording.
  • the specific method please refer to the corresponding description in the distributed recording mode, which will not be repeated here.
  • the audio data stream input by the WiFi interface can be input into the time alignment sub-module at the same time as the audio data stream input by the local recording algorithm module. time alignment in.
  • the audio decoding sub-module in the multi-camera recording algorithm module can first decompress the compressed audio data to obtain the original audio data, and then decompress the compressed audio data.
  • the obtained original audio data is transmitted to the time alignment sub-module for time alignment.
  • time alignment method reference may be made to the description of the time alignment in the case of multiple transmitters in the distributed recording mode, which will not be repeated here.
  • the audio data after time alignment by the time alignment submodule can be sent to the recording processing submodule for recording processing.
  • the multi-camera recording algorithm module directly sends the audio data processed by the recording to the camera APP at the application layer. It can be understood that, for the manner in which the multi-camera recording algorithm module directly sends the audio data after recording processing to the camera APP at the application layer, reference may be made to the corresponding description in the distributed recording mode, which will not be repeated here.
  • the audio decoding sub-module is an optional sub-module. If the audio data sent by the transmitter has not been compressed, the audio data can be directly sent to the time alignment sub-module for time alignment.
  • WiFi is used as a possible method, and of course, any wireless transmission methods such as BT and ZigBee can be used for equivalent replacement, which is not limited in this application.
  • a terminal device may have a receiver architecture and a transmitter architecture at the same time, and for the terminal device, it is selected as one of a receiver or a transmitter according to specific usage conditions.
  • the solutions described in FIG. 5 to FIG. 13 can be understood as the receiver and the transmitter are fixed in each use process.
  • the identities of the receiver and the transmitter may be interchanged.
  • the receiver may display prompt information on the display screen for the exchange of identities.
  • the user can choose the receiver to exchange identities with a certain transmitter according to the prompt information.
  • the receiver can first determine which transmitters can also be used as receivers at the same time, and display the corresponding identifiers of the corresponding transmitters, so that the user can select one of the displayed one or more transmitter identifiers to communicate with the receiver.
  • Identity swap When the identities are exchanged, the original transmitter acts as a new receiver to realize all the functions of the receiver, and at the same time, the original receiver acts as a new transmitter to realize all the functions of the transmitter.
  • FIG. 14 is a flowchart of a distributed recording method provided by an embodiment of the present application.
  • the present application also provides a distributed recording method, which can implement the solutions described in FIG. 4 to FIG. 11 above.
  • the method is mainly used in distributed recording mode, and the method can include the following steps:
  • the receiver receives input information from the user, and determines, according to the input information, that the recording mode is the distributed recording mode.
  • the receiver sends a distributed recording request to one or more transmitters.
  • the transmitter receives the distributed recording request, and sends a distributed recording response message if the distributed recording is allowed.
  • the receiver receives a distributed recording response message sent by one or more transmitters, and sends an instruction to start distributed recording to the transmitters that allow distributed recording.
  • the transmitter that allows distributed recording receives an instruction to start distributed recording, and collects audio data according to the instruction.
  • one or more receivers receive audio data, and send the received audio data to the camera application.
  • the method described in FIG. 14 is mainly aimed at the distributed recording scenario of one transmitter and one receiver, and the distributed recording scenario of multiple transmitters and one receiver.
  • the specific implementation process please refer to FIGS. 4 to 11 .
  • the scheme described above will not be repeated here.
  • FIG. 15 is a flowchart of another distributed recording method provided by an embodiment of the present application.
  • the present application also provides another distributed recording method, which can implement the solutions described in FIGS. 7 to 9 , 12 , and 13 above.
  • the method is mainly used in the multi-machine collaborative recording mode, and the method may include the following steps:
  • the receiver receives input information from the user, and determines, according to the input information, that the recording mode is a multi-machine collaborative recording mode.
  • the receiver sends a distributed recording request (or a multi-machine collaborative recording request) to one or more transmitters.
  • the transmitter receives a distributed recording request (or a multi-machine cooperative recording request), and if the distributed recording (or multi-machine cooperative recording) is allowed, sends a distributed recording response message (or a multi-machine cooperative recording response message).
  • the receiver receives a distributed recording response message (or a multi-machine cooperative recording response message) sent by one or more transmitters, and sends an instruction to start distributed recording (or an instruction to start a multi-machine cooperative recording) to a system that allows distribution A transmitter for type recording (or multi-camera cooperative recording).
  • the transmitter that allows distributed recording receives an instruction to start distributed recording (or an instruction to initiate multi-machine collaborative recording), and collects audio data according to the instruction.
  • the transmitter that allows distributed recording sends the collected audio data to the receiver.
  • S1507 can also be executed.
  • the receiver starts the recording of the local machine, and collects the audio data of the local machine.
  • execution time of S1507 may be any time after S1501 and before S1508, which is not limited in this application.
  • the receiver receives the multiple audio data and the local audio data, and sends the received multiple audio data and the local audio data to the camera application.
  • the method described in Fig. 15 is mainly aimed at the multi-camera collaborative recording scenario of one transmitter and one receiver, and the multi-camera collaborative recording scenario of multiple transmitters and one receiver.
  • the specific implementation process please refer to Fig. 7 to Fig. 9.
  • the solutions described in FIG. 12 and FIG. 13 will not be repeated here.
  • FIG. 16 is a flowchart of yet another distributed recording method provided by an embodiment of the present application.
  • the present application also provides another distributed recording method, which can implement the solutions described in FIG. 4 to FIG. 11 above.
  • This method is mainly used in the local recording mode, and the method can include the following steps:
  • the receiver receives the input information of the user, and determines the recording mode as the local recording mode according to the input information.
  • the receiver starts the recording of the local machine, and collects the audio data of the local machine.
  • the receiver sends the local audio data to the camera application.
  • the method shown in FIG. 16 is mainly aimed at the local recording scene of one receiver.
  • FIG. 17 is a schematic diagram of a distributed recording device according to an embodiment of the present application.
  • a distributed recording device 1700 is provided, and the device 1700 may be the transmitter or the receiver in the solutions involved in the above-mentioned FIG. 4 to FIG. 13 .
  • the device 1700 may include a processor 1710, an external memory interface 1720, an internal memory 1721, a universal serial bus (USB) interface 1730, a charge management module 1740, a power management module 1741, a battery 1742, an antenna 1, an antenna 2 , mobile communication module 1750, wireless communication module 1760, audio module 1770, speaker 1770A, receiver 1770B, microphone 1770C, headphone jack 1770D, sensor module 1780, buttons 1790, motor 1791, indicator 1792, camera 1793, display screen 1794, and Subscriber identification module (subscriber identification module, SIM) card interface 1795 and so on.
  • SIM Subscriber identification module
  • the sensor module 1780 may include a pressure sensor 1780A, a gyroscope sensor 1780B, an air pressure sensor 1780C, a magnetic sensor 1780D, an acceleration sensor 1780E, a distance sensor 1780F, a proximity light sensor 1780G, a fingerprint sensor 1780H, a temperature sensor 1780J, a touch sensor 1780K, and an environment Light sensor 1780L, bone conduction sensor 1780M, etc.
  • the distributed recording device 1700 may include more or less components than shown, or combine some components, or separate some components, or arrange different components.
  • the illustrated components may be implemented in hardware, software, or a combination of software and hardware.
  • the processor 1710 may include one or more processing units, for example, the processor 1710 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor, and/or neural-network processing unit (neural-network processing unit, NPU), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
  • application processor application processor, AP
  • modem processor graphics processor
  • image signal processor image signal processor
  • ISP image signal processor
  • controller video codec
  • digital signal processor digital signal processor
  • baseband processor baseband processor
  • neural-network processing unit neural-network processing unit
  • the controller can generate an operation control signal according to the instruction operation code and timing signal, and complete the control of fetching and executing instructions.
  • a memory may also be provided in the processor 1710 for storing instructions and data.
  • the memory in processor 1710 is cache memory. This memory may hold instructions or data that have just been used or recycled by the processor 1710 . If the processor 1710 needs to use the instruction or data again, it can be called directly from the memory. Repeated accesses are avoided and the latency of the processor 1710 is reduced, thereby increasing the efficiency of the system.
  • the processor 1710 may include one or more interfaces.
  • the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transceiver (universal asynchronous transmitter) receiver/transmitter, UART) interface, mobile industry processor interface (MIPI), general-purpose input/output (GPIO) interface, subscriber identity module (SIM) interface, and / or universal serial bus (universal serial bus, USB) interface, etc.
  • I2C integrated circuit
  • I2S integrated circuit built-in audio
  • PCM pulse code modulation
  • PCM pulse code modulation
  • UART universal asynchronous transceiver
  • MIPI mobile industry processor interface
  • GPIO general-purpose input/output
  • SIM subscriber identity module
  • USB universal serial bus
  • the USB interface 1730 is an interface that conforms to the USB standard specification, which can be a Mini USB interface, a Micro USB interface, a USB Type C interface, and the like.
  • the USB interface 1730 can be used to connect a charger to charge the distributed recording device 1700, and can also be used to transmit data between the distributed recording device 1700 and peripheral devices. It can also be used to connect headphones to play or capture audio through the headphones.
  • the interface can also be used to connect other electronic devices, such as AR devices.
  • the interface connection relationship between the modules illustrated in the embodiment of the present invention is only a schematic illustration, and does not constitute a structural limitation of the distributed recording device 1700 .
  • the distributed recording device 1700 may also adopt different interface connection manners in the foregoing embodiments, or a combination of multiple interface connection manners.
  • the wireless communication function of the distributed recording device 1700 can be implemented by the antenna 1, the antenna 2, the mobile communication module 1750, the wireless communication module 1760, the modulation and demodulation processor, the baseband processor, and the like.
  • Antenna 1 and Antenna 2 are used to transmit and receive electromagnetic wave signals.
  • Each antenna in distributed recording device 1700 may be used to cover a single or multiple communication frequency bands. Different antennas can also be reused to improve antenna utilization.
  • the antenna 1 can be multiplexed as a diversity antenna of the wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
  • the mobile communication module 1750 can provide wireless communication solutions including 2G/3G/4G/5G etc. applied on the distributed recording device 1700 .
  • the mobile communication module 1750 may include at least one filter, switch, power amplifier, low noise amplifier (LNA), and the like.
  • the mobile communication module 1750 can receive electromagnetic waves from the antenna 1, filter and amplify the received electromagnetic waves, and transmit them to the modulation and demodulation processor for demodulation.
  • the mobile communication module 1750 can also amplify the signal modulated by the modulation and demodulation processor, and then convert it into electromagnetic waves for radiation through the antenna 1 .
  • at least part of the functional modules of the mobile communication module 1750 may be provided in the processor 1710 .
  • at least part of the functional modules of the mobile communication module 1750 may be provided in the same device as at least part of the modules of the processor 1710 .
  • the modem processor may include a modulator and a demodulator.
  • the modulator is used to modulate the low frequency baseband signal to be sent into a medium and high frequency signal.
  • the demodulator is used to demodulate the received electromagnetic wave signal into a low frequency baseband signal. Then the demodulator transmits the demodulated low-frequency baseband signal to the baseband processor for processing.
  • the low frequency baseband signal is processed by the baseband processor and passed to the application processor.
  • the application processor outputs sound signals through audio devices (not limited to speaker 1770A, receiver 1770B, etc.), or displays images or videos through display screen 1794 .
  • the modem processor may be a stand-alone device.
  • the modem processor may be independent of the processor 1710, and be provided in the same device as the mobile communication module 1750 or other functional modules.
  • the wireless communication module 1760 can provide applications on the electronic device 1700 including wireless local area networks (WLAN) (such as WiFi networks), Bluetooth, global navigation satellite system (GNSS), frequency modulation (frequency modulation, FM), near field communication technology (near field communication, NFC), infrared technology (infrared, IR) and other wireless communication solutions.
  • WLAN wireless local area networks
  • GNSS global navigation satellite system
  • FM frequency modulation
  • NFC near field communication technology
  • infrared technology infrared, IR
  • the wireless communication module 1760 may be one or more devices integrating at least one communication processing module.
  • the wireless communication module 1760 receives electromagnetic waves via the antenna 2 , frequency modulates and filters the electromagnetic wave signals, and sends the processed signals to the processor 1710 .
  • the wireless communication module 1760 can also receive the signal to be sent from the processor 1710 , perform frequency modulation on it, amplify it, and convert it into electromagnetic waves for radiation through the antenna 2 .
  • the antenna 1 of the distributed recording device 1700 is coupled with the mobile communication module 1750, and the antenna 2 is coupled with the wireless communication module 1760, so that the distributed recording device 1700 can communicate with the network and other devices through wireless communication technology.
  • Display screen 1794 is used to display images, videos, and the like.
  • Display screen 1794 includes a display panel.
  • the display panel can be a liquid crystal display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode or an active-matrix organic light-emitting diode (active-matrix organic light).
  • LED diode AMOLED
  • flexible light-emitting diode flexible light-emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light-emitting diode (quantum dot light emitting diodes, QLED) and so on.
  • the distributed recording device 1700 may include 1 or N display screens 1794, where N is a positive integer greater than 1.
  • the distributed recording device 1700 can realize the shooting function through the ISP, the camera 1793, the video codec, the GPU, the display screen 1794 and the application processor.
  • the camera 1793 is used to capture still images or video.
  • the object is projected through the lens to generate an optical image onto the photosensitive element.
  • the photosensitive element may be a charge coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor.
  • CMOS complementary metal-oxide-semiconductor
  • the photosensitive element converts the optical signal into an electrical signal, and then transmits the electrical signal to the ISP to convert it into a digital image signal.
  • the ISP outputs the digital image signal to the DSP for processing.
  • DSP converts digital image signals into standard RGB, YUV and other formats of image signals.
  • the distributed recording device 1700 may include 1 or M cameras 1793 , where M is a positive integer greater than 1.
  • the external memory interface 1720 can be used to connect an external memory card, such as a Micro SD card, to expand the storage capacity of the distributed recording device 1700.
  • the external memory card communicates with the processor 1710 through the external memory interface 1720 to realize the data storage function. For example to save files like music, video etc in external memory card.
  • Internal memory 1721 may be used to store computer executable program code, which includes instructions.
  • the internal memory 1721 may include a storage program area and a storage data area.
  • the storage program area can store an operating system, an application program required for at least one function (such as a sound playback function, an image playback function, etc.), and the like.
  • the storage data area may store data (such as audio data, etc.) created during the use of the distributed recording device 1700 and the like.
  • the internal memory 1721 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, universal flash storage (UFS), and the like.
  • the processor 1710 executes various functional applications and data processing of the distributed recording apparatus 1700 by executing instructions stored in the internal memory 1721 and/or instructions stored in the memory provided in the processor.
  • the distributed recording device 1700 can implement audio functions through an audio module 1770, a speaker 1770A, a receiver 1770B, a microphone 1770C, an earphone interface 1770D, and an application processor. Such as music playback, recording, etc.
  • the audio module 1770 is used for converting digital audio information into analog audio signal output, and also for converting analog audio input into digital audio signal. Audio module 1770 may also be used to encode and decode audio signals. such as compression and decompression. In some embodiments, the audio module 1770 may be provided in the processor 1710, or some functional modules of the audio module 1770 may be provided in the processor 1710, so that the processor can process the digital audio signal. In one example, the audio module 1770 may be, for example, a codec.
  • Microphone 1770C also called “microphone”, “microphone”, is used to convert sound signals into electrical signals.
  • the sound signal may be input to the microphone 1770C by sounding in close proximity to the microphone 1770C.
  • the distributed recording device 1700 may be provided with at least one microphone 1770C, for example, 3-4 microphones. In other embodiments, the distributed recording device 1700 may include two microphones 1770C, so that a noise reduction function can also be implemented while collecting sound signals. In other embodiments, the distributed recording device 1700 may also be provided with three, four or more microphones 1770C to collect sound signals, reduce noise, identify sound sources, and implement directional recording functions.
  • the headphone jack 1770D is used to connect wired headphones.
  • the earphone interface 1770D may be a USB interface 1730, or a 3.5mm open mobile terminal platform (OMTP) standard interface, a cellular telecommunications industry association of the USA (CTIA) standard interface.
  • OMTP open mobile terminal platform
  • CTIA cellular telecommunications industry association of the USA
  • the software system of the distributed recording device 1700 may adopt a layered architecture, an event-driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture.
  • the embodiment of the present application may be an Android system using a layered architecture.
  • the distributed recording device 1700 can perform any one of the functions of the transmitter or the receiver in the above-mentioned FIGS. 4 to 16 , for details, please refer to the descriptions in FIGS. 4 to 16 , and details are not repeated here.
  • the terminal equipment is used as the transmitter and receiver. Because the terminal equipment has a high degree of even memory and a large market stock, it is guaranteed that the user can clearly record and shoot anytime, anywhere without purchasing any other external equipment. The subject's voice as well as the narration and narration while getting a good recording.
  • non-transitory English: non-transitory
  • the storage medium is non-transitory ( English: non-transitory) media, such as random access memory, read only memory, flash memory, hard disk, solid state disk, magnetic tape (English: magnetic tape), floppy disk (English: floppy disk), optical disc (English: optical disc) and any combination thereof.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computational Linguistics (AREA)
  • Quality & Reliability (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • General Engineering & Computer Science (AREA)
  • Studio Devices (AREA)

Abstract

一种分布式录音方法和设备(1700),通过接收机发送分布式录音指令,以便控制一个或多个发射机进行录音,并将录音得到的音频数据发送至上层录音应用中。保障了用户无需购买任何外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说。

Description

一种录音方法和设备
本申请要求在2020年8月25日提交中国国家知识产权局、申请号为202010863700.9、发明名称为“一种录音方法和设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及音频领域,尤其涉及一种分布式的录音方法和设备。
背景技术
现如今,随着短视频以及直播行业的兴起,人们通过手机进行拍摄视频和进行直播已经越来越普遍,并且相关的软件也越来越多。当人们通过手机进行拍摄时,若拍摄对象与摄像设备距离较远时,拍摄对象的声音极易被环境中的噪声所掩盖,导致拍摄对象的声音很难被清晰地录制。例如在嘈杂的环境中进行拍摄时,旁白和解说的声音也会因为环境中的噪声,导致无法被清晰的录制。
比如,当用户作为主播进行直播时,一般会佩戴有线耳机进行收音,而有线耳机的连接线会造成用户活动极为不便,并限制活动范围。因此,通常情况下,用户需要额外购买外置录音配件,或者使用无线麦克风(microphone,MIC)设备。可以理解的是无线MIC设备可以包括发射机和接收机。显然额外购买的外置录音配件或者无线MIC设备的价格比较昂贵,并且设备尺寸也较大。在使用过程中需要和手机进行有线连接或固定,一些设备甚至需要手动设置MIC的朝向和相应的参数,并且在使用完毕后还得进行拆卸。显然,上述设备仅适合专业人员进行使用,并不适用于普通用户。
因此,在不增加任何其它设备的前提下,可以简单、方便以及随时随地对拍摄对象的声音进行清晰的录制,成为了急需解决的问题,对用户的录音体验的提升具有重要意义。
发明内容
本申请实施例提供了一种录音方法,在无线局域网络中,采用分布式录音的方式,接收机通过发送分布式录音指令至一个或多个发射机,以便一个或多个发射机根据分布式录音指令进行分布式录音,并将录音得到的音频数据反馈至接收机。接收机将接收到的一个或多个音频数据进行时间对齐、降噪和/或解压缩,并发送至上层录音应用中。可以保障用户无需购买任何外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说。
第一方面,提供了一种分布式录音系统,系统包括:至少两个终端设备,至少两个终端设备中包含第一设备和至少一个第二设备。第一设备和第二设备可以是手机、平板电脑、可穿戴设备等等。第一设备用于接收用户的输入信息,并根据输入信息确 定录音模式;第一设备还用于,当录音模式为分布式录音模式时,发送启动分布式录音指令到至少一个第二设备;第二设备用于接收启动分布式录音指令,并根据启动分布式录音指令周期性采集第一音频数据,其中,第一音频数据为第二设备采集到的本地音频数据;第二设备还用于,将周期性采集到的第一音频数据,并对第一音频数据进行人声增强和/或降噪处理,再以相同周期发送至第一设备;第一设备还用于,接收至少一个第一音频数据,以便摄像应用调用至少一个第一音频数据。本申请通过接收机控制一个或多个发射机进行录音,并将录音得到的音频数据发送至上层录音应用中。保障了用户无需购买任何外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说。
在一个可能的实施方式中,第二设备还用于:根据启动分布式录音指令确定录音参数;根据录音参数采集第一音频数据。
在一个可能的实施方式中,在第二设备将周期性采集到的第一音频数据以相同周期发送至第一设备之前,第二设备还用于:将周期性采集到的第一音频数据进行压缩处理,得到压缩后的第一音频数据。本申请对采集到的音频数据进行压缩,可以有效减少传输过程中的数据大小,以及保障了传输过程中数据的安全性。
在一个可能的实施方式中,第一设备还用于:当第一音频数据为压缩过的数据,对压缩的第一音频数据进行音频解压缩,得到解压缩后的第一音频数据。本申请还可以对接收到的压缩的音频数据进行解压缩,在传输数据减小数据量的同时,保障接收后可以无损还原。
在一个可能的实施方式中,分布式录音模式包括:多机协同录音模式;当录音模式为多机协同录音模式时,在摄像应用调用至少一个第一音频数据之前,第一设备还用于:采集第二音频数据,摄像应用调用至少一个第一音频数据和第二音频数据,其中,第二音频数据为第一设备采集到的本地音频数据。本申请还可以结合接收机采集的本地音频数据,保障了拍摄视频的过程中可以清晰的录制到拍摄对象的声音以及旁白和解说。
在一个可能的实施方式中,在摄像应用调用至少一个第一音频数据之前,第一设备还用于:当第一音频数据为多个时,对多个第一音频数据进行时间对齐操作;或当录音模式为多机协同录音模式时,对至少一个第一音频数据和第二音频数据进行时间对齐操作。本申请在接收到多个音频数据后,对多个音频数据进行时间对齐,避免了不同音频之间由于传输导致的时间误差。
在一个可能的实施方式中,在摄像应用调用至少一个第二音频数据之前,第一设备还用于:对第二音频数据进行降噪和/或人声增强处理。本申请还可以对音频进行降噪、人声增强等预处理,使得上传至摄像应用的音频更加清晰。
在一个可能的实施方式中,分布式录音模式包括:第一子分布式录音模式和第二子分布式录音模式;当录音模式为第一子分布式录音模式时,第一设备还用于发送分布式录音指令到一个第二设备,以及接收第二设备发送的第一音频数据;或者,当录音模式为第二子分布式录音模式时,第一设备还用于发送分布式录音指令到多个第二设备,以及接收多个第二设备发送的第一音频数据。
在一个可能的实施方式中,第一设备在发送启动分布式录音指令到至少一个第二 设备之前,系统还包括:第一设备还用于,发送分布式录音请求到至少一个第二设备;第二设备还用于,接收第一设备发送的分布式录音请求;第二设备还用于,在第二设备的显示屏上显示提示信息,提示信息用于提示是否允许该第二设备进行分布式录音;第二设备还用于,向第一设备发送分布式录音响应消息,分布式录音响应消息用于表示是否允许该第二设备进行分布式录音;第一设备还用于,接收至少一个第二设备发送的分布式录音响应消息;第一设备还用于,向允许进行分布式录音的至少一个第二设备发送启动分布式录音指令。
在一个可能的实施方式中,录音模式还包括本地录音模式;第一设备还用于,根据用户输入的输入信息,在本地录音模式、第一子分布式录音模式、第二子分布式录音模式和多机协同录音模式中进行切换。
在一个可能的实施方式中,第一设备还用于:在第一设备的显示屏上显示分布式录音图标。
在一个可能的实施方式中,第二设备还用于:在第二设备的显示屏上显示分布式录音提示图标。
第二方面,提供了一种录音终端设备,录音终端设备为第一设备或第二设备;录音终端设备包括:显示器、处理器、存储器、发送器和接收器;当录音终端设备为第一设备时,接收器,用于接收用户的输入信息,并根据输入信息确定录音模式;处理器,用于当录音模式为分布式录音模式时,控制发送器发送启动分布式录音指令到至少一个第二设备;当录音终端设备为第二设备时,接收器,用于接收启动分布式录音指令,并根据启动分布式录音指令周期性采集第一音频数据,其中,第一音频数据为第二设备采集到的本地音频数据;发送器,用于将周期性采集到的第一音频数据,并对第一音频数据进行人声增强和/或降噪处理,再以相同周期发送至第一设备;当录音终端设备为第一设备时,接收器还用于,接收至少一个第一音频数据;处理器还用于,控制摄像应用调用至少一个第一音频数据。本申请通过接收机控制一个或多个发射机进行录音,并将录音得到的音频数据发送至上层录音应用中。保障了用户无需购买任何外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说。
在一个可能的实施方式中,当录音终端设备为第二设备时,处理器用于:根据启动分布式录音指令确定录音参数;根据录音参数控制麦克风采集第一音频数据。
在一个可能的实施方式中,当录音终端设备为第二设备时,处理器还用于:将周期性采集到的第一音频数据进行压缩处理,得到压缩后的第一音频数据。本申请对采集到的音频数据进行压缩,可以有效减少传输过程中的数据大小,以及保障了传输过程中数据的安全性。
在一个可能的实施方式中,当录音终端设备为第二设备时,处理器还用于:当第一音频数据为压缩过的数据,对压缩的第一音频数据进行音频解压缩,得到解压缩后的第一音频数据。本申请还可以对接收到的压缩的音频数据进行解压缩,在传输数据减小数据量的同时,保障接收后可以无损还原。
在一个可能的实施方式中,分布式录音模式包括:多机协同录音模式;当录音终端设备为第一设备时,处理器还用于:当录音模式为多机协同录音模式时,采集第二音频数据,摄像应用调用至少一个第一音频数据和第二音频数据,其中,第二音频数 据为第一设备采集到的本地音频数据。本申请还可以结合接收机采集的本地音频数据,保障了拍摄视频的过程中可以清晰的录制到拍摄对象的声音以及旁白和解说。
在一个可能的实施方式中,当录音终端设备为第一设备时,处理器还用于:当第一音频数据为多个时,对多个第一音频数据进行时间对齐操作;或当录音模式为多机协同录音模式时,对至少一个第一音频数据和第二音频数据进行时间对齐操作。本申请在接收到多个音频数据后,对多个音频数据进行时间对齐,避免了不同音频之间由于传输导致的时间误差。
在一个可能的实施方式中,当录音终端设备为第一设备时,处理器还用于:对第二音频数据进行降噪和/或人声增强处理。本申请还可以对音频进行降噪、人声增强等预处理,使得上传至摄像应用的音频更加清晰。
在一个可能的实施方式中,分布式录音模式包括:第一子分布式录音模式和第二子分布式录音模式;当录音终端设备为第一设备时,发送器还用于,当录音模式为第一子分布式录音模式时,发送分布式录音指令到一个第二设备;或者,发送器还用于,当录音模式为第二子分布式录音模式时,发送分布式录音指令到多个第二设备;接收器还用于,当录音模式为第一子分布式录音模式时,接收第二设备发送的第一音频数据;或者,当录音模式为第二子分布式录音模式时,接收多个第二设备发送的第一音频数据。
在一个可能的实施方式中,当录音终端设备为第一设备时,发送器还用于,发送分布式录音请求到至少一个第二设备;当录音终端设备为第二设备时,接收器还用于,接收第一设备发送的分布式录音请求;显示器还用于,在第二设备的显示屏上显示提示信息,提示信息用于提示是否允许该第二设备进行分布式录音;
发送器还用于,向第一设备发送分布式录音响应消息,分布式录音响应消息用于表示是否允许该第二设备进行分布式录音;
当录音终端设备为第一设备时,接收器还用于,接收至少一个第二设备发送的分布式录音响应消息;
发送器还用于,向允许进行分布式录音的至少一个第二设备发送启动分布式录音指令。
在一个可能的实施方式中,录音模式还包括本地录音模式;当录音终端设备为第一设备时,处理器还用于,根据用户输入的输入信息,在本地录音模式、第一子分布式录音模式、第二子分布式录音模式和多机协同录音模式中进行切换。
在一个可能的实施方式中,当录音终端设备为第一设备时,显示器用于:在第一设备的显示屏上显示分布式录音图标。
在一个可能的实施方式中,当录音终端设备为第二设备时,显示器用于:在第二设备的显示屏上显示分布式录音提示图标。
第三方面,提供了一种计算机可读存储介质,计算机可读存储介质中存储有指令,当指令在终端上运行时,使得终端执行第一方面任意一项的方法。
第四方面,提供了一种包含指令的计算机程序设备,当其在终端上运行时,使得终端执行第一方面中的任一项的方法。
本申请公开了一种分布式录音方法和设备,通过接收机发送分布式录音指令,以 便控制一个或多个发射机进行录音,并将录音得到的音频数据发送至上层录音应用中。保障了用户无需购买任何外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说。
附图说明
图1为本申请实施例提供的一种分布式录音场景示意图;
图2a为一种专业录音设备示意图;
图2b为图2a所示录音设备使用示意图;
图3为另一种专业录音设备示意图;
图4为本申请实施例提供的一种一对一分布式录音场景示意图;
图5为本申请实施例提供的一种接收机架构示意图;
图6为本申请实施例提供的一种终端设备摄像界面示意图;
图7为本申请实施例提供的一种发射机架构示意图;
图8为本申请实施例提供的一种显示提示信息界面示意图;
图9为本申请实施例提供的一种分布式录音界面示意图;
图10为本申请实施例提供的一种一对多分布式录音场景示意图;
图11为本申请实施例提供的另一种接收机架构示意图;
图12为本申请实施例提供的一种多机协同分布式录音场景示意图;
图13为本申请实施例提供的又一种接收机架构示意图;
图14为本申请实施例提供的一种分布式录音方法流程图;
图15为本申请实施例提供的另一种分布式录音方法流程图;
图16为本申请实施例提供的又一种分布式录音方法流程图;
图17为本申请实施例提供的一种分布式录音设备示意图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行描述。
本申请主要应用在使用终端设备进行拍摄视频的场景。例如图1示出的,在该场景下,拍摄者正在对被摄对象进行视频拍摄,此时被摄对象与拍摄者的距离相对较远,因此,拍摄者使用的摄像器材无法清晰的将被摄对象的声音进行采集并录制。同时,若拍摄视频时还存在旁白和解说,则由于位置的差异,拍摄者使用的摄像器材同样无法清晰的将旁白和解说的声音进行采集并录制。
在一些方案中,通常采用另购一些外置的专业录音配件,例如图2a示出的。此类专业录音配件通过有线连接的方式与终端设备相连接,并向终端设备发送音频数流,同时此类专业录音配件可以被终端设备进行控制。其连接方式例如图2b示出的,可以看出图2a所示的专业录音设备通过有线连接的方式,与下方的终端设备相连接。此时,终端设备上可以运行与该专业录音设备配套的应用(application,APP)进行录音。当然,还可以在终端设备上显示波形、频谱等信息。
此类专业录音配件所采用的MIC器件的尺寸与性能指标有着紧密的关系。其关注的性能指标例如灵敏度、带宽、频响平坦度、声学过载点等等信息。该类专业录音配 件通常使用尺寸较大的MIC器件或者采用指向性MIC器件,并具备较强的信号处理能力。此类专业录音配件通过将MIC采集到的音频数据先在配件的内部进行处理,然后通过通用串行总线(universal serial bus,USB)接口或其它等效的接口将音频数据传输至相连接的终端设备上。对于终端设备而言,还可以采用信号处理算法对处理过的音频数据进行二次处理,从而实现多样化的录音功能。
然而,对于上述方案,由于专业录音设备尺寸较大,导致便携性较差。在使用时需要将设备与终端设备相连接,并通过终端设备对MIC进行手动调节方向和增益等参数,使用完毕后还需取下,使得操作过程过于繁琐复杂。同时,对于拍摄距离较远的情况,或是环境较为嘈杂时,录音效果也不是很好,导致此类设备仅适用于专业用户,并不适合普通用户。
在另一些方案中,采用了无线MIC的方案,该方案包括有发射机和接收机。其中,被摄对象可以佩戴发射机,例如具备无线功能的领夹MIC,此时可以将该无线领夹MIC固定在被摄对象的衣领上,以便拾取人声,例如图3中左半部分示出的。而接收机则可以与终端设备进行有线连接。该方案通过无线领夹MIC拾取音频数据,并通过无线的方式传输至接收机,接收机将接收到的音频数据通过USB接口或其它等效的接口传输至相连接的终端设备上。通常一个接收机可以连接1至2台发射机。
但是,该方案中,无线领夹MIC和接收机功能单一,体积较大,用户需要单独购买。性能好的通常在500元以上,而更专业的型号则高达几千元。同时,使用时结合搜集需要与终端设备相连接,并在使用完毕后对接收机和发射机进行拆卸。这导致其仅适用于专业用户,并不适合普通用户随时随地适用。
因此,本申请提出了一种可以将终端设备变成无线录音MIC的方法,包括将部分终端设备变成发射机,以及将某一个终端设备变为接收机。当用户通过终端设备进行拍摄时,无需单独购买其它设备,而是采用另一个终端设备即可进行外围拾音。通过接收机发送分布式录音指令,控制一个或多个发射机进行录音,并将录音得到的音频数据发送至上层录音应用中。保障了用户无需购买任何外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说,从而达到专业化的录音效果。
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行详细描述。
图4为本申请实施例提供的一种一对一分布式录音场景示意图。
可以看出该场景下,采用一个接收机匹配一个发射机。拍摄者使用接收机进行视频拍摄,此时发射机位于被摄对象的周围,以便清晰的将被摄对象的声音进行拾音。其中,接收机可以是第一设备,发射机可以是第二设备。本申请中的“第一”、“第二”并不限制其先后顺序,而是仅仅用于区分不同的终端设备。
可以理解的是,本申请中的发射机与接收机需要预先连接在同一个无线网络环境中,例如通过无线保真(wireless fidelity,WiFi)、蓝牙(bluetooth,BT)、紫峰(ZigBee)等任意无线传输方式接入同一无线网络环境。换句话说,发射机与接收机需要处于同一个无线局域网中。
在该场景下,本申请提供了图5所示的一种接收机架构示意图。
可以理解的是,该接收机架构可以应用在终端设备上。本申请所涉及的终端设备可以但不限于手机、电视、音响、可穿戴设备、平板电脑、个人数字助理(personal  digitalassistant,PDA)、膝上型计算机(laptop)、移动电脑、增强现实(augmented reality,AR)设备、虚拟现实(virtual reality,VR)设备、人工智能(artificial intelligence,AI)设备等任意终端设备或便携式终端设备。
如图5示出的,接收机架构包括有APP层、架构(framework)层和硬件层。其中APP层包括有摄像APP,当然还可包括其他任何可能的APP,本申请在此不作限定。
在一个例子中,以用户使用的角度对本方案进行描述。例如,用户首先通过接收机选择录音模式,如分布式录音模式、多机协同录音模式或本地录音模式。当用户选择分布式录音模式或多机协同录音模式后,接收机可以通过向自身设备所处的无线网络中发送分布式录音请求,例如可以采用广播的形式进行发送。然后当处于同一无线网络中的发射机接收到分布式录音请求,可以根据发射机的用户进行确认是否允许进行分布式录音或多机协同录音。当发射机的用户确定允许该发射机进行分布式录音或多机协同录音后,该发射机可以向所处的无线网络中发送响应消息。以便接收机接收到该响应消息后,确认该发射机允许进行录音。上述过程可以理解为发射机与接收机相互确认的过程。此时,接收机可以确认一个或多个允许录音的发射机。接收机可以发送启动分布式录音或多机协同录音的指令,以便确认的一个或多个发射机进行录音,并周期性的将录制的音频数据发送至接收机。当周期较短时,可以看成发射机实时发送音频数据,接收机也实时接收音频数据。当接收机接收到发射机发送的音频数据后,可以被接收机的上层摄像APP进行调用。当然接收机还可以将接收到的音频数据进行实时上传,如上传至某服务器,从而实现在线直播等功能。
例如在图4所示的场景下,假设拍摄者准备使用终端设备对被摄对象进行摄像或直播。此时拍摄者所使用的终端设备可称为接收机。拍摄者首先可以打开接收机上的摄像APP,然后根据接收机显示的界面,点击该界面上显示的分布式录音图标,从而选择相应的录音模式。其中录音模式可以包括分布式录音和本机录音。当然也可以是某些设置菜单的选项中显示分布式录音图标,或是在其它一些可能的用户界面(user interface,UI)上显示分布式录音图标。录音模式可以包括分布式录音模式、多机协同录音模式和本机录音模式。在一个例子中,若拍摄者不点击分布式录音图标进行录音模式选择时,摄像APP可以选择本地录音模式作为默认的录音模式。
较为具体的,图6示出了一种终端设备摄像界面示意图。可以看出,该界面为用户使用接收机上的摄像APP进行摄像或直播的显示界面。当拍摄者进行摄像或直播的过程中,可以通过界面左下方的分布式录音图标601选择录音模式。当然可以理解的是,通过分布式录音图标601选择录音模式的动作也可以发生在摄像或直播之前。也就是说,拍摄者可以预先通过分布式录音图标601选择好相应的录音模式,然后再进行摄像或直播。拍摄者可以通过点击分布式录音图标601,以实现在不同录音模式下的切换。例如,在图4的场景下,拍摄者可以通过点击分布式录音图标601选择分布式录音模式。由于此场景下仅包括1个发射机和1个接收机,因此在该场景下,分布式录音模式可以为第一子分布式录音模式,用于表示一个发射机与一个接收机通过无线方式进行连接,并进行分布式录音。
当然对于图6所示的摄像界面,还可以包括用于开启/停止录像的按键602、用于暂停录像的按键603、用于拍摄录像中某一帧画面的按键604、用于调节画面大小的滑 动条605以及用于控制闪光灯开启的按键606。当然该界面还可以包括视频录制时间、以及位于界面上方的任务栏。其中该任务栏中可以包括运营商、WiFi、电量以及时间等信息。当然该任务栏中还可以包括其它可能的图标或信息,本申请再次不做限定。同时,对于该摄像界面中还可以包括除图中示出的图标以外的其它任意具有某些特定功能的图标。可以理解的是,图6仅仅示意了摄像界面的一种可能形式,其中的分布式录音图标以及其它图标的样式可以根据实际情况进行任意设计,本申请在此不作限定。
继续回到图5,当用户通过摄像APP选择了分布式录音模式进行录像录音时,摄像APP向framework层中的分布式录音系统配置模块发送模式选择指令,其中,该模式选择指令可以是用于告知分布式录音系统配置模块用户选择了分布式录音模式的第一指令。
当摄像APP发送的为第一指令,则分布式录音系统配置模块在接收到第一指令后,可以确定拍摄者选择的录音模式为分布式录音模式。之后,分布式录音系统配置模块根据第一指令生成分布式录音请求,并向WiFi接口发送该分布式录音请求。WiFi接口在接收到分布式录音请求后,会将该请求转发至硬件层的WiFi子系统。并通过WiFi子系统将该请求发送至被摄对象周围的一个或多个发射机。可以理解的是WiFi接口为framework层中抽象出来的软件接口。其中,WiFi子系统可以包括硬件接口、WiFi协议栈和天线。对于本申请涉及的WiFi协议栈可以是硬件实体化的WiFi协议栈,用于对数据进行相应处理,从而使得处理后的数据满足WiFi协议格式。较为具体的,WiFi接口将分布式录音请求转发至硬件层的硬件接口。硬件接口在接收到分布式录音请求后,将分布式录音请求转发至WiFi协议栈,并在WiFi协议栈内对分布式录音请求进行相应的处理,并通过天线发送至发射机。可以理解的是,在WiFi协议栈内对分布式录音请求进行相应的处理的具体方式可以参考当前的已知技术,为方便描述在此不再赘述。
当分布式录音系统配置模块接收到第一指令后,还可以生成分布式录音指令,并向多机录音算法模块发送该分布式录音指令。其中,多机录音算法模块用于接收来自不同终端设备的音频数据,并将接收到的一个或多个音频数据进行音频解压缩、录音处理、时间对齐等操作,并将处理后的音频数据发送至应用层的摄像APP。在对音频数据处理之前,多机录音算法模块可以根据分布式录音系统配置模块发送的指令进行配置,例如,多机录音算法模块在接收到分布式录音指令后,会根据分布式录音指令进行配置,选择WiFi接口的输出作为音频数据流的输入,而对于本机录音算法模块的输出则不作为输入。
接收机的WiFi子系统在发送分布式录音请求之后,还可以接收来自一个或多个发射机发送的分布式录音响应消息。当然,对于图4所示的场景,可以是仅接收到来自一个发射机发送的分布式录音响应消息。此时的分布式录音场景即为1个发射机与1个接收机的分布式录音场景。较为具体的,可以通过天线接收到与某个发射机发送的分布式录音响应消息。天线在接收到分布式录音响应消息后,可以经过WiFi协议栈对接收到的分布式录音响应消息进行相应的处理,再通过硬件接口将处理后的分布式录音响应消息传输至framework层中的WiFi接口。可以理解的是,在WiFi协议栈内对 接收到音频数据进行相应的处理的具体方式可以参考当前的已知技术,为方便描述在此不再赘述。WiFi接口将分布式录音响应消息发送至分布式录音系统配置模块,以便分布式录音系统配置模块可以根据分布式录音响应消息确定哪些发射机允许进行分布式录音。
当分布式录音系统配置模块根据分布式录音响应消息确定了某个发射机允许进行分布式录音之后,分布式录音系统配置模块生成启动分布式录音指令,并向WiFi接口发送该启动分布式录音指令。在一个例子中,上述的生成启动分布式录音指令,可以是接收机通过显示屏发出提示信息,以便用户可以根据该提示信息确定是否立即启动分布式录音。若接收机接收到用户选择立即启动分布式录音的操作指令后,可以控制分布式录音系统配置模块生成启动分布式录音指令。当然在另一个例子中,接收机的分布式录音系统配置模块可以在根据分布式录音响应消息确定了某个发射机允许进行分布式录音之后,主动生成启动分布式录音指令。再一个例子中,还可以是接收机的分布式录音系统配置模块在根据分布式录音响应消息确定了某个发射机允许进行分布式录音之后,接收到用户主动触发的立即启动分布式录音的操作指令。可以理解的是,在该例子中,接收机可能并未提示用户是否立即启动分布式录音,而是用户通过触摸屏上的实体或虚拟按钮,主动提供启动分布式录音的操作指令。WiFi接口在接收到启动分布式录音指令后,会将该启动分布式录音指令转发至硬件层的WiFi子系统。并通过WiFi子系统将该启动分布式录音指令发送至允许进行分布式录音的发射机。可以理解的是,发送启动分布式录音指令与发送分布式录音请求的过程相同,为方便描述在此不再赘述。
接收机的WiFi子系统在发送启动分布式录音指令之后,还可以接收来自允许进行分布式录音的发射机发送的音频数据。较为具体的,可以通过天线接收到该发射机发送的音频数据。然后天线在接收到音频数据后,可以经过WiFi协议栈对接收到音频数据进行相应的处理,再通过硬件接口将处理后的音频数据传输至framework层中的WiFi接口。可以理解的是,在WiFi协议栈内对接收到音频数据进行相应的处理的具体方式可以参考当前的已知技术,为方便描述在此不再赘述。
framework层中的WiFi接口在接收到处理后的音频数据后,将该音频数据直接转发至多机录音算法模块中。以便多机录音算法模块中的一个或多个子模块对接收到的音频数据进行处理。在一个例子中,若发射机发送的音频数据是经过压缩处理过的音频数据,则多机录音算法模块中的音频解码子模块可以先将压缩过的音频数据进行解压缩处理,得到原始的音频数据,然后将得到的原始的音频数据经录音处理子模块进行录音处理。可以理解的是对音频数据的压缩过程也可以看做是对音频数据的编码过程,同理,则对压缩过的音频数据进行解压缩过程也可以看做是解码过程。其中,录音处理可以包括对音频数据的降噪、人声增强以及其它可能的操作。经过多机录音算法模块将录音处理后的音频数据直接发送至应用层的摄像APP。在一个例子中,多机录音算法模块在将音频数据进行录音处理后,可以通过录音记录来源(audio record.source)接口将录音处理后的音频数据发送至应用层的摄像APP,以便摄像APP使用该音频数据。例如将音频数据与摄像画面组合成为视频文件,或是将音频数据与摄像画面组合成为视频文件后通过现有方式转发给其他终端设备,从而实现视频直播。当 然由于接收机与发射机通过WiFi连接在同一无线局域网中,接收机此时可以通过其它网络连接方式将合成的视频文件传输至某服务器或者直接传输至其它终端设备。例如可以通过蜂窝网络、蓝牙或ZigBee等等。当然若采用WiFi直连(direct)技术,假设接收机与发射机使用了2.4GHz频段进行WiFi互联,则还可以通过5GHz频段实现与广域网进行数据交互。可以理解的是,若接收机与发射机通过蓝牙或ZigBee等其它无线方式进行连接,则接收机可以直接通过2.4GHz频段或5GHz频段的WiFi与广域网相连接。
可以理解是的,音频解码子模块作为可选的子模块,若发射机发送的音频数据并未经过压缩处理,则可以直接将音频数据经录音处理子模块进行录音处理。
在另一个例子中,分布式录音系统配置模块接收摄像APP发送的模式选择指令,还可以是用于告知分布式录音系统配置模块用户选择了本机录音模式的第二指令。
当摄像APP发送的为第二指令,则分布式录音系统配置模块在接收到第二指令后,可以确定拍摄者选择的录音模式为本机录音模式。之后,分布式录音系统配置模块还可以根据第二指令生成本地录音指令,并向多机录音算法模块发送该本地录音指令。例如,多机录音算法模块在接收到本地录音指令后,会根据本地录音指令进行配置,选择本机录音算法模块的输出作为音频数据流的输入。
若拍摄者选择了本机录音模式,则摄像APP可以通过设置参数(setparameter)接口直接向本机录音算法模块发送本机录音控制指令,以便本机算法模块控制硬件层的一个或多个MIC采集到音频数据的模拟信号。其中,本机录音控制指令可以为一个字节的长度。当本机录音算法模块接收到本机录音控制指令之后,可以控制该接收机上的全部MIC进行采集音频数据。通常情况下,MIC的数量为2-4个,例如在一个例子中为3个MIC。当然,本机录音算法模块还可以根据本机录音控制指令控制该接收机上MIC的调节拾音范围、噪声强度等等。例如,多MIC指向性降噪,拾音范围可以是大于3米等等。MIC将采集到的模拟信号通过模数转换器转换为数字信号,再将音频数据的数字信号传输至framework层中的本机录音算法模块进行相应的处理。其中,本机录音算法模块对音频数据的数字信号进行处理,例如可以是预加重、滤波、混响等等中的一种或几种组合的任意可能处理方式。本机录音算法模块可以通过与多机录音算法模块之间建立的音频通路将音频数据传输至多机录音算法模块,此时多机录音算法模块不具有任何功能,仅用于将音频数据转发至APP层的摄像APP。在一个例子中,模数转换器可以采用编译码器(coder-decoder,Codec)将采集到的模拟信号转换为数字信号。可以理解的是,本申请中若无特殊说明,则音频数据指代的是音频数据的数字信号。
可以理解的是,若用户并未选择某个录音模式时,多机录音算法模块将选择本机录音算法模块的输出,作为默认的音频数据流的输入。
在图4所示的场景下与接收机相对应的,本申请还提供了图7所示的一种发射机架构示意图。
可以理解的是,该发射机架构可以应用在终端设备上。终端设备可以但不限于手机、可穿戴设备、平板电脑、PDA、膝上型计算机、移动电脑、AR设备、VR设备、AI设备等任意终端设备或便携式终端设备。
其中,发射机架构包括有APP层、framework层和硬件层。首先,位于硬件层的WiFi子系统可以接收由接收机发送的分布式录音请求,较为具体的,例如,发射机可以通过天线接收到接收机发送的分布式录音请求。并将该分布式录音请求发送至WiFi协议栈。WiFi协议栈再将分布式录音请求通过硬件接口转发至framework层中的WiFi接口。WiFi接口可以将接收到分布式录音请求转发至分布式录音发射机配置模块。分布式录音发射机配置模块在接收到分布式录音请求之后,可以向APP层发送提示信息,该提示信息显示在显示屏上,用于提示用户是否允许该终端设备作为发射机进行分布式录音。例如图8所示,可以看到显示屏界面上出现用于提示的弹窗,该弹窗上可以包括文字、图形、图案等信息,用于提示用户是否进行分布式录音。同时还可以具有选择按钮,例如“是”、“否”,以便用户进行选择。可以理解的是,用户可以是拍摄者、被摄对象或是其它使用该终端设备的用户。
若用户选择允许该终端设备作为发射机进行分布式录音,则APP层将根据用户在显示屏上的操作,将第一信息发送至framework层中的分布式录音发射机配置模块。分布式录音发射机配置模块可以根据第一信息确定将该终端设备作为发射机,并生成分布式录音响应消息。然后,分布式录音发射机配置模块将分布式录音响应消息通过WiFi接口发送至硬件层的WiFi子系统,并经过WiFi子系统发送至接收机,以便接收机根据该分布式录音响应消息向该发射机发送启动分布式录音指令。当然,若用户选择不允许该终端设备作为发射机进行分布式录音,则APP层将根据用户在显示屏上的操作,将第二信息发送至framework层中的分布式录音发射机配置模块。分布式录音发射机配置模块可以根据第二信息确定该终端设备不作为发射机,并生成分布式录音响应消息。
发射机的WiFi子系统还可以接收接收机发送的启动分布式录音指令。较为具体的,例如,发射机可以通过天线接收到接收机发送的启动分布式录音指令。并将该启动分布式录音指令发送至WiFi协议栈。WiFi协议栈再将启动分布式录音指令通过硬件接口转发至framework层中的WiFi接口。WiFi接口可以将接收到启动分布式录音指令转发至分布式录音发射机配置模块。
其中,分布式录音发射机配置模块在接收到启动分布式录音指令后,可以通过setparameter接口向本机录音算法模块发送启动分布式录音指令,本机录音算法模块在接收到启动分布式录音指令后,可以根据该启动分布式录音指令调节算法参数以适应分布式录音的需求。例如,启动分布式录音指令可以为1字节长度,当本机录音算法模块接收到启动分布式录音指令之后,可以控制该接收机上的部分MIC进行采集音频数据的模拟信号。例如可以控制1个或2个MIC,其中,若控制一个MIC则可以是位于终端设备顶部的MIC,若控制两个MIC则可以是位于终端设备顶部的MIC以及底部的MIC。当然,本机录音算法模块还可以根据启动分布式录音指令控制该接收机上MIC的调节拾音范围、噪声强度等等。例如,单MIC降噪或是双MIC降噪,拾音范围可以是1米内等等。当MIC采集完音频数据的模拟信号之后,可以将音频数据的模拟信号发送至模数转换器,并将采集到的模拟信号转换为数字信号,然后将本机音频数据的数字信号发送至本机录音算法模块。此时,本机录音算法模块还可以对音频数据的数字信号进行处理,例如可以是预加重、滤波、混响等等中的一种或几种组合的 任意可能处理方式。并在处理后,将音频数据的数字信号发送至分布式录音发射机配置模块。当然,在一些例子中,本机录音算法模块还可以将本机采集到的音频数据进行压缩处理。然后将压缩后的音频数据发送至分布式录音发射机配置模块。
在一个例子中,发射机可以周期性的采集音频数据,当该周期较短时,例如几毫秒、几微秒甚至几纳秒时,则可以近似看做实时采集音频信号。可以理解的是,发射机可以以相同的周期发送周期性采集到的音频信号,即在一个周期内,发射机采集音频信号并发送至接收机。对于接收机而言,则可以按照相同的周期接收音频信号。
对于接收机而言,若发射机通过有线或无线的方式连接有耳机。则此时可以使用耳机的MIC进行拾音。对于拍摄对象而言就可以将发射机放入口袋或衣服兜等其他位置。此时发射机可以作为“无线领夹麦克风”。
对于发射机的分布式录音发射机配置模块在接收到启动分布式录音指令后,可以将分布式录音提示图标相关的数据发送至APP层,并在APP层显示分布式录音提示图标。例如图9示出的终端设备显示屏上方的任务栏中,还可以具有分布式录音提示图标801。该分布式录音提示图标801用于表示此终端设备作为发射机已经进入分布式录音模式,并且正在进行分布式录音。可以理解的是,图9仅仅示意了一种可能的分布式录音提示图标801,具体分布式录音提示图标801的样式以及显示位置可以根据实际情况进行任意设计,本申请在此不作限定。
分布式录音发射机配置模块还在接收到本机录音算法模块发送的音频数据后,还可以将接收到的音频数据通过WiFi接口发送至硬件层的WiFi子系统。WiFi子系统还可以用于将音频数据发送至接收机。较为具体的,例如WiFi接口将音频数据发送至位于硬件层的硬件接口,再经用硬件接口将音频数据传输至WiFi协议栈,并在WiFi协议栈内对音频数据进行相应的处理,然后WiFi协议栈将处理后的音频数据经天线将发送至接收机。可以理解的是,在WiFi协议栈内对音频数据进行相应的处理的具体方式可以参考当前的已知技术,为方便描述在此不再赘述。
在另一个例子中,与图4所示的场景相比,接收机还可以接收到多个发射机发送的分布式录音响应消息,并控制多个发射机进行拾音。此时的分布式录音场景即为多个发射机对1个接收机的分布式录音场景,可以例如图10示出的,拍摄者通过接收机进行视频拍摄,此时被摄对象周围可以存在多个发射机,例如发射机1、发射机2……发射机N,其中N为正整数。当然发射机也可以位于其它位置,例如发射机3,可以用于录制旁白或者解说。当然,若存在多个被摄对象,则发射机也可以分配给多个被摄对象,本申请对于发射机的位置不作具体限定。由于此场景下包括多个发射机和1个接收机,因此在该场景下,分布式录音模式可以为第二子分布式录音模式,用于表示多个发射机与一个接收机通过无线方式进行连接,并进行分布式录音。
在图10所示的场景下,发射机与图7和图9所示的发射机相同,具体可以参考相应的描述,本申请在此不再赘述。对于接收机架构,相比图5所示的接收机的架构,区别仅在多机录音算法模块。
图11为本申请实施例提供的另一种接收机架构示意图。由于图10的场景为一个接收机和多个发射机的分布式录音场景,由于不涉及本地录音模式,因此图11中并未示出本地录音算法模块以及相应的模数转换器和多个MIC。对于为一个接收机和多个 发射机的分布式录音场景下的接收机,其分布式录音系统配置模块在接收到摄像APP发送的第一指令后,可以通过WiFi接口将分布式录音请求传输至硬件层的WiFi子模块,并通过WiFi子模块将分布式录音指令发送至多个发射机。并在接收到多个发射机发送的分布式录音响应消息后,还可以通过相同的方式将启动分布式录音指令通过WiFi接口传输至硬件层的WiFi子模块,以及通过WiFi子模块将启动分布式录音指令发送至允许进行分布式录音的多个发射机。从而保障多个发射机可以同时向接收机发送采集到的音频数据。其具体实现方式可以参考图5中相应的描述,在此不再赘述。
若存在多个发射机允许进行分布式录音,则接收机的WiFi子模块还可以接收来自多个发射机发送的音频数据。并将接收到的多个音频数据传输至framework层的WiFi接口,并通过WiFi接口将接收的多个音频数据传输至多机录音算法模块中。由于发射机所处的位置可能并不相同,因此在通过无线传输音频数据的过程中会导致传输时间不同,因此接收时间必然也会不同。因此对于非同一时间接收到的音频数据,可以采用多机录音算法模块中的时间对齐子模块将多个音频数据进行时间对齐,从而使得输出的音频数据时间顺序是正确的,避免相互之间的干扰。例如可以通过相应的检测以及延时缓冲区实现多个音频数据的时间对齐,其具体实现方式可以采用现有方式,在此不再赘述。在对多个音频数据进行时间对齐之后,可以启用多机录音算法模块中的录音处理子模块对接收到的多个音频数据进行录音处理。例如可以进行降噪、降混响以及人声增强等可能的操作。具体实现方式可以参考图5中相应部分的描述,在此不再赘述。当然,对于接收到的多个音频数据,若接收到的是经过压缩的音频数据,此时还可以启用多机录音算法模块中的音频解码子模块,将压缩过的音频数据进行解压操作,并得到原始的音频数据。应当注意的是,对多个音频数据进行时间对齐和录音处理的顺序可以根据实际情况进行调整,例如先对多个音频数据进行时间对齐再进行录音处理,或是先对多个音频数据进行录音处理再进行时间对齐。但可以理解的是,为了避免录音处理时丢失部分音频数据导致时间对齐出现的问题,因此通常情况下,选择先对多个音频数据进行时间对齐再进行录音处理。
可以理解的是,图11中的摄像APP也可以在显示屏上显示分布式录音图标,其与图5中摄像APP显示的方式完全相同,具体可参考相应的描述,在此不再赘述。
图12为本申请实施例提供的一种多机协同分布式录音场景示意图。
图12与图10相比,此时作为视频拍摄的接收机也同时进行录音。而与图5相比,则作为视频拍摄的接收机进行录音的同时还包括多个发射机同时进行录音。显然,在图12所示的场景下,接收机和多个发射机协同进行录音,从而保障可以清晰录制拍摄对象的声音以及旁白和解说。
在该场景下,例如图13所示,本申请提供了又一种接收机架构示意图。摄像APP可以在显示屏上显示分布式录音图标,以便用户在使用时,可以通过该分布式录音图标选择不同的录音模式。录音模式可以包括分布式录音、多机协同录音和本机录音。其具体的显示方式与图5中摄像APP的显示方式完全相同,具体可参考相应的描述,在此不再赘述。
在一个例子中,分布式录音系统配置模块接收摄像APP发送的模式选择指令,还可以是用于告知分布式录音系统配置模块用户选择了多机协同录音模式的第三指令。 当framework层的分布式录音系统配置模块接收到摄像APP发送的第三指令后,分布式录音系统配置模块可以确定拍摄者选择的录音模式为多机协同录音模式。之后,分布式录音系统配置模块根据第三指令生成分布式录音请求,并向WiFi接口发送该分布式录音请求。WiFi接口在接收到分布式录音请求后,会将该分布式录音请求转发至硬件层的WiFi子系统。并通过WiFi子系统将该请求发送至被摄对象周围的一个或多个发射机。具体发送分布式录音请求的方式可以参考图5部分相应的描述,在此不再赘述。当然在另一个例子中,可以分布式录音系统配置模块根据第三指令生成多机协同录音请求,可以理解的是,多机协同录音请求的作用与上述分布式录音请求的作用相同。
同时,当分布式录音系统配置模块接收到第三指令后,还可以生成多机协同录音指令,并向多机录音算法模块发送该多机协同录音指令。在对音频数据处理之前,多机录音算法模块可以根据分布式录音系统配置模块发送的指令进行配置,例如,多机录音算法模块在接收到多机协同录音指令后,会根据多机协同录音指令进行配置,选择WiFi接口的输出以及本机录音算法模块的输出同时作为音频数据流的输入。
若拍摄者选择了多机协同录音模式,则摄像APP可以通过setparameter接口直接向本机录音算法模块发送本机录音控制指令,以便本机算法模块控制硬件层的一个或多个MIC采集到音频数据的模拟信号。本机算法模块运行的图具体方式与本机录音模式相同,可以参考本机录音模式部分的相应描述,在此不再赘述。
接收机的WiFi子系统在发送分布式录音请求之后,还可以接收来自一个或多个发射机发送的分布式录音响应消息。当然,若接收机发送的为多机协同录音请求,则可以接收多机协同录音响应消息。可以理解的是多机协同录音响应消息的功能与作用和分布式录音响应消息相同,后续均以分布式录音响应消息为例进行描述,但可以想到的是分布式录音响应消息可以替换为多机协同录音响应消息。对于图12所示的场景,可以是仅接收到来自一个发射机发送的分布式录音响应消息。此时的分布式录音场景即为1个发射机与1个接收机的分布式录音场景。又或是可以接收到来自多个发射机发送的分布式录音响应消息。此时的分布式录音场景为多个发射机与1个接收机的分布式录音场景。较为具体的,可以通过天线接收到一个或多个发射机发送的分布式录音响应消息。天线在接收到分布式录音响应消息后,可以经过WiFi协议栈对接收到的分布式录音响应消息进行相应的处理,再通过硬件接口将处理后的分布式录音响应消息传输至framework层中的WiFi接口。可以理解的是,在WiFi协议栈内对接收到音频数据进行相应的处理的具体方式可以参考当前的已知技术,为方便描述在此不再赘述。WiFi接口将分布式录音响应消息发送至分布式录音系统配置模块,以便分布式录音系统配置模块可以根据分布式录音响应消息确定哪些发射机允许进行分布式录音。
当分布式录音系统配置模块根据分布式录音响应消息确定了一个或多个发射机允许进行分布式录音之后,分布式录音系统配置模块生成启动分布式录音指令,并向WiFi接口发送该启动分布式录音指令。WiFi接口在接收到启动分布式录音指令后,会将该启动分布式录音指令转发至硬件层的WiFi子系统。并通过WiFi子系统将该启动分布式录音指令发送至允许进行分布式录音的发射机。可以理解的是,发送启动分布式录音指令与发送分布式录音请求的过程相同,为方便描述在此不再赘述。在另一个 例子中,当分布式录音系统配置模块根据多机协同录音响应消息确定了一个或多个发射机允许进行分布式录音之后,分布式录音系统配置模块还可以生成启动多机协同录音指令。可以理解的是,启动多机协同录音指令与启动分布式录音指令的功能、作用完全相同。
接收机的WiFi子系统在发送启动分布式录音指令(或启动多机协同录音指令)之后,还可以接收来自允许进行分布式录音的发射机发送的音频数据。其具体方式可以参考分布式录音模式中相应的描述,在此不再赘述。
由于多机协同录音模式中多机录音算法模块同时具有两个音频数据的输入接口,因此,对于WiFi接口输入的音频数据流可以与本机录音算法模块输入的音频数据流同时输入时间对齐子模块中进行时间对齐。当然,若发射机发送的音频数据是经过压缩处理过的音频数据,则多机录音算法模块中的音频解码子模块可以先将压缩过的音频数据进行解压处理,得到原始的音频数据,然后将得到的原始的音频数据传输至时间对齐子模块中进行时间对齐。具体时间对齐的方式可以参考分布式录音模式中多个发射机情况下时间对齐的描述,在此不再赘述。经过时间对齐子模块进行时间对齐后的音频数据可以发送至录音处理子模块进行录音处理,其具体录音处理方式可以参考上述分布式录音模式中相应的描述,在此不再赘述。多机录音算法模块将录音处理后的音频数据直接发送至应用层的摄像APP。可以理解是的,多机录音算法模块将录音处理后的音频数据直接发送至应用层的摄像APP的方式可以参考分布式录音模式中相应的描述,在此不再赘述。同时,音频解码子模块作为可选的子模块,若发射机发送的音频数据并未经过压缩处理,则可以直接将音频数据发送至时间对齐子模块进行时间对齐。
对于上述图5至图13所述的方案,仅以WiFi作为一种可能的方式,当然还可以采用BT、ZigBee等任意无线传输方式进行等效替换,本申请在此不作限定。
可以理解的是,对于终端设备可以同时具有接收机的架构和发射机的架构,对于终端设备而言,其根据具体使用情况选择作为接收机或发射机中的一种。
在又一个例子中,上述图5至图13所述的方案可以理解为在每次使用过程中接收机与发射机是固定不变的。可以理解的是,在某些情况下,接收机与发射机的身份是可以互换的。例如接收机可以在显示屏上显示用于互换身份的提示信息。用户可以根据该提示信息选择该接收机与某个发射机进行身份互换。其中,接收机可以首先确定哪些发射机同时还可以作为接收机,并将相应的发射机对应的标识进行显示,以便用户根据显示的一个或多个发射机标识,选择其中一个与该接收机进行身份互换。当身份互换后,原发射机作为新的接收机实现接收机的所有功能,同时原接收机作为新的发射机实现发射机的所有功能。
图14为本申请实施例提供的一种分布式录音方法流程图。
本申请还提供了一种分布式录音方法,该方法可以实现上述图4至图11所述的方案。该方法主要用于分布式录音模式,方法可以包括以下步骤:
S1401,接收机接收用户的输入信息,并根据输入信息确定录音模式为分布式录音模式。
S1402,接收机发送分布式录音请求到一个或多个发射机。
S1403,发射机接收分布式录音请求,若允许进行分布式录音,则发送分布式录音响应消息。
S1404,接收机接收到一个或多个发射机发送的分布式录音响应消息,并将启动分布式录音指令发送至允许进行分布式录音的发射机。
S1405,允许进行分布式录音的发射机接收到启动分布式录音指令,并根据该指令采集音频数据。
S1406,允许进行分布式录音的发射机将采集的音频数据发送至接收机。
S1407,接收机一个或多个接收音频数据,并将接收到的音频数据发送至摄像应用。
图14所述方法主要针对于1个发射机与1个接收机的分布式录音场景,以及多个发射机与1个接收机的分布式录音场景,具体实现过程可参考图4至图11所述的方案,在此不再赘述。
图15为本申请实施例提供的另一种分布式录音方法流程图。
本申请还提供了另一种分布式录音方法,该方法可以实现上述图7至图9、图12、图13所述的方案。该方法主要用于多机协同录音模式,方法可以包括以下步骤:
S1501,接收机接收用户的输入信息,并根据输入信息确定录音模式为多机协同录音模式。
S1502,接收机发送分布式录音请求(或多机协同录音请求)到一个或多个发射机。
S1503,发射机接收分布式录音请求(或多机协同录音请求),若允许进行分布式录音(或多机协同录音),则发送分布式录音响应消息(或多机协同录音响应消息)。
S1504,接收机接收到一个或多个发射机发送的分布式录音响应消息(或多机协同录音响应消息),并将启动分布式录音指令(或启动多机协同录音指令)发送至允许进行分布式录音(或多机协同录音)的发射机。
S1505,允许进行分布式录音(或多机协同录音)的发射机接收到启动分布式录音指令(或启动多机协同录音指令),并根据该指令采集音频数据。
S1506,允许进行分布式录音(或多机协同录音)的发射机将采集的音频数据发送至接收机。
可以理解的是当确定录音模式为多机协同录音模式时,还可以执行S1507。
S1507,接收机启动本机录音,并采集本机音频数据。
应当注意的是,S1507执行的时刻可以在S1501之后以及S1508之前的任意时刻,本申请在此不作限定。
S1508,接收机接收到多个音频数据以及本机音频数据,并将接收到的多个音频数据和本机音频数据发送至摄像应用。
图15所述方法主要针对于1个发射机与1个接收机的多机协同录音场景,以及多个发射机与1个接收机的多机协同录音场景,具体实现过程可参考图7至图9、图12、图13所述的方案,在此不再赘述。
图16为本申请实施例提供的又一种分布式录音方法流程图。
本申请还提供了又一种分布式录音方法,该方法可以实现上述图4至图11所述的方案。该方法主要用于本机录音模式,方法可以包括以下步骤:
S1601,接收机接收用户的输入信息,并根据输入信息确定录音模式为本机录音模 式。
S1602,接收机启动本机录音,并采集本机音频数据。
S1603,接收机将本机音频数据发送至摄像应用。
图16所述方法主要针对于1个接收机的本机录音场景,具体实现过程可以参考图4至图11所述的方案,在此不再赘述。
图17为本申请实施例提供的一种分布式录音设备示意图。
如图17所示,提供了一种分布式录音设备1700,该设备1700可以为上述图4至图13所涉及方案中的发射机或接收机。
该设备1700可以包括处理器1710,外部存储器接口1720,内部存储器1721,通用串行总线(universal serial bus,USB)接口1730,充电管理模块1740,电源管理模块1741,电池1742,天线1,天线2,移动通信模块1750,无线通信模块1760,音频模块1770,扬声器1770A,受话器1770B,麦克风1770C,耳机接口1770D,传感器模块1780,按键1790,马达1791,指示器1792,摄像头1793,显示屏1794,以及用户标识模块(subscriber identification module,SIM)卡接口1795等。其中,传感器模块1780可以包括压力传感器1780A,陀螺仪传感器1780B,气压传感器1780C,磁传感器1780D,加速度传感器1780E,距离传感器1780F,接近光传感器1780G,指纹传感器1780H,温度传感器1780J,触摸传感器1780K,环境光传感器1780L,骨传导传感器1780M等。
可以理解的是,本发明实施例示意的结构并不构成对分布式录音设备1700的具体限定。在本申请另一些实施例中,分布式录音设备1700可以包括比图示更多或更少的部件,或者组合某些部件,或者拆分某些部件,或者不同的部件布置。图示的部件可以以硬件,软件或软件和硬件的组合实现。
处理器1710可以包括一个或多个处理单元,例如:处理器1710可以包括应用处理器(application processor,AP),调制解调处理器,图形处理器(graphics processing unit,GPU),图像信号处理器(image signal processor,ISP),控制器,视频编解码器,数字信号处理器(digital signal processor,DSP),基带处理器,和/或神经网络处理器(neural-network processing unit,NPU)等。其中,不同的处理单元可以是独立的器件,也可以集成在一个或多个处理器中。
控制器可以根据指令操作码和时序信号,产生操作控制信号,完成取指令和执行指令的控制。
处理器1710中还可以设置存储器,用于存储指令和数据。在一些实施例中,处理器1710中的存储器为高速缓冲存储器。该存储器可以保存处理器1710刚用过或循环使用的指令或数据。如果处理器1710需要再次使用该指令或数据,可从所述存储器中直接调用。避免了重复存取,减少了处理器1710的等待时间,因而提高了系统的效率。
在一些实施例中,处理器1710可以包括一个或多个接口。接口可以包括集成电路(inter-integrated circuit,I2C)接口,集成电路内置音频(inter-integrated circuit sound,I2S)接口,脉冲编码调制(pulse code modulation,PCM)接口,通用异步收发传输器(universal asynchronous receiver/transmitter,UART)接口,移动产业处理器接口(mobile industry processor interface,MIPI),通用输入输出(general-purpose input/output, GPIO)接口,用户标识模块(subscriber identity module,SIM)接口,和/或通用串行总线(universal serial bus,USB)接口等。
USB接口1730是符合USB标准规范的接口,具体可以是Mini USB接口,Micro USB接口,USB Type C接口等。USB接口1730可以用于连接充电器为分布式录音设备1700充电,也可以用于分布式录音设备1700与外围设备之间传输数据。也可以用于连接耳机,通过耳机播放或采集音频。该接口还可以用于连接其他电子设备,例如AR设备等。
可以理解的是,本发明实施例示意的各模块间的接口连接关系,只是示意性说明,并不构成对分布式录音设备1700的结构限定。在本申请另一些实施例中,分布式录音设备1700也可以采用上述实施例中不同的接口连接方式,或多种接口连接方式的组合。
分布式录音设备1700的无线通信功能可以通过天线1,天线2,移动通信模块1750,无线通信模块1760,调制解调处理器以及基带处理器等实现。
天线1和天线2用于发射和接收电磁波信号。分布式录音设备1700中的每个天线可用于覆盖单个或多个通信频带。不同的天线还可以复用,以提高天线的利用率。例如:可以将天线1复用为无线局域网的分集天线。在另外一些实施例中,天线可以和调谐开关结合使用。
移动通信模块1750可以提供应用在分布式录音设备1700上的包括2G/3G/4G/5G等无线通信的解决方案。移动通信模块1750可以包括至少一个滤波器,开关,功率放大器,低噪声放大器(low noise amplifier,LNA)等。移动通信模块1750可以由天线1接收电磁波,并对接收的电磁波进行滤波,放大等处理,传送至调制解调处理器进行解调。移动通信模块1750还可以对经调制解调处理器调制后的信号放大,经天线1转为电磁波辐射出去。在一些实施例中,移动通信模块1750的至少部分功能模块可以被设置于处理器1710中。在一些实施例中,移动通信模块1750的至少部分功能模块可以与处理器1710的至少部分模块被设置在同一个器件中。
调制解调处理器可以包括调制器和解调器。其中,调制器用于将待发送的低频基带信号调制成中高频信号。解调器用于将接收的电磁波信号解调为低频基带信号。随后解调器将解调得到的低频基带信号传送至基带处理器处理。低频基带信号经基带处理器处理后,被传递给应用处理器。应用处理器通过音频设备(不限于扬声器1770A,受话器1770B等)输出声音信号,或通过显示屏1794显示图像或视频。在一些实施例中,调制解调处理器可以是独立的器件。在另一些实施例中,调制解调处理器可以独立于处理器1710,与移动通信模块1750或其他功能模块设置在同一个器件中。
无线通信模块1760可以提供应用在电子设备1700上的包括无线局域网(wireless local area networks,WLAN)(如WiFi网络),蓝牙,全球导航卫星系统(global navigation satellite system,GNSS),调频(frequency modulation,FM),近距离无线通信技术(near field communication,NFC),红外技术(infrared,IR)等无线通信的解决方案。无线通信模块1760可以是集成至少一个通信处理模块的一个或多个器件。无线通信模块1760经由天线2接收电磁波,将电磁波信号调频以及滤波处理,将处理后的信号发送到处理器1710。无线通信模块1760还可以从处理器1710接收待发送的信号,对其进行调频,放大,经天线2转为电磁波辐射出去。
在一些实施例中,分布式录音设备1700的天线1和移动通信模块1750耦合,天线2和无线通信模块1760耦合,使得分布式录音设备1700可以通过无线通信技术与网络以及其他设备通信。
显示屏1794用于显示图像,视频等。显示屏1794包括显示面板。显示面板可以采用液晶显示屏(liquid crystal display,LCD),有机发光二极管(organic light-emitting diode,OLED),有源矩阵有机发光二极体或主动矩阵有机发光二极体(active-matrix organic light emitting diode的,AMOLED),柔性发光二极管(flex light-emitting diode,FLED),Miniled,MicroLed,Micro-oLed,量子点发光二极管(quantum dot light emitting diodes,QLED)等。在一些实施例中,分布式录音设备1700可以包括1个或N个显示屏1794,N为大于1的正整数。
分布式录音设备1700可以通过ISP,摄像头1793,视频编解码器,GPU,显示屏1794以及应用处理器等实现拍摄功能。
摄像头1793用于捕获静态图像或视频。物体通过镜头生成光学图像投射到感光元件。感光元件可以是电荷耦合器件(charge coupled device,CCD)或互补金属氧化物半导体(complementary metal-oxide-semiconductor,CMOS)光电晶体管。感光元件把光信号转换成电信号,之后将电信号传递给ISP转换成数字图像信号。ISP将数字图像信号输出到DSP加工处理。DSP将数字图像信号转换成标准的RGB,YUV等格式的图像信号。在一些实施例中,分布式录音设备1700可以包括1个或M个摄像头1793,M为大于1的正整数。
外部存储器接口1720可以用于连接外部存储卡,例如Micro SD卡,实现扩展分布式录音设备1700的存储能力。外部存储卡通过外部存储器接口1720与处理器1710通信,实现数据存储功能。例如将音乐,视频等文件保存在外部存储卡中。
内部存储器1721可以用于存储计算机可执行程序代码,所述可执行程序代码包括指令。内部存储器1721可以包括存储程序区和存储数据区。其中,存储程序区可存储操作系统,至少一个功能所需的应用程序(比如声音播放功能,图像播放功能等)等。存储数据区可存储分布式录音设备1700使用过程中所创建的数据(比如音频数据等)等。此外,内部存储器1721可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件,闪存器件,通用闪存存储器(universal flash storage,UFS)等。处理器1710通过运行存储在内部存储器1721的指令,和/或存储在设置于处理器中的存储器的指令,执行分布式录音设备1700的各种功能应用以及数据处理。
分布式录音设备1700可以通过音频模块1770,扬声器1770A,受话器1770B,麦克风1770C,耳机接口1770D,以及应用处理器等实现音频功能。例如音乐播放,录音等。
音频模块1770用于将数字音频信息转换成模拟音频信号输出,也用于将模拟音频输入转换为数字音频信号。音频模块1770还可以用于对音频信号编码和解码。例如进行压缩和解压缩。在一些实施例中,音频模块1770可以设置于处理器1710中,或将音频模块1770的部分功能模块设置于处理器1710中,以便处理器对数字音频信号进行处理。在一个例子中,音频模块1770例如可以是codec。
麦克风1770C,也称“话筒”,“传声器”,用于将声音信号转换为电信号。当 进行录音时,可以通过在靠近麦克风1770C的发声,以便将声音信号输入到麦克风1770C。分布式录音设备1700可以设置至少一个麦克风1770C,例如设置3-4个麦克风。在另一些实施例中,分布式录音设备1700可以包括两个麦克风1770C,以便在采集声音信号的同时,还可以实现降噪功能。在另一些实施例中,分布式录音设备1700还可以设置三个,四个或更多麦克风1770C,实现采集声音信号,降噪,还可以识别声音来源,实现定向录音功能等。
耳机接口1770D用于连接有线耳机。耳机接口1770D可以是USB接口1730,也可以是3.5mm的开放移动电子设备平台(open mobile terminal platform,OMTP)标准接口,美国蜂窝电信工业协会(cellular telecommunications industry association of the USA,CTIA)标准接口。
分布式录音设备1700的软件系统可以采用分层架构,事件驱动架构,微核架构,微服务架构,或云架构。本申请实施例可以是采用分层架构的Android系统。
分布式录音设备1700可以执行上述图4至图16中的发射机或接收机的任意一项功能,具体可以参考图4至图16的描述,在此不再赘述。
本申请通过将终端设备作为发射机和接收机,由于终端设备的偶记程度很高,并且市场存量很大,因此保障了用户无需购买任何其它外置设备的前提下,可以随时随地清晰录制拍摄对象的声音以及旁白和解说,同时获得良好的录音效果。
本领域普通技术人员应该还可以进一步意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、计算机软件或者二者的结合来实现,为了清楚地说明硬件和软件的可互换性,在上述说明中已经按照功能一般性地描述了各示例的组成及步骤。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分步骤是可以通过程序来指令处理器完成,所述的程序可以存储于计算机可读存储介质中,所述存储介质是非短暂性(英文:non-transitory)介质,例如随机存取存储器,只读存储器,快闪存储器,硬盘,固态硬盘,磁带(英文:magnetic tape),软盘(英文:floppy disk),光盘(英文:optical disc)及其任意组合。
以上所述,仅为本申请较佳的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到的变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应该以权利要求的保护范围为准。

Claims (25)

  1. 一种录音系统,其特征在于,所述系统包括:至少两个终端设备,所述至少两个终端设备中包含第一设备和至少一个第二设备;
    所述第一设备,用于接收用户的输入信息,并根据所述输入信息确定录音模式;
    所述第一设备还用于,当所述录音模式为分布式录音模式时,发送启动分布式录音指令到至少一个所述第二设备;
    所述第二设备,用于接收所述启动分布式录音指令,并根据所述启动分布式录音指令周期性采集第一音频数据,其中,所述第一音频数据为所述第二设备采集到的本地音频数据;
    所述第二设备还用于,将周期性采集到的所述第一音频数据,并对所述第一音频数据进行人声增强和/或降噪处理,再以相同周期发送至所述第一设备;
    所述第一设备还用于,接收至少一个所述第一音频数据,以便摄像应用调用至少一个所述第一音频数据。
  2. 如权利要求1所述的系统,其特征在于,所述第二设备还用于:
    根据所述启动分布式录音指令确定录音参数;
    根据所述录音参数采集所述第一音频数据。
  3. 如权利要求1或2所述的系统,其特征在于,在所述第二设备将周期性采集到的所述第一音频数据以相同周期发送至所述第一设备之前,所述第二设备还用于:
    将周期性采集到的所述第一音频数据进行压缩处理,得到压缩后的所述第一音频数据。
  4. 如权利要求1-3任一所述的系统,其特征在于,所述第一设备还用于:
    当所述第一音频数据为压缩过的数据,对压缩的所述第一音频数据进行音频解压缩,得到解压缩后的第一音频数据。
  5. 如权利要求1-4任一所述的系统,其特征在于,所述分布式录音模式包括:多机协同录音模式;
    当所述录音模式为多机协同录音模式时,在所述摄像应用调用至少一个所述第一音频数据之前,所述第一设备还用于:
    采集第二音频数据,所述摄像应用调用至少一个所述第一音频数据和所述第二音频数据,其中,所述第二音频数据为所述第一设备采集到的本地音频数据。
  6. 如权利要求1-5任一所述的系统,其特征在于,在所述摄像应用调用至少一个所述第一音频数据之前,所述第一设备还用于:
    当所述第一音频数据为多个时,对多个所述第一音频数据进行时间对齐操作;或
    当所述录音模式为多机协同录音模式时,对至少一个所述第一音频数据和第二音频数据进行时间对齐操作。
  7. 如权利要求5或6所述的系统,其特征在于,在所述摄像应用调用至少一个所述第二音频数据之前,所述第一设备还用于:
    对所述第二音频数据进行降噪和/或人声增强处理。
  8. 如权利要求1-7任一所述的系统,其特征在于,所述分布式录音模式包括:第一子分布式录音模式和第二子分布式录音模式;
    当所述录音模式为第一子分布式录音模式时,所述第一设备还用于发送分布式录音指令到一个所述第二设备,以及接收所述第二设备发送的所述第一音频数据;或者,
    当所述录音模式为第二子分布式录音模式时,所述第一设备还用于发送分布式录音指令到多个所述第二设备,以及接收多个所述第二设备发送的所述第一音频数据。
  9. 如权利要求1-8任一所述的系统,其特征在于,所述第一设备在发送启动分布式录音指令到至少一个第二设备之前,所述系统还包括:
    所述第一设备还用于,发送分布式录音请求到至少一个第二设备;
    所述第二设备还用于,接收所述第一设备发送的分布式录音请求;
    所述第二设备还用于,在所述第二设备的显示屏上显示提示信息,所述提示信息用于提示是否允许该所述第二设备进行分布式录音;
    所述第二设备还用于,向所述第一设备发送分布式录音响应消息,所述分布式录音响应消息用于表示是否允许该所述第二设备进行分布式录音;
    所述第一设备还用于,接收至少一个第二设备发送的分布式录音响应消息;
    所述第一设备还用于,向允许进行分布式录音的至少一个所述第二设备发送所述启动分布式录音指令。
  10. 如权利要求1-9任一所述的系统,其特征在于,所述录音模式还包括本地录音模式;
    所述第一设备还用于,根据用户输入的输入信息,在所述本地录音模式、第一子分布式录音模式、第二子分布式录音模式和多机协同录音模式中进行切换。
  11. 如权利要求1-10任一所述的系统,其特征在于,所述第一设备还用于:
    在所述第一设备的显示屏上显示分布式录音图标。
  12. 如权利要求1-11任一所述的系统,其特征在于,所述第二设备还用于:
    在所述第二设备的显示屏上显示分布式录音提示图标。
  13. 一种录音终端设备,其特征在于,所述录音终端设备为第一设备或第二设备;所述录音终端设备包括:显示器、处理器、存储器、发送器和接收器;
    当所述录音终端设备为第一设备时,所述接收器,用于接收用户的输入信息,并根据所述输入信息确定录音模式;
    所述处理器,用于当所述录音模式为分布式录音模式时,控制所述发送器发送启动分布式录音指令到至少一个所述第二设备;
    当所述录音终端设备为第二设备时,所述接收器,用于接收所述启动分布式录音指令,并根据所述启动分布式录音指令周期性采集第一音频数据,其中,所述第一音频数据为所述第二设备采集到的本地音频数据;
    所述发送器,用于将周期性采集到的所述第一音频数据,并对所述第一音频数据进行人声增强和/或降噪处理,再以相同周期发送至所述第一设备;
    当所述录音终端设备为第一设备时,所述接收器还用于,接收至少一个所述第一音频数据;
    所述处理器还用于,控制摄像应用调用至少一个所述第一音频数据。
  14. 如权利要求13所述的录音终端设备,其特征在于,当所述录音终端设备为第 二设备时,所述处理器用于:
    根据所述启动分布式录音指令确定录音参数;
    根据所述录音参数控制麦克风采集所述第一音频数据。
  15. 如权利要求13或14所述的录音终端设备,其特征在于,当所述录音终端设备为第二设备时,所述处理器还用于:
    将周期性采集到的所述第一音频数据进行压缩处理,得到压缩后的所述第一音频数据。
  16. 如权利要求13-15任一所述的录音终端设备,其特征在于,当所述录音终端设备为第二设备时,所述处理器还用于:
    当所述第一音频数据为压缩过的数据,对压缩的所述第一音频数据进行音频解压缩,得到解压缩后的第一音频数据。
  17. 如权利要求13-16任一所述的录音终端设备,其特征在于,所述分布式录音模式包括:多机协同录音模式;
    当所述录音终端设备为第一设备时,所述处理器还用于:
    当所述录音模式为多机协同录音模式时,采集第二音频数据,所述摄像应用调用至少一个所述第一音频数据和所述第二音频数据,其中,所述第二音频数据为所述第一设备采集到的本地音频数据。
  18. 如权利要求13-17任一所述的录音终端设备,其特征在于,当所述录音终端设备为第一设备时,所述处理器还用于:
    当所述第一音频数据为多个时,对多个所述第一音频数据进行时间对齐操作;或
    当所述录音模式为多机协同录音模式时,对至少一个所述第一音频数据和第二音频数据进行时间对齐操作。
  19. 如权利要求17或18所述的录音终端设备,其特征在于,当所述录音终端设备为第一设备时,所述处理器还用于:
    对所述第二音频数据进行降噪和/或人声增强处理。
  20. 如权利要求13-19任一所述的录音终端设备,其特征在于,所述分布式录音模式包括:第一子分布式录音模式和第二子分布式录音模式;
    当所述录音终端设备为第一设备时,
    所述发送器还用于,当所述录音模式为第一子分布式录音模式时,发送分布式录音指令到一个所述第二设备;或者,所述发送器还用于,当所述录音模式为第二子分布式录音模式时,发送分布式录音指令到多个所述第二设备;
    所述接收器还用于,当所述录音模式为第一子分布式录音模式时,接收所述第二设备发送的所述第一音频数据;或者,当所述录音模式为第二子分布式录音模式时,接收多个所述第二设备发送的所述第一音频数据。
  21. 如权利要求13-20任一所述的录音终端设备,其特征在于,当所述录音终端设备为第一设备时,所述发送器还用于,发送分布式录音请求到至少一个第二设备;
    当所述录音终端设备为第二设备时,所述接收器还用于,接收所述第一设备发送的分布式录音请求;
    所述显示器还用于,在所述第二设备的显示屏上显示提示信息,所述提示信息用 于提示是否允许该所述第二设备进行分布式录音;
    所述发送器还用于,向所述第一设备发送分布式录音响应消息,所述分布式录音响应消息用于表示是否允许该所述第二设备进行分布式录音;
    当所述录音终端设备为第一设备时,所述接收器还用于,接收至少一个第二设备发送的分布式录音响应消息;
    所述发送器还用于,向允许进行分布式录音的至少一个所述第二设备发送所述启动分布式录音指令。
  22. 如权利要求13-21任一所述的录音终端设备,其特征在于,所述录音模式还包括本地录音模式;
    当所述录音终端设备为第一设备时,所述处理器还用于,根据用户输入的输入信息,在所述本地录音模式、第一子分布式录音模式、第二子分布式录音模式和多机协同录音模式中进行切换。
  23. 如权利要求13-22任一所述的录音终端设备,其特征在于,当所述录音终端设备为第一设备时,所述显示器用于:
    在所述第一设备的显示屏上显示分布式录音图标。
  24. 如权利要求13-23任一所述的录音终端设备,其特征在于,当所述录音终端设备为第二设备时,所述显示器用于:
    在所述第二设备的显示屏上显示分布式录音提示图标。
  25. 一种计算机可读存储介质,所述计算机可读存储介质中存储有指令,其特征在于,当所述指令在终端上运行时,使得所述终端执行如权利要求1-12任意一项所述的方法。
PCT/CN2021/107868 2020-08-25 2021-07-22 一种录音方法和设备 WO2022042151A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US18/042,808 US20230319217A1 (en) 2020-08-25 2021-07-22 Recording Method and Device
EP21859991.8A EP4187535A4 (en) 2020-08-25 2021-07-22 AUDIO RECORDING METHOD AND DEVICE

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010863700.9 2020-08-25
CN202010863700.9A CN114121026A (zh) 2020-08-25 2020-08-25 一种录音方法和设备

Publications (1)

Publication Number Publication Date
WO2022042151A1 true WO2022042151A1 (zh) 2022-03-03

Family

ID=80352628

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/107868 WO2022042151A1 (zh) 2020-08-25 2021-07-22 一种录音方法和设备

Country Status (4)

Country Link
US (1) US20230319217A1 (zh)
EP (1) EP4187535A4 (zh)
CN (1) CN114121026A (zh)
WO (1) WO2022042151A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114422710A (zh) * 2022-03-29 2022-04-29 荣耀终端有限公司 电子设备的录像控制方法、电子设备及可读介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150032797A1 (en) * 2013-07-23 2015-01-29 Adobe Systems Incorporated Distributed audio playback and recording
CN107393568A (zh) * 2017-08-16 2017-11-24 广东小天才科技有限公司 一种多媒体文件的录制方法、系统及终端设备
CN109413215A (zh) * 2018-12-24 2019-03-01 苏州思必驰信息科技有限公司 用于录音设备和用户终端的数据同步方法和系统
CN109412886A (zh) * 2018-09-26 2019-03-01 北京文香信息技术有限公司 一种分布式录播系统及其通信检测方法、录播控制方法
WO2020032788A1 (en) * 2018-08-10 2020-02-13 Poqeapp B.V. Computer-implemented method for communication between a first electronic device and a second electronic device
CN110827845A (zh) * 2019-11-18 2020-02-21 西安闻泰电子科技有限公司 录音方法、装置、设备及存储介质

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9325781B2 (en) * 2005-01-31 2016-04-26 Invention Science Fund I, Llc Audio sharing
US8983844B1 (en) * 2012-07-31 2015-03-17 Amazon Technologies, Inc. Transmission of noise parameters for improving automatic speech recognition
US9438993B2 (en) * 2013-03-08 2016-09-06 Blackberry Limited Methods and devices to generate multiple-channel audio recordings
US9947364B2 (en) * 2015-09-16 2018-04-17 Google Llc Enhancing audio using multiple recording devices

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150032797A1 (en) * 2013-07-23 2015-01-29 Adobe Systems Incorporated Distributed audio playback and recording
CN107393568A (zh) * 2017-08-16 2017-11-24 广东小天才科技有限公司 一种多媒体文件的录制方法、系统及终端设备
WO2020032788A1 (en) * 2018-08-10 2020-02-13 Poqeapp B.V. Computer-implemented method for communication between a first electronic device and a second electronic device
CN109412886A (zh) * 2018-09-26 2019-03-01 北京文香信息技术有限公司 一种分布式录播系统及其通信检测方法、录播控制方法
CN109413215A (zh) * 2018-12-24 2019-03-01 苏州思必驰信息科技有限公司 用于录音设备和用户终端的数据同步方法和系统
CN110827845A (zh) * 2019-11-18 2020-02-21 西安闻泰电子科技有限公司 录音方法、装置、设备及存储介质

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP4187535A4

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114422710A (zh) * 2022-03-29 2022-04-29 荣耀终端有限公司 电子设备的录像控制方法、电子设备及可读介质

Also Published As

Publication number Publication date
CN114121026A (zh) 2022-03-01
EP4187535A4 (en) 2024-01-17
EP4187535A1 (en) 2023-05-31
US20230319217A1 (en) 2023-10-05

Similar Documents

Publication Publication Date Title
US11812098B2 (en) Projected audio and video playing method and electronic device
CN114009055B (zh) 一种投屏显示方法及电子设备
CN108243481B (zh) 文件传输方法及装置
WO2022121775A1 (zh) 一种投屏方法及设备
WO2020151385A1 (zh) 一种控制智能家居设备连网的方法及设备
WO2021129521A1 (zh) 蓝牙通信方法和装置
WO2021104114A1 (zh) 一种提供无线保真WiFi网络接入服务的方法及电子设备
WO2021017909A1 (zh) 一种通过nfc标签实现功能的方法、电子设备及系统
WO2022062739A1 (zh) 一种多设备蓝牙配对的通信方法和系统
WO2022148319A1 (zh) 视频切换方法、装置、存储介质及设备
WO2022042151A1 (zh) 一种录音方法和设备
WO2024012345A1 (zh) 一种投屏画面的处理方法及相关装置
WO2023273763A1 (zh) 一种视频数据的传输方法及装置
US11368611B2 (en) Control method for camera device, camera device, camera system, and storage medium
WO2022042261A1 (zh) 屏幕共享方法、电子设备及系统
KR100640429B1 (ko) 다기능 장치 및 그의 제어방법
CN115525366A (zh) 一种投屏方法及相关装置
CN116981108B (zh) 无线投屏连接方法、移动终端及计算机可读存储介质
WO2023160214A1 (zh) 蓝牙耳机、音频输出方法及音频输出系统
WO2023071730A1 (zh) 声纹注册方法及电子设备
WO2024001735A1 (zh) 网络连接方法、电子设备及存储介质
WO2022095581A1 (zh) 数据传输方法及终端设备
CN115145517A (zh) 一种投屏方法、电子设备和系统
CN114840157A (zh) 一种组合显示方法、移动设备及电子设备
CN115119214A (zh) 一种立体声组网方法、系统及相关装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21859991

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2021859991

Country of ref document: EP

Effective date: 20230224

NENP Non-entry into the national phase

Ref country code: DE