WO2024001138A1 - 传输音频数据的装置、方法以及电子设备 - Google Patents

传输音频数据的装置、方法以及电子设备 Download PDF

Info

Publication number
WO2024001138A1
WO2024001138A1 PCT/CN2022/143772 CN2022143772W WO2024001138A1 WO 2024001138 A1 WO2024001138 A1 WO 2024001138A1 CN 2022143772 W CN2022143772 W CN 2022143772W WO 2024001138 A1 WO2024001138 A1 WO 2024001138A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
data
binary data
interface
module
Prior art date
Application number
PCT/CN2022/143772
Other languages
English (en)
French (fr)
Inventor
杨建新
余庆华
王泷
Original Assignee
哲库科技(上海)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 哲库科技(上海)有限公司 filed Critical 哲库科技(上海)有限公司
Publication of WO2024001138A1 publication Critical patent/WO2024001138A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/80Services using short range communication, e.g. near-field communication [NFC], radio-frequency identification [RFID] or low energy communication
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/173Transcoding, i.e. converting between two coded representations avoiding cascaded coding-decoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/724094Interfacing with a device worn on the user's body to provide access to telephonic functionalities, e.g. accepting a call, reading or composing a message
    • H04M1/724097Worn on the head
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/72412User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories using two-way short-range wireless interfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2250/00Details of telephonic subscriber devices
    • H04M2250/02Details of telephonic subscriber devices including a Bluetooth interface

Definitions

  • the embodiments of the present application relate to the field of audio technology, and more specifically, to an apparatus, method and electronic equipment for transmitting audio data.
  • the embodiments of the present application provide a device, method and electronic equipment for transmitting audio data. Various aspects of the embodiments of the present application are introduced below.
  • a device for transmitting audio data including: a main processing module configured to: convert initial PCM audio data into a first binary data group and a second binary data group; and to process the first binary data group.
  • a binary data group and the second binary data group are respectively subjected to audio encoding to obtain corresponding first encoding packets and second encoding packets;
  • a wireless audio module is connected to the main processing module through a first interface, Configured to respectively transmit the first encoded packet and the second encoded packet through a wireless channel.
  • a device for transmitting audio data including: a main processing module configured to generate initial PCM data; a wireless audio module connected to the main processing module through a first interface and configured to: generate initial PCM audio data Converting into a first binary data group and a second binary data group; and performing audio encoding on the first binary data group and the second binary data group respectively to obtain the corresponding first encoding packet and the second encoded packet; and transmitting the first encoded packet and the second encoded packet respectively through a wireless channel.
  • a method for transmitting audio data is provided.
  • the method is applied to a device for transmitting audio data.
  • the device is provided with a main processing module and a wireless audio module, and the main processing module and the wireless audio module pass
  • the first interface is connected, and the method includes: converting the initial PCM audio data into a first binary data group and a second binary data group; converting the first binary data group and the second binary data group;
  • the system data groups are respectively subjected to audio encoding to obtain corresponding first encoded packets and second encoded packets; and the transmitting module is used to transmit the first encoded packets and the second encoded packets respectively.
  • a fourth aspect provides an electronic device, including: a memory for storing code; and a processor for loading and executing the code in the memory to implement the method of transmitting audio data as described in the third aspect.
  • a chip including a processor configured to perform the method as described in the third aspect.
  • a sixth aspect provides an intelligent terminal, including the chip described in the fifth aspect.
  • the embodiment of the present application converts the initial PCM audio data into multiple binary data groups and transmits them separately, which helps to reduce the amount of retransmitted data and the wireless bandwidth required for audio data transmission, thereby avoiding audio lagging.
  • FIG. 1 is a schematic diagram of the basic flow of Bluetooth audio data processing.
  • FIG. 2 is a schematic diagram of another basic flow of Bluetooth audio data processing.
  • Figure 3 is a schematic structural diagram of a device for transmitting audio data provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a possible basic flow of audio data processing in the device of FIG. 3 .
  • FIG. 5 is a schematic diagram of another possible basic flow of audio data processing in the device of FIG. 3 .
  • Figure 6 is a schematic flowchart of a method for transmitting audio data provided by an embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
  • Figure 8 is a schematic structural diagram of a chip provided by an embodiment of the present application.
  • the electronic device mentioned in the embodiment of this application may refer to any type of electronic device capable of transmitting wireless audio signals.
  • the electronic device may be, for example, a mobile phone or a wireless headset.
  • the following is an example combining the data transmission process of mobile phones and wireless headsets.
  • wireless earphones In wireless earphones, the wire in the middle is replaced by radio waves, and the audio signal from the transmitter is sent to the earphone at the receiver through the radio waves.
  • wireless headphones are more convenient to use and are not bound and restricted by wires, so they have been widely used.
  • TWS wireless stereo
  • TWS headsets Due to the improvement in delay performance of TWS headsets, wired headsets have gradually replaced wired headsets in games. With the development of the noise reduction function of TWS headphones, in noisy scenes such as subways, buses or airports, using TWS noise reduction headphones will provide a better listening experience than ordinary wired headphones. Therefore, people are increasingly using wireless methods such as Bluetooth to spread and share audio.
  • FIG. 1 is a schematic diagram of the basic flow of Bluetooth audio data processing.
  • the mobile phone may include an audio digital signal processing (ADSP) module 110 and a Bluetooth (blue tooth, BT) transmitter module 120.
  • ADSP audio digital signal processing
  • BT Bluetooth
  • ADSP module 110 may encode or decode audio source data. Encoding and decoding technology can effectively reduce the space occupied by digital storage and improve the operating efficiency of the system.
  • the ADSP module 110 is also provided with a buffer to cache data that needs to be transmitted.
  • the BT transmitting module 120 may include a BT transceiver 122, which is used to convert the audio digital signal into an analog signal and transmit the analog signal through a wireless channel.
  • the BT side may also be provided with a buffer 121 to cache data that needs to be received.
  • the data processing process of audio signals is roughly divided into the following steps:
  • Step 1 The audio processing module 110 encodes the audio source data.
  • the audio source data format usually adopts pulse code modulation (PCM).
  • PCM pulse code modulation
  • the source data of audio signals is mostly 16-bit, 24-bit or 32-bit data.
  • the encoded data is temporarily stored in the buffer on the audio processing module 110 side.
  • Step 2 Transmit the data in the buffer on the audio processing module 110 side to the buffer 121 on the BT side through the data bus.
  • the BT transceiver 122 converts the audio digital signal into an analog signal and wirelessly transmits it to the TWS headset 130 .
  • data bus such as integrated circuit built-in audio bus (inter-IC sound, I2S), low-power inter-chip serial media bus (serial low-power inter-chip media bus, SLIMbus), Soundwire or serial bus. Serial peripheral interface (SPI), etc.
  • Step 3 The TWS earphone 130 receives the analog audio signal, and after corresponding analog-to-digital conversion processing, the sound is played from the earpiece.
  • the I2S interface is naturally used to transmit synchronous data. If the audio source file is encoded and converted into asynchronous data, it will take up less space. When decoding on the headset side, it may be necessary to use a private protocol for decoding.
  • the I2S interface is usually used to transmit the data before audio encoding.
  • FIG. 2 is a schematic diagram of another basic flow of Bluetooth audio data processing.
  • the mobile phone may include an audio ADSP module 210 and a BT transmitting module 220.
  • the BT transmitting module 220 may include a controller 222 and a BT transceiver 223. Controller 222 may encode or decode audio data. In addition, the BT side may also be provided with a buffer 221 to cache data that needs to be received.
  • the audio processing module 210 temporarily stores the audio source data into a buffer on the audio processing module 210 side.
  • the data in the buffer on the audio processing module 210 side is transmitted to the buffer 221 on the BT side through the data bus.
  • the BT controller 222 encodes the audio data, and then the BT transceiver 223 converts the audio digital signal into an analog signal and transmits it to TWS headset 230 receives.
  • Bluetooth is a low-cost, high-capacity short-distance wireless communication specification. Bluetooth technology is globally open and has good compatibility around the world. It usually uses radio waves from 2.4 to 2.485GHz.
  • each data packet in wireless transmission can be 5ms or 7.5ms. If retransmitted, it will seriously occupy the bandwidth. If the interference is serious and the number of retransmissions is too many, the wireless bandwidth of Bluetooth cannot meet this scenario, or instant communication cannot be achieved.
  • Bluetooth audio data transmission failure mentioned above is just an example, and the embodiments of the present application can be applied to any type of scenario in which wireless audio data transmission fails.
  • an embodiment of the present application proposes a device for transmitting audio data.
  • the embodiment of the present application will be described in detail below.
  • FIG. 3 is a schematic structural diagram of a device for transmitting audio data provided by an embodiment of the present application.
  • the device 300 for transmitting audio data may include a main processing module 310 and a wireless audio module 320.
  • the main processing module 310 may be, for example, an application processor (application processor, AP), or may be an ADSP module. In some embodiments, main processing module 310 may be configured to encode and decode initial PCM audio data. Generally, the processing performance of the main processing module is good, and the workload of the wireless audio module 320 can be reduced by performing the audio encoding and decoding work through the main processing module 310 .
  • the wireless audio module 320 may be, for example, a WiFi module, a 4G communication module or a Bluetooth module, etc.
  • Wireless audio module 320 may include a Bluetooth controller that may be configured to perform transmission of audio packets. Depending on the situation, the Bluetooth controller can also be configured to perform audio codecs. When the Bluetooth controller is independent of the main processor module, performing audio encoding and decoding or part of the work/steps through the Bluetooth controller can reduce the workload of the main processor module.
  • the main processing module may be formed as a first chip, and the wireless audio module may be formed as an independent second chip.
  • the two chips are jointly disposed on the device for transmitting audio data provided in the embodiment of the present application, and cooperate to perform audio encoding and decoding work.
  • the main processing module 310 transmits audio data to the wireless audio module 320 through the first interface.
  • the first interface can be a bus interface, for example.
  • the first interface may be an audio interface that supports asynchronous transmission of data, such as SLIMbus, Soundwire or SPI.
  • the main processing module 310 converts the initial PCM audio data into a first binary data group and a second binary data group, and performs audio encoding on the first binary data group and the second binary data group respectively to obtain Corresponding first encoding packet and second encoding packet.
  • the main processing module 310 transmits the encoded PCM audio data to the wireless audio module 320 through the first interface.
  • the main processing module 310 may employ lossless audio encoding.
  • the main processing module 310 may use free lossless audio codec (FLAC).
  • FLAC free lossless audio codec
  • the main processing module 310 may employ lossy audio encoding.
  • lossy audio encoding For example, sub-band coding (SBC) and advanced audio coding (AAC) can be used.
  • SBC sub-band coding
  • AAC advanced audio coding
  • the initial PCM audio data is PCM decoded audio data. After PCM decoding, several frames of PCM audio data can be obtained.
  • the initial PCM audio data may be one of several PCM audio data in one frame of data. One frame is, for example, 10 milliseconds.
  • the number of bits of the initial PCM audio data can represent the sampling depth.
  • N-bit initial PCM audio data can represent a sampling depth of N.
  • N can have multiple values, such as 32, 24, 16, etc. The larger N is, the more detailed the recording of sound intensity is.
  • the initial PCM audio data may include N pieces of data from low bits to high bits.
  • N the initial PCM audio data can be 24 binary data arranged from the lowest bit to the highest bit.
  • the initial PCM audio data can be converted to N-bit data based on the principle of memory alignment.
  • the initial PCM audio data can be converted into multiple binary data groups based on the number of bits.
  • the plurality of binary data groups may include a first binary data group and a second binary data group.
  • the first encoded packet may be generated by audio encoding the first binary data group, that is, the audio data in the first encoded packet may be the data in the first binary data group.
  • the second encoded packet may be generated by audio encoding the second binary data group, that is to say, the audio data in the second encoded packet may be the data in the second binary data group.
  • the conversion of the initial PCM audio data can be split based on the number of bits, or it can be extracted based on the number of bits.
  • the initial PCM audio data is split into multiple binary data groups.
  • the initial PCM audio data is formed as N-bit binary data, which can be split, and the high-M bit data of the N-bit data is used to form the first binary data group.
  • the second binary data group can also be formed with the lower L-bit data among the N-bit data.
  • specific bits or combinations thereof in the initial PCM audio data may be extracted separately to form multiple binary data groups. For example, odd-numbered bits or even-numbered bits in the initial PCM audio data (N-bit binary data) are respectively extracted to form a first binary data group or a second binary data group respectively.
  • the first set of binary data may correspond to a first plurality of binary bits of the initial PCM audio data.
  • the second set of binary data may correspond to a second batch of one or more binary bits of the original PCM audio data.
  • the first binary data group may include a continuously distributed plurality of binary bits in the initial PCM audio data. In other embodiments, the first binary data group may include a plurality of discretely distributed binary bits in the original PCM audio data. For example, the first binary data group may include odd or even bits in the original PCM audio data.
  • the second binary data group may include one binary bit in the initial PCM audio data, or a plurality of continuously distributed binary bits. In other embodiments, the second binary data group may include a plurality of discretely distributed binary bits in the original PCM audio data. For example, the second binary data group may include even or odd bits in the original PCM audio data.
  • the number of bits corresponding to the first binary data group and the second binary data group may be different.
  • the number of bits in the first binary data set may be greater than the number of bits in the second binary data set.
  • the first binary data group may correspond to the upper 24 bits of the initial PCM audio data
  • the second binary data group may correspond to the lower 8 bits of the initial PCM audio data.
  • the number of bits corresponding to the first binary data group may be smaller than the number of bits corresponding to the second binary data group.
  • the first binary data group may correspond to the upper 24 bits of the initial PCM audio data
  • the second binary data group may correspond to the lower 48 bits of the initial PCM audio data.
  • the number of bits corresponding to the first binary data group and the second binary data group may be the same.
  • the first binary data group may correspond to the upper 16 bits of the initial PCM audio data
  • the second binary data group may correspond to the lower 16 bits of the initial PCM audio data.
  • the first binary data group may correspond to the upper 8 bits of the initial PCM audio data
  • the second binary data group may correspond to the middle 8 bits or lower 8 bits of the initial PCM audio data.
  • the first binary data group may be high-bit data split from the initial PCM audio data; the second binary data group may be low-bit data split from the initial PCM audio data.
  • the second binary data group may also be the middle data and low data split from the initial PCM audio data.
  • the second binary data group may include a mid-bit group and/or a low-bit group.
  • the mid-bit group is composed of mid-bit data split from the initial PCM audio data
  • the low-bit group is composed of low-bit data split from the initial PCM audio data. composition.
  • the first binary data group may be the upper 16-bit data split from the initial PCM audio data;
  • the second binary data group may include the middle bit group and the low bit group , the middle bit group consists of the middle 4-bit binary data separated from the initial PCM audio data, and the low-bit group consists of the low 4-bit binary data separated from the initial PCM audio data.
  • the receiving end usually includes a digital to analog converter (DAC).
  • DAC digital to analog converter
  • audio signals are stored and transmitted digitally, and DACs can convert digital signals into analog signals so that they can be recognized by the outside world.
  • the signal to noise ratio is related to the sampling depth of the PCM audio data.
  • Sampling depth can also be expressed as the number of bits of PCM audio data. The greater the sampling depth, the higher the SNR.
  • the SNR of 24-bit sampling depth can reach 144dB
  • the SNR of 16-bit sampling depth can reach 96dB
  • the SNR of the lower 8-bit sampling depth is 48dB
  • the SNR of the lower 4-bit sampling depth is 24dB.
  • the SNR index of a better DAC can generally only reach 110dB. From a digital domain perspective, the theoretical SNR of a digital signal with a 20-bit sampling depth is 120dB. Therefore, in some embodiments, for initial PCM audio data of more than 24 bits, the theoretical SNR of the PCM audio data of the upper 20 bits is higher than the actual SNR that the DAC can achieve, and the SNR index of the DAC can be reached, which is 4 bits or more lower. After the data passes through the DAC, it is basically hidden in the noise.
  • the number of bits in the first binary data group can be set to be close to, equal to, or greater than 20 bits, thereby ensuring successful transmission of the first encoded packet that has a great impact on sound quality.
  • the first encoded packet and the second encoded packet may also contain additional information.
  • the additional information may include, for example, identification information and/or verification information.
  • the identification information can be used to identify the position of the encoded packet in the code stream. In other words, the identification information can mark the time sequence of the encoded packets in the code stream. After the audio playback device groups packets based on this identification information, it can ensure that the order of the code streams is correct.
  • the first encoding packet and the second encoding packet mentioned above may contain the same identification information, and the audio playback device can find the encoding packet corresponding to the initial PCM audio data according to the identification information, and assemble the packets.
  • the identification information may be a time sequence code.
  • the time sequence code can time-mark the first encoded packet and the second encoded packet, so that the audio playback device can group the packets in time sequence.
  • the time sequence code may be a long integer variable ranging from 0 to 2 63 -1. When the variable of the time series code increases to 2 63 -1, the next value can be 0.
  • the sender can be identified by the time series value corresponding to each encoded packet. For example, if the time series value 100 is compiled into the time series bit corresponding to the encoding packet at that moment, the time series value corresponding to the next encoding packet will be 101.
  • the verification information can verify the audio data in the encoded packet so that the receiving end can determine the integrity of the encoded packet.
  • the verification information may be a check code. For example, you can perform XOR calculation on the next value starting from the first bit value of the encoded packet, and the obtained calculation result can be used as the check code.
  • the wireless audio module 320 receives the first encoding packet and the second encoding packet transmitted by the main processing module 310 through the first interface.
  • a buffer may be provided on the wireless audio module 320 side to cache data that needs to be received.
  • the wireless audio module 320 is configured to respectively transmit the first encoding packet and the second encoding packet through the wireless channel, so that the audio data is received by the wireless audio playback device.
  • the first interface may be an audio interface that supports synchronous data transmission.
  • the main processing module 310 transmits the initial PCM audio data to the wireless audio module 320 through the first interface.
  • the wireless audio module 320 converts the initial PCM audio data into a first binary data group and a second binary data group, and performs audio encoding on the first binary data group and the second binary data group respectively, Obtain the corresponding first encoding packet and second encoding packet.
  • the wireless audio module 320 transmits the first encoded packet and the second encoded packet through the wireless channel so that the audio playback device can receive them.
  • the wireless audio module 320 is a Bluetooth transmitting module.
  • the device 300 for transmitting audio data is a mobile phone.
  • the wireless audio playback device After receiving the audio signal, the wireless audio playback device performs audio decoding on the first encoded packet and the second encoded packet to obtain the first binary data group and the second binary data group of the PCM audio data. In some implementations, if the transmission of the first encoded packet fails, the transmitting module 320 receives the retransmission instruction and can retransmit the first encoded packet.
  • the compression rate using FLAC or other formats is about 70%. Then the bandwidth for one transmission is 6.45Mbps, and the bandwidth for one retransmission is 12.9Mbps. In some embodiments, if it is divided into 16 bits + 4 bits + 4 bits for transmission, retransmitting a 16-bit + 4-bit encoded packet requires up to 11.83Mbps bandwidth. In some embodiments, if the transmission is divided into 16 bits + 8 bits, retransmitting the first encoded packet corresponding to the 16-bit first binary data group only requires a maximum bandwidth of 9.68 Mbps.
  • the wireless audio playback device may adopt a variety of processing methods, such as compensating the second binary data group or ignoring the error. For example, target PCM audio data is generated based on the compensation data of the first binary data group and the second binary group.
  • the first binary data group corresponding to the first encoding packet may be the upper 16-bit binary data, and the bits of the second binary data group The number can be the lower 8 bits. If the first encoding packet is received but the second encoding packet is not received, compensation processing is performed on the second binary data group, and theoretically the SNR of the reframed audio signal can reach 96dB.
  • the first binary data group corresponding to the first encoding packet may be the upper 16-bit binary data.
  • the second encoding packet may include a middle-bit group and a low-bit group.
  • the middle-bit group may be the middle 4-bit binary data
  • the low-bit group may be the lower 4-bit binary data. If the first coded packet is received, the second coded packet corresponding to the mid-bit group is received, and the second coded packet corresponding to the low-bit group is not received, the low-bit group of the second binary data group can be compensated, so theoretically The SNR of the reframed audio signal PCB can reach 120dB.
  • the initial PCM audio data is divided into multiple binary data groups for transmission, which reduces the number of data bits corresponding to a single encoding packet and can be retransmitted at a smaller granularity, thereby reducing the amount of retransmitted data and the transmission time. extension. This helps the receiver adopt flexible processing solutions when packets are lost and reduce bandwidth usage caused by retransmissions.
  • FIG. 4 is a schematic diagram of a possible basic flow of audio data processing in the device of FIG. 3 .
  • the first interface in Figure 4 may be an audio interface that supports asynchronous transmission of data, such as SLIMbus, Soundwire or SPI.
  • the initial PCM audio data is encoded and decoded in the main processing module.
  • the main processing module can be an ADSP module, and the wireless audio module 420 can be a BT module.
  • the device for transmitting audio data may include a main processing module 410 and a wireless audio module 420 as described above.
  • the main processing module 410 may convert the initial PCM audio data into a first binary data group and a second binary data group, and perform audio encoding on the first binary data group and the second binary data group respectively, Obtain the corresponding first encoding packet and second encoding packet.
  • the first binary data group may be high-bit data split from the initial PCM audio data; the second binary data group may be low-bit data split from the initial PCM audio data.
  • the second binary data group may also be mid-bit data or low-bit data split from the initial PCM audio data.
  • the second binary data group may include a mid-bit group or a low-bit group.
  • the mid-bit group is composed of mid-bit data split from the initial PCM audio data
  • the low-bit group is composed of low-bit data split from the initial PCM audio data.
  • the first binary data group may be the upper 16-bit data split from the initial PCM audio data;
  • the second binary data group may include the middle bit group and the low bit group , the middle bit group consists of the middle 4-bit binary data separated from the initial PCM audio data, and the low-bit group consists of the low 4-bit data separated from the initial PCM audio data.
  • a buffer 421 may also be provided on the wireless audio module 420 side to cache data that needs to be received.
  • the main processing module 410 transmits the packet-coded audio data to the BT transmitting module 422 of the wireless audio module 420 through the first interface.
  • the BT transmitting module 422 transmits the first encoding packet and the second encoding packet respectively through the wireless channel, so that the audio data is received by the wireless audio playback device 430.
  • FIG. 5 is a schematic diagram of another possible basic flow of audio data processing in the device of FIG. 3 .
  • the first interface in Figure 5 may support an audio interface for synchronous data transmission, such as an I2S interface.
  • the device for transmitting audio data may include an ADSP module 510 and a wireless audio module 520.
  • ADSP module 510 is the main processing module and can read audio source files.
  • a buffer is provided on the ADSP module 510 side to cache the PCM audio data that needs to be transmitted.
  • the wireless audio module 520 is a BT module and may include an encoding module 522 and a transmitting module 523.
  • the BT side may also be provided with a buffer 521 to cache PCM audio data that needs to be received.
  • the ADSP module 510 transmits the initial PCM audio data to the buffer 521 on the wireless audio module 520 side through the first interface.
  • the encoding module 522 in the wireless audio module 520 converts the initial PCM audio data into a first binary data group and a second binary data group, and processes the first binary data group and the second binary data group respectively. Perform audio encoding to obtain the corresponding first encoding packet and second encoding packet.
  • the BT transmitting module 523 transmits the first encoding packet and the second encoding packet respectively through the wireless channel, so that the audio data is received by the wireless audio playback device 530.
  • the device embodiment of the present application is described in detail above with reference to FIGS. 1 to 5 , and the method embodiment of the present application is described in detail below with reference to FIG. 6 . It should be understood that the description of the method embodiments corresponds to the description of the device embodiments. Therefore, the parts not described in detail can be referred to the previous device embodiments.
  • Figure 6 is a schematic flow chart of a method for transmitting audio data provided by an embodiment of the present application.
  • the method of Figure 6 can be applied to the device for transmitting audio data mentioned in any of the embodiments of Figures 3 to 5.
  • the device for transmitting audio data is provided with a main processing module and a wireless audio module, and the main processing module and the wireless audio module transmit audio data through the first interface.
  • the method in Figure 6 includes steps S610 to S630, and these steps will be described in detail below.
  • step S610 the initial PCM audio data is converted into a first binary data group and a second binary data group.
  • step S620 audio encoding is performed on the first binary data group and the second binary data group respectively to obtain corresponding first encoded packets and second encoded packets.
  • step S630 the wireless audio module is used to transmit the first encoded packet and the second encoded packet respectively, so as to be received by the wireless audio playback device.
  • the first interface is an audio interface that only supports synchronous data transmission
  • both the encoding module and the transmitting module are integrated in the wireless audio module.
  • the first interface is an I2S interface.
  • the encoding module is integrated in the main processing module, and the transmitting module is integrated in the wireless audio module.
  • the first interface may be an interface such as SLIMbus, Soundwire or SPI.
  • the transmitting module is a Bluetooth transmitting module.
  • the device for transmitting audio data is a mobile phone.
  • FIG. 7 is a schematic structural diagram of an electronic device provided by an embodiment of the present application. As shown in FIG. 7 , the electronic device may include a memory 710 and a processor 720 .
  • Memory 710 is used to store code.
  • the processor 720 is configured to load and execute the code in the memory 710 to implement any method of transmitting audio data as described above.
  • FIG. 8 is a schematic structural diagram of a chip provided by an embodiment of the present application.
  • the chip 800 may include a processor 810 configured to perform any method of transmitting audio data as described above.
  • the processor 800 is configured to: convert the initial PCM audio data into a first binary data group and a second binary data group; and perform separate processing on the first binary data group and the second binary data group. Audio encoding to obtain the corresponding first encoding packet and second encoding packet.
  • An embodiment of the present application also provides an intelligent terminal, including the chip 800 as described above.
  • the disclosed systems, devices and methods can be implemented in other ways.
  • the device embodiments described above are only illustrative.
  • the division of the units is only a logical function division. In actual implementation, there may be other division methods.
  • multiple units or components may be combined or can be integrated into another system, or some features can be ignored, or not implemented.
  • the coupling or direct coupling or communication connection between each other shown or discussed may be through some interfaces, and the indirect coupling or communication connection of the devices or units may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in various embodiments of the present application may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.

Abstract

提供了一种传输音频数据的装置、方法以及电子设备。该装置包括:主处理模块,配置成:将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;以及对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;无线音频模块,与主处理模块通过第一接口相连,配置成通过无线信道分别传输第一编码包和第二编码包。本申请实施例将初始PCM音频数据转换成多个二进制数据组,并分别进行传输,有助于降低重传的数据量,从而降低传输音频数据所需的带宽。

Description

传输音频数据的装置、方法以及电子设备
本申请要求于2022年06月30日提交中国专利局、申请号为202210762229.3、申请名称为“传输音频数据的装置、方法以及电子设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请实施例涉及音频技术领域,并且更为具体地,涉及一种传输音频数据的装置、方法以及电子设备。
背景技术
近年来随着无线通信技术的发展,手机、无线耳机等电子设备迅速普及,人们越来越多的应用蓝牙等无线方式来传播和分享音频。在无线传输时,由于种种干扰因素,如果音频播放设备收到的数据有错误时,需要进行重传。重传一般会严重占用带宽,重传次数过多时,无线带宽满足不要求,就会出现音频卡顿现象。
发明内容
本申请实施例提供了一种传输音频数据的装置、方法以及电子设备,下面对本申请实施例的各个方面进行介绍。
第一方面,提供一种传输音频数据的装置,包括:主处理模块,配置成:将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;以及对所述第一二进制数据组和所述第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;无线音频模块,与所述主处理模块通过第一接口相连,配置成通过无线信道分别传输所述第一编码包和所述第二编码包。
第二方面,提供一种传输音频数据的装置,包括:主处理模块,配置成生成初始PCM数据;无线音频模块,与所述主处理模块通过第一接口相连,配置成:将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;以及对所述第一二进制数据组和所述第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;以及通过无线信道分别传输所述第一编码包和所述第二编码包。
第三方面,提供一种传输音频数据的方法,所述方法应用于传输音频数据的装置,所述装置设置有主处理模块和无线音频模块,且所述主处理模块和所述无线音频模块通过第一接口相连,所述方法包括:将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;对所述第一二进制数据组和所述第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;以及利用所述发射模块分别传输所述第一编码包和所述第二编码包。
第四方面,提供一种电子设备,包括:存储器,用于存储代码;处理器,用于加载并执行所述存储器的代码,以实现如第三方面所述的传输音频数据的方法。
第五方面,提供一种芯片,包括处理器,所述处理器配置成执行如第三方面所述的方法。
第六方面,提供一种智能终端,包括如第五方面所述的芯片。
本申请实施例将初始PCM音频数据转换成多个二进制数据组,并分别进行传输,有助于降低重传数据量和音频数据传输所需的无线带宽,从而避免音频卡顿现象。
附图说明
图1是一种蓝牙音频数据处理的基本流程的示意图。
图2是另一种蓝牙音频数据处理的基本流程的示意图。
图3是本申请实施例提供的传输音频数据的装置的结构示意图。
图4是图3装置的一种可能的音频数据处理的基本流程的示意图。
图5是图3装置的另一种可能的音频数据处理的基本流程的示意图。
图6是本申请实施例提供的传输音频数据的方法的流程示意图。
图7是本申请实施例提供的电子设备的结构示意图。
图8是本申请实施例提供的芯片的结构示意图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅是本申请的一部分实施例,而不是全部的实施例。
近年来,随着无线通讯技术的发展,无线通信技术的应用越来越广泛,手机、无线耳机等电子设备迅速普及。需要说明的是,本申请实施例提及的电子设备可以指能够传输无线音频信号的任意类型的电子设备。该电子设备例如可以是手机、无线耳机。下面结合手机和无线耳机的数据传输流程进行举例说明。
无线耳机是中间的线被电波代替,发射端的音频信号通过电波发送到接收端的耳机中。无线耳机和有线耳机相比,使用比较方便,不受有线的束缚和限制,因此得到了广泛的应用。随着无线耳机,尤其是真无线立体声(true wireless stereo,TWS)耳机的爆发式增长,消费者在生活和工作中使用无线耳机已非常常见,比如听音乐,打电话。
同时,消费者对音频的播放质量的要求越来越高。例如,很多消费者希望通过无线音频播放设备收听高质量音乐,或无损(lossless)音乐。
由于TWS耳机延时性能的提高,在游戏中也逐步取代了有线耳机。伴随着TWS耳机降噪功能的发展,在地铁、公交车或机场等这些嘈杂场景中,使用TWS降噪耳机比普通有线耳机有更好的听觉体验。因此,人们越来越多的应用蓝牙等无线方式来传播和分享音频。
图1是一种蓝牙音频数据处理的基本流程的示意图。如图1所示,手机可以包括音频数字信号处理(audio digital signal processing,ADSP)模块110和蓝牙(blue tooth,BT)发射模块120。
ADSP模块110可以对音频的源数据进行编码或译码。编译码技术能有效减少数字存储占用的空间,提高系统的运行效率。ADSP模块110侧还设有缓冲器,以便对需要传输的数据进行缓存。
BT发射模块120可以包括BT收发器122,BT收发器122用于将音频数字信号转换成模拟信号,并通过无线信道将模拟信号发射出去。此外,BT侧也可以设有缓冲器121,以便对需要接收的数据进行缓存。
音频信号的数据处理过程大致分为以下步骤:
步骤一:音频处理模块110对音频的源数据进行编码,音频的源数据格式通常采用脉冲编码调制(pulse code modulation,PCM)。音频信号的源数据多为16位、24位或32位数据。编码后的数据暂存到音频处理模块110侧的缓冲器中。
步骤二:通过数据总线将音频处理模块110侧缓冲区的数据传输到BT侧的缓冲区121中,BT收发器122将音频数字信号转换成模拟信号,并无线传输给TWS耳机130。数据总线的方式可以有多种,比如集成电路内置音频总线(inter-IC sound,I2S)、低功耗芯片间串行媒体总线(serial low-power inter-chip media bus,SLIMbus)、Soundwire或串行外设接口(serial peripheral interface,SPI)等。
步骤三:TWS耳机130接收模拟音频信号,经过相应的模数转换处理,从听筒播放出声音。
值得提出的是,I2S接口天然用于传输同步数据,如果音频源文件经过编码后,转变成了异步数据,占用空间会变少。在耳机端解码时有可能需要使用私有协议来解码,通常利用I2S接口传输音频编码前的数据。
图2是另一种蓝牙音频数据处理的基本流程的示意图。如图2所示,手机可以包括音频ADSP模块210和BT发射模块220。
BT发射模块220可以包括控制器222和BT收发器223。控制器222可以对音频数据进行编码或译码。此外,BT侧也可以设有缓冲器221,以便对需要接收的数据进行缓存。
音频信号的数据处理过程大致如下:
音频处理模块210将音频的源数据暂存到音频处理模块210侧的缓冲器中。通过数据总线将音频处理模块210侧缓冲区的数据传输到BT侧的缓冲区221中,由BT控制器222对音频数据进行编码,而后BT收发器223将音频数字信号转换成模拟信号,发射给TWS耳机230接收。
蓝牙是一种低成本大容量的短距离无线通信规范。蓝牙技术是全球开放的,在全球范围内具有很好的兼容性,通常使用2.4~2.485GHz的无线电波。
在上述无线传输时,由于种种因素,比如其它频段(如Wi-Fi 2.4GHz)、其它蓝牙设备的干扰,以及设备本身的射频性能的限制,如果蓝牙耳机侧收到的数据丢失或出错,那么现有处理方案通常需要重传。无线传输中每个数据包的长度可以有5ms或者7.5ms时长,如果重传会造成严重占用带宽的结果。如果干扰严重,重传次数过多时,蓝牙的无线带宽满足不了此种场景,或者不能实现即时通讯。
需要说明的是,上文提及的蓝牙音频数据传输失败仅是一个示例,本申请实施例可应用于具有无线音频数据传输失败的任意类型的场景。
因此,如何开发一种应对无线音频数据传输失败的方案是需要解决的问题。
针对上述问题,本申请实施例提出一种传输音频数据的装置,下面对本申请实施例进行详细描述。
图3是本申请实施例提供的一种传输音频数据的装置的结构示意图。该传输音频数据的装置300可以包括主处理模块310和无线音频模块320。
主处理模块310例如可以为应用处理器(application processor,AP),或者可以是ADSP模块。在一些实施例中,主处理模块310可配置成对初始PCM音频数据进行编解码。通常主处理模块的处理性能较好,通过主处理模块310执行音频编解码工作,可降低无线音频模块320的工作负担。
无线音频模块320例如可以为WiFi模块、4G通信模块或蓝牙模块等。无线音频模块320可包括蓝牙控制器,蓝牙控制器可配置成执行音频包的传输。视情况而定,蓝牙控制器还可配置成执行音频编解码。在蓝牙控制器独立于主处理器模块的情况下,通过蓝牙控制器执行音频编解码或其中的部分工作/步骤,可降低主处理器模块的工作负担。
在本申请一些实施例中,主处理模块可形成为第一芯片,无线音频模块可形成为独立的第二芯片。两块芯片共同设置于本申请实施例提供的传输音频数据的装置上,协作执行音频编解码工作。
主处理模块310通过第一接口和无线音频模块320之间传输音频数据。第一接口例如可以是总线接口。
在一些实施例中,第一接口可以为支持数据异步传输的音频接口,例如为SLIMbus,Soundwire或SPI。
主处理模块310将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组,并对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包。主处理模块310通过第一接口将编码后的PCM音频数据传输至无线音频模块320中。
在一些实施例中,主处理模块310可以采用无损音频编码。例如,主处理模块310可以采用自由无损音频编码(free lossless audio codec,FLAC)。
在一些实施例中,主处理模块310可以采用有损音频编码。例如,可以采用子带编码(sub-band coding,SBC)、高级音频编码(advanced audio coding,AAC)。
初始PCM音频数据为PCM解码后的音频数据。PCM解码后可以得到若干帧的PCM音频数据。初始PCM音频数据可以是一帧数据中若干PCM音频数据中的一个。一帧例如是10毫秒。
初始PCM音频数据的位数可以表示采样深度。例如N位初始PCM音频数据可以表示其采样深度为N。N可以有多种取值,例如32、24、16等。N越大,代表对声音强度的记录就越精细。
初始PCM音频数据的位数用N表示时,该初始PCM音频数据可以包括从低位到高位的N个数据。例如,N为24时,初始PCM音频数据可以是24个从最低位到最高位排列的二进制数据。
初始PCM音频数据可以基于内存对齐的原则,对N位数据进行转换。初始PCM音频数据可以基于位数转换为多个二进制数据组。该多个二进制数据组可以包括第一 二进制数据组和第二二进制数据组。第一编码包可以通过对第一二进制数据组进行音频编码生成,也就是说,第一编码包中的音频数据可以是第一二进制数据组中的数据。第二编码包可以通过对第二二进制数据组进行音频编码生成,也就是说,第二编码包中的音频数据可以是第二二进制数据组中的数据。
对初始PCM音频数据进行转换可以是基于位数进行拆分,也可以是基于位数进行提取。在一些实施例中,将初始PCM音频数据拆分为多个二进制数据组。例如,初始PCM音频数据形成为N位二进制数据,可对其进行拆分,以N位数据中的高M位数据形成第一二进制数据组。又如,还可以以N位数据中的低L位数据形成第二二进制数据组。在另一些实施例中,可分别提取初始PCM音频数据中的特定位或其组合,以形成多个二进制数据组。例如,分别提取初始PCM音频数据(N位二进制数据)中的奇数位或偶数位,以分别形成第一二进制数据组或第二二进制数据组。
第一二进制数据组可以对应于初始PCM音频数据的第一批多个二进制位。第二二进制数据组可以对应于初始PCM音频数据的第二批一个或多个二进制位。
在一些实施例中,第一二进制数据组可以包括初始PCM音频数据中的连续分布的多个二进制位。在另一些实施例中,第一二进制数据组可以包括初始PCM音频数据中的离散分布的多个二进制位。例如,第一二进制数据组可以包括初始PCM音频数据中的奇数位或偶数位。
在一些实施例中,第二二进制数据组可以包括初始PCM音频数据中的一个二进制位,或连续分布的多个二进制位。在另一些实施例中,第二二进制数据组可以包括初始PCM音频数据中的离散分布的多个二进制位。例如,第二二进制数据组可以包括初始PCM音频数据中的偶数位或奇数位。
在一些实施例中,第一二进制数据组和第二二进制数据组对应的位数可以不同。
在一些实施例中,第一二进制数据组的位数可以大于第二二进制数据组的位数。例如,针对32位初始PCM音频数据,第一二进制数据组可以对应初始PCM音频数据的高24位,第二二进制数据组可以对应初始PCM音频数据的低8位。
在一些实施例中,第一二进制数据组对应的位数可以小于第二二进制数据组对应的位数。例如,针对64位初始PCM音频数据,第一二进制数据组可以对应初始PCM音频数据的高24位,第二二进制数据组可以对应初始PCM音频数据的低48位。
在一些实施例中,第一二进制数据组和第二二进制数据组对应的位数可以相同。例如,针对32位初始PCM音频数据,第一二进制数据组可以对应初始PCM音频数据的高16位,第二二进制数据组可以对应初始PCM音频数据的低16位。又如,针 对32位初始PCM音频数据,第一二进制数据组可以对应初始PCM音频数据的高8位,第二二进制数据组可以对应初始PCM音频数据的中8位或低8位。
下面分别以32位和24位的初始PCM音频数据为例,给出初始PCM音频数据的几个具体的拆分示例。在下面的示例中,第一二进制数据组可以是从初始PCM音频数据中拆分出的高位数据;第二二进制数据组可以是初始PCM音频数据中拆分出的低位数据。第二二进制数据组也可以是初始PCM音频数据中拆分出的中位数据和低位数据。第二二进制数据组可以包括中位组和/或低位组,中位组由初始PCM音频数据中拆分出的中位数据组成,低位组由初始PCM音频数据中拆分出的低位数据组成。
对于32位的初始PCM音频数据,可以有以下几种拆分类型:24位+8位、24位+4位+4位、16位+16位、16位+8位+8位、16位+8位+4位+4位、16位+4位+4位+4位、8位+8位+8位等。
对于24位的初始PCM音频数据,可以有以下几种拆分类型:16位+8位、16位+4位+4位,8位+8位+8位,8位+8位+4位+4位等。对于16位+4位+4位类型,第一二进制数据组可以是从初始PCM音频数据中拆分出的高16位数据;第二二进制数据组可以包括中位组和低位组,中位组由初始PCM音频数据中拆分出的中4位二进制数据组成,低位组由初始PCM音频数据中拆分出的低4位二进制数据组成。
接收端通常包括数模转换器(digital to analog converter,DAC)。在很多数字系统中(例如计算机),音频信号以数字方式存储和传输,而DAC可以将数字信号转换为模拟信号,从而使得它们能够被外界识别。
DAC对PCM音频数据进行数模转换时,信噪比(signal to noise ratio,SNR)与PCM音频数据的采样深度有关。采样深度,也可以表示为PCM音频数据的位数,采样深度越大,SNR越高。理论上,24位采样深度的SNR可以达到144dB,16位采样深度的SNR可以达到96dB,低8位采样深度的SNR是48dB,低4位采样深度的SNR是24dB。
但是,较好的DAC的SNR指标一般只能达到110dB,而从数字域角度看,20位采样深度的数字信号理论SNR为120dB。因此,在一些实施例中,对于24位以上的初始PCM音频数据,高20位的PCM音频数据的理论SNR比DAC能达到的实际SNR要高,可以达到DAC的SNR指标,低4位或以上数据经过DAC以后基本上都隐藏在噪声之中了。
因此,在一些实施例中,可以将第一二进制数据组的位数设置为接近、等于或大于20位,从而确保对音质影响大的第一编码包传输成功。
除了音频数据之外,第一编码包和第二编码包还可以包含附加信息。该附加信息例如可以包括例如标识信息和/或校验信息。
标识信息可用于标识编码包在码流中的位置。也就是说,标识信息可以对编码包在码流中的时间先后进行标记。音频播放设备根据该标识信息进行组包后,可以保证码流的顺序正确。前文提到的第一编码包和第二编码包可以包含相同的标识信息,音频播放设备可以根据标识信息找到初始PCM音频数据对应的编码包,并进行组包。
在一些实施例中,标识信息可以是时间序列码。时间序列码可以对第一编码包和第二编码包进行时间标记,以便音频播放设备按时间序列进行组包。在一些实施例中,时间序列码可以是一个0到2 63-1的长整形的变量。当时间序列码的变量增加到2 63-1时,下一个值可以为0。发送端可以按每个编码包对应的时间序列值进行标识。例如将时间序列值100编入该时刻对应编码包的时间序列位,下一个编码包对应的时间序列值就是101。
校验信息可以对编码包内的音频数据进行校验,以便接收端确定编码包的完整性。在一些实施例中,检验信息可以是校验码。例如,可以从编码包第1位比特值开始对下一个值进行异或计算,得到的计算结果即可作为校验码。
无线音频模块320通过第一接口接收主处理模块310传输的第一编码包和第二编码包。无线音频模块320侧可以设有缓冲器,以便对需要接收的数据进行缓存。无线音频模块320配置成通过无线信道分别传输第一编码包和第二编码包,以便音频数据被无线音频播放设备接收。
在一些实施例中,第一接口可以为支持数据同步传输的音频接口。例如为I2S接口。主处理模块310通过第一接口将初始PCM音频数据传输至无线音频模块320。由无线音频模块320将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组,并对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包。无线音频模块320通过无线信道将第一编码包和第二编码包发射出去,以便音频播放装置接收。
可选地,无线音频模块320为蓝牙发射模块。
可选地,传输音频数据的装置300为手机。
无线音频播放设备在接收到音频信号后,对第一编码包和第二编码包进行音频解码,得到PCM音频数据的第一二进制数据组和第二二进制数据组。在一些实现方式中,如果第一编码包的传输失败,发射模块320接收到重传指令,可以重传第一编码包。
对于24位的PCM音频数据,假定采样频率为192KHz,双通道传输,采用FLAC或其它格式压缩率约为70%。那么传输一次带宽为6.45Mbps,重传一次需要12.9Mbps带宽。在一些实施例中,如果分成16位+4位+4位传输,重传一次16位+4位的编码包最多需要11.83Mbps带宽。在一些实施例中,如果分成16位+8位传输,重传一次16位的第一二进制数据组对应的第一编码包最多只需要9.68Mbps带宽。
在一些实现方式中,如果接收到第一编码包且未接收到第二编码包,无线音频播放设备可以采用多种处理方式,如对第二二进制数据组进行补偿或忽略错误。如根据第一二进制数据组和第二二进制组的补偿数据,生成目标PCM音频数据。
在一些实施例中,如果24位初始PCM音频数据的分为两组发送,第一编码包对应的第一二进制数据组可以是高16位二进制数据,第二二进制数据组的位数可以是低8位。如果接收到第一编码包且未接收到第二编码包,则对第二二进制数据组进行补偿处理,理论上重新组帧后的音频信号的SNR可以达到96dB。
在一些实施例中,如果24位初始PCM音频数据的分为三组,第一编码包对应的第一二进制数据组可以是高16位二进制数据。第二编码包可以包括中位组和低位组,中位组可以为中间4位二进制数据,低位组为低4位二进制数据。如果接收到第一编码包,接收到中位组对应的第二编码包且未接收到低位组对应的第二编码包,可以对第二二进制数据组的低位组进行补偿处理,这样理论上重新组帧后的音频信号PCB的SNR可以达到120dB。如果接收到第一编码包且未接收到第二二进制数据组的中位组对应的第二编码包,可以对第二二进制数据组的中位组进行补偿处理,这样,理论上重新组帧后的音频信号PCB的SNR可以达到96dB。
本申请实施例中将初始PCM音频数据分成多个二进制数据组发送,减小了单个编码包对应的数据位数,可以以更小的粒度进行重传,从而可以降低重传数据量和传输时延。有助于接收方在丢包时采用灵活的处理方案,减少重传带来的带宽占用。
图4是图3装置的一种可能的音频数据处理的基本流程的示意图。图4中的第一接口可以为支持数据异步传输的音频接口,例如为SLIMbus、Soundwire或SPI。在主处理模块中对初始PCM音频数据进行编解码处理,主处理模块可以为ADSP模块,无线音频模块420可以为BT模块。如图4所示,传输音频数据的装置可以包括如上文所述的主处理模块410和无线音频模块420。
主处理模块410可以将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组,并对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包。
第一二进制数据组可以是从初始PCM音频数据中拆分出的高位数据;第二二进制数据组可以是初始PCM音频数据中拆分出的低位数据。第二二进制数据组也可以是初始PCM音频数据中拆分出的中位数据或低位数据。第二二进制数据组可以包括中位组或低位组,中位组由初始PCM音频数据中拆分出的中位数据组成,低位组由初始PCM音频数据中拆分出的低位数据组成。
例如,对于24位的初始PCM音频数据,可以有以下几种分组类型:16位+8位、16位+4位+4位。对于16位+4位+4位类型,第一二进制数据组可以是从初始PCM音频数据中拆分出的高16位数据;第二二进制数据组可以包括中位组和低位组,中位组由初始PCM音频数据中拆分出的中4位二进制数据组成,低位组由初始PCM音频数据中拆分出的低4位数据组成。
无线音频模块420侧也可以设有缓冲器421,以便对需要接收的数据进行缓存。主处理模块410通过第一接口将分组编码处理后的音频数据传输至无线音频模块420的BT发射模块422中。BT发射模块422通过无线信道分别传输第一编码包和第二编码包,以便音频数据被无线音频播放设备430接收。
图5是图3装置的另一种可能的音频数据处理的基本流程的示意图。图5中的第一接口可以支持数据同步传输的音频接口,例如为I2S接口。如图5所示,传输音频数据的装置可以包括ADSP模块510和无线音频模块520。
ADSP模块510为主处理模块,可以读取音源文件。ADSP模块510侧设有缓冲器,以便对需要传输的PCM音频数据进行缓存。
无线音频模块520为BT模块,可以包括编码模块522和发射模块523。BT侧也可以设有缓冲器521,以便对需要接收的PCM音频数据进行缓存。
ADSP模块510通过第一接口将初始PCM音频数据传输至无线音频模块520侧的缓冲器521中。无线音频模块520中的编码模块522将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组,并对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包。
BT发射模块523通过无线信道分别传输第一编码包和第二编码包,以便音频数据被无线音频播放设备530接收。
上文结合图1-图5,详细描述了本申请的装置实施例,下面结合图6,详细描述本申请的方法实施例。应理解,方法实施例的描述与装置实施例的描述相互对应,因此,未详细描述的部分可以参见前面装置实施例。
图6是本申请实施例提供的一种传输音频数据的方法的示意性流程图。图6的方 法可应用于前文图3-图5任一实施例提及的传输音频数据的装置。该传输音频数据的装置内部设置有主处理模块和无线音频模块,且主处理模块和无线音频模块通过第一接口传输音频数据。
图6的方法包括步骤S610至步骤S630,下面对这些步骤进行详细描述。
在步骤S610中,将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组。
在步骤S620中,对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包。
在步骤S630中,利用无线音频模块分别传输第一编码包和第二编码包,以便被无线音频播放设备接收。
可选地,如果第一接口为仅支持数据同步传输的音频接口,则编码模块和发射模块均集成在无线音频模块中。例如,第一接口为I2S接口。
可选地,如果第一接口为支持数据异步传输的音频接口,则编码模块集成在主处理模块中,且发射模块集成在无线音频模块中。例如,第一接口可以为SLIMbus、Soundwire或SPI等接口。
可选地,发射模块为蓝牙发射模块。
可选地,传输音频数据的装置为手机。
图7是本申请实施例提供的一种电子设备的结构示意图。如图7所示,该电子设备可以包括存储器710和处理器720。
存储器710用于存储代码。
处理器720用于加载并执行存储器710的代码,以实现如前文描述的任一种传输音频数据的方法。
图8是本申请实施例提供的一种芯片的结构示意图。如图8所示,芯片800可以包括处理器810,处理器810配置成执行如前文描述的任一种传输音频数据的方法。例如,处理器800配置成:将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;以及对第一二进制数据组和第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包。
本申请实施例还提供一种智能终端,包括如前文描述的芯片800。
应理解,在本申请的各种实施例中,“第一”、“第二”等是用于区别不同的对象,而不是用于描述特定顺序,上述各过程的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成 任何限定。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
在本申请所提供的几个实施例中,应该理解到,当称某一部分与另一部分“连接”或“相连”时,其意味着该部分不仅可以“直接连接”,而且也可以“电连接”,同时另一个元件介入其中。另外,术语“连接”也意指该部分“物理地连接”以及“无线地连接”。另外,当称某一部分“包含”某一元件时,除非另行加以陈述,否则,其意味着该某一部分可以包括另一元件,而不是排除所述另一个元件。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。
以上所述,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以所述权利要求的保护范围为准。

Claims (16)

  1. 一种传输音频数据的装置,其特征在于,包括:
    主处理模块,配置成:
    将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;以及
    对所述第一二进制数据组和所述第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;
    无线音频模块,与所述主处理模块通过第一接口相连,配置成通过无线信道分别传输所述第一编码包和所述第二编码包。
  2. 根据权利要求1所述的装置,其特征在于,所述第一接口为支持数据异步传输的音频接口。
  3. 根据权利要求1所述的装置,其特征在于,所述第一接口为SLIMbus,Soundwire或SPI。
  4. 根据权利要求1-3中任一项所述的装置,其特征在于,所述主处理模块为设置于所述装置上的第一芯片,所述无线音频模块为设置于所述装置上的第二芯片。
  5. 一种传输音频数据的装置,其特征在于,包括:
    主处理模块,配置成生成初始PCM数据;
    无线音频模块,与所述主处理模块通过第一接口相连,配置成:
    将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;以及
    对所述第一二进制数据组和所述第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;以及
    通过无线信道分别传输所述第一编码包和所述第二编码包。
  6. 根据权利要求5所述的装置,其特征在于,所述第一接口为支持数据同步传输的音频接口。
  7. 根据权利要求6所述的装置,其特征在于,所述第一接口为I2S接口。
  8. 根据权利要求5-7中任一项所述的装置,其特征在于,所述主处理模块为设置于所述装置上的第一芯片,所述无线音频模块为设置于所述装置上的第二芯片。
  9. 一种传输音频数据的方法,其特征在于,所述方法应用于传输音频数据的装置,所述装置设置有主处理模块和无线音频模块,且所述主处理模块和所述无线音频模块通过第一接口相连,
    所述方法包括:
    将初始PCM音频数据转换为第一二进制数据组和第二二进制数据组;
    对所述第一二进制数据组和所述第二二进制数据组分别进行音频编码,得到对应的第一编码包和第二编码包;以及
    利用所述无线音频模块分别传输所述第一编码包和所述第二编码包。
  10. 根据权利要求9所述的方法,其特征在于,所述第一接口为支持数据同步传输的音频接口,所述第一编码包和所述第二编码包由所述无线音频模块生成。
  11. 根据权利要求10所述的方法,其特征在于,所述第一接口为I2S接口。
  12. 根据权利要求9所述的方法,其特征在于,所述第一接口为支持数据异步传输的音频接口,所述第一编码包和所述第二编码包由所述主处理模块生成。
  13. 根据权利要求12所述的方法,其特征在于,所述第一接口为SLIMbus,Soundwire或SPI。
  14. 一种电子设备,其特征在于,包括:
    存储器,用于存储代码;
    处理器,用于加载并执行所述存储器的代码,以实现如权利要求9-13任一所述的传输音频数据的方法。
  15. 一种芯片,其特征在于,包括处理器,所述处理器配置成执行如权利要求9-13中任一项所述的方法。
  16. 一种智能终端,其特征在于,包括如权利要求15所述的芯片。
PCT/CN2022/143772 2022-06-30 2022-12-30 传输音频数据的装置、方法以及电子设备 WO2024001138A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210762229.3A CN115243238A (zh) 2022-06-30 2022-06-30 传输音频数据的装置、方法以及电子设备
CN202210762229.3 2022-06-30

Publications (1)

Publication Number Publication Date
WO2024001138A1 true WO2024001138A1 (zh) 2024-01-04

Family

ID=83670696

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/143772 WO2024001138A1 (zh) 2022-06-30 2022-12-30 传输音频数据的装置、方法以及电子设备

Country Status (2)

Country Link
CN (1) CN115243238A (zh)
WO (1) WO2024001138A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115243238A (zh) * 2022-06-30 2022-10-25 哲库科技(上海)有限公司 传输音频数据的装置、方法以及电子设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080114478A1 (en) * 2006-11-09 2008-05-15 David Wu Method and System for Multi-Channel PCM Audio Grouping in Hardware
CN105681819A (zh) * 2014-11-18 2016-06-15 中兴通讯股份有限公司 信号的发送、接收方法及装置、系统
CN115243238A (zh) * 2022-06-30 2022-10-25 哲库科技(上海)有限公司 传输音频数据的装置、方法以及电子设备
CN115278294A (zh) * 2022-06-22 2022-11-01 哲库科技(上海)有限公司 传输音频数据的方法、装置、电子设备及音频播放设备

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080114478A1 (en) * 2006-11-09 2008-05-15 David Wu Method and System for Multi-Channel PCM Audio Grouping in Hardware
CN105681819A (zh) * 2014-11-18 2016-06-15 中兴通讯股份有限公司 信号的发送、接收方法及装置、系统
CN115278294A (zh) * 2022-06-22 2022-11-01 哲库科技(上海)有限公司 传输音频数据的方法、装置、电子设备及音频播放设备
CN115243238A (zh) * 2022-06-30 2022-10-25 哲库科技(上海)有限公司 传输音频数据的装置、方法以及电子设备

Also Published As

Publication number Publication date
CN115243238A (zh) 2022-10-25

Similar Documents

Publication Publication Date Title
US10271152B1 (en) Synchronization of wireless headphones
US10932043B2 (en) Wireless audio system and method for wirelessly communicating audio information using the same
CN111294783B (zh) 音频数据传输的方法、装置、芯片和电子设备
WO2022033107A1 (zh) 蓝牙低功耗音频数据分组传输方法及装置
TW200901744A (en) Headset having wirelessly linked earpieces
US20030002473A1 (en) Enhanced cordless telephone platform using BLUETOOTH technology
WO2023246282A1 (zh) 传输音频数据的方法、装置、电子设备及音频播放设备
CN108718361B (zh) 一种音频文件播放方法及无线接听设备
CN112312369A (zh) 一种低延迟无线音频发送和接收方法及设备
WO2024001138A1 (zh) 传输音频数据的装置、方法以及电子设备
CN101789853B (zh) 处理重传请求方法、终端和对重传请求进行管理的方法
CN102832968A (zh) 一种利用音频接口在手机与设备之间进行通讯的方法
WO2021160040A1 (zh) 音频传输方法及电子设备
WO2021223243A1 (zh) 音频数据传输的方法、装置、芯片和电子设备
CN113365129A (zh) 蓝牙音频数据处理方法、发射器、接收器及收发设备
CA2952736C (en) Offset selection for error correction data
CN102845050A (zh) 在用于VoIP呼叫的带蓝牙耳机的移动通话器中绕过转码器的方法
CN110366752B (zh) 一种语音分频传输方法、源端、播放端、源端电路和播放端电路
CN100446519C (zh) 一种在通话过程中播放音乐的手机及其方法
US11545161B2 (en) Wireless communication device, and method and apparatus for processing voice data
US10778479B1 (en) Systems and methods for wireless transmission of audio information
CN115240692A (zh) 接收音频数据的方法、装置以及音频播放设备
CN220108215U (zh) 一种无线耳机
CN114584956B (zh) 蓝牙通信方法及系统、蓝牙接收端和发射端的通信方法
CN101741407B (zh) 一种具有收音功能的手机及其实现方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22949213

Country of ref document: EP

Kind code of ref document: A1