WO2024092478A1 - 音频处理设备和方法 - Google Patents

音频处理设备和方法 Download PDF

Info

Publication number
WO2024092478A1
WO2024092478A1 PCT/CN2022/128860 CN2022128860W WO2024092478A1 WO 2024092478 A1 WO2024092478 A1 WO 2024092478A1 CN 2022128860 W CN2022128860 W CN 2022128860W WO 2024092478 A1 WO2024092478 A1 WO 2024092478A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio data
sound
audio
channel
target
Prior art date
Application number
PCT/CN2022/128860
Other languages
English (en)
French (fr)
Inventor
田庆元
王梁
邱景博
曾繁荣
Original Assignee
深圳市韶音科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市韶音科技有限公司 filed Critical 深圳市韶音科技有限公司
Priority to PCT/CN2022/128860 priority Critical patent/WO2024092478A1/zh
Publication of WO2024092478A1 publication Critical patent/WO2024092478A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/12Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers

Definitions

  • the present invention relates to the field of audio processing, and in particular to an audio processing device and method.
  • the existing audio processing method is to directly send the audio data of one channel to multiple sound units for playback.
  • This specification provides an audio processing device and method with better sound quality.
  • the present specification provides an audio processing device, comprising: a sound module, comprising K sound units, each of which has different audio characteristics, wherein K is an integer greater than 1; and an audio processing circuit, configured to: obtain initial audio data, convert the initial audio data into target audio data of K channels, the target audio data of each channel being adapted to the audio characteristics of the corresponding sound unit, and input the target audio data of the K channels into the corresponding K sound units, so that each sound unit converts the corresponding target audio data into target audio to form a reverberation sound.
  • the audio processing circuit in order to convert the initial audio data into target audio data of K channels, and the target audio data of each channel is adapted to the audio characteristics of the corresponding sound-emitting unit, the audio processing circuit: copies the initial audio data into the initial audio data of K channels; and performs spectrum adjustment on the initial audio data of each channel so that the adjusted target audio data is adapted to the audio characteristics of the corresponding sound-emitting unit.
  • the frequency spectrum of the initial audio data includes K frequency bands; and the i-th sounding unit among the K sounding units has a desired sound effect in the i-th frequency band, where i is any integer in [1, K].
  • the audio processing circuit in order to perform frequency spectrum adjustment on the initial audio data of each channel so that the adjusted target audio data adapts to the audio characteristics of the corresponding sounding unit, the audio processing circuit: for the initial audio data of the i-th channel, retains or enhances the amplitude on the i-th frequency band, and attenuates the amplitude in other frequency bands, to obtain the target audio data of the i-th channel, and the target audio data of the i-th channel adapts to the audio characteristics of the sounding unit corresponding to the i-th channel.
  • the desired sound effect includes at least one of a degree of restoration exceeding a preset value, a preset sound effect, or a target sound effect.
  • the audio processing circuit in order to perform spectral adjustment on the initial audio data of each channel so that the adjusted target audio data adapts to the audio characteristics of the corresponding sound-emitting unit, the audio processing circuit: for the initial audio data of the i-th channel, retains the audio data of the i-th frequency band, filters out the audio data of other frequency bands, and obtains the target audio data of the i-th channel, and the target audio data of the i-th channel adapts to the audio characteristics of the sound-emitting unit corresponding to the i-th channel.
  • the audio processing circuit further includes K spectrum adjustment circuits, wherein the i-th spectrum adjustment circuit performs the spectrum adjustment on the initial audio data of the i-th channel when in operation.
  • the sound-emitting unit includes at least high-frequency, medium-frequency and low-frequency speakers, and the K frequency bands cover the high-frequency, the medium-frequency and the low-frequency.
  • the audio processing circuit also includes a DAC module, which, when working, receives the target audio data of the K channels; converts the target audio data of the K channels into K analog electrical signals; and inputs the K analog electrical signals into corresponding sound units respectively.
  • a DAC module which, when working, receives the target audio data of the K channels; converts the target audio data of the K channels into K analog electrical signals; and inputs the K analog electrical signals into corresponding sound units respectively.
  • the audio processing circuit in order to input the target audio data of the K channels into the corresponding K sounding units respectively, the audio processing circuit: combines the target audio of each channel to obtain integrated audio data; and inputs the integrated audio data into the corresponding K sounding units respectively.
  • the integrated audio data includes K segments of sub-data, wherein the i-th segment of sub-data includes the target audio data of the i-th channel and the corresponding i-th identifier, where i is any integer in [1, K].
  • the audio processing circuit also includes a digital audio interface, and when in operation, the integrated audio data is respectively input into the corresponding K sound-emitting units through the digital audio interface, wherein each of the K sound-emitting units includes an identification circuit and at least one speaker, and when the i-th identification circuit is in operation: the integrated audio data is received, the corresponding i-th identifier is identified in the integrated audio data, and sub-data corresponding to other identifiers are filtered out, and the target audio data corresponding to the i-th identifier is converted into target audio covering the i-th frequency band, and the target audio is sent to the at least one speaker.
  • the audio processing circuit also includes a digital audio interface, and when the audio digital interface is in operation: receives the integrated audio data; identifies an identifier corresponding to each sound-emitting unit and target audio data corresponding to the identifier in the integrated audio data; and sends the target audio data to the corresponding sound-emitting unit.
  • the K sound-emitting units play the corresponding target audio simultaneously with the same phase when operating.
  • the audio processing device is a headset.
  • the correspondence between the sound channels and the sound generating units includes one of a one-to-one correspondence, a many-to-one correspondence, or a one-to-many correspondence.
  • the audio processing circuit when at least one of the K sound-emitting units changes or receives a sound effect adjustment request, the audio processing circuit adjusts the correspondence between the sound channel and the sound-emitting unit.
  • the present specification also provides an audio processing method for headphones, comprising: obtaining initial audio data through the audio processing circuit of the headphones; converting the initial audio data into target audio data of K channels, the target audio data of each channel being adapted to the audio characteristics of a corresponding sound unit among K sound units in the headphones, wherein K is an integer greater than 1, and the audio characteristics of each sound unit are different; and inputting the target audio data of the K channels into the corresponding K sound units, respectively, so that the K sound units output reverberation sounds.
  • the converting of the initial audio data into target audio data of K channels, wherein the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit includes: copying the initial audio data into the initial audio data of K channels; and performing spectral adjustment on the initial audio data of each channel so that the adjusted target audio data is adapted to the audio characteristics of one sound unit among the corresponding K sound units in the headphones.
  • the initial audio data includes K frequency bands; and the i-th sound-emitting unit among the K generating units has a desired sound effect in the i-th frequency band, where i is any integer in [1, K], wherein the spectrum of the initial audio data of each channel is adjusted so that the adjusted target audio data adapts to the audio characteristics of one of the corresponding K sound-emitting units in the headset, including: for the initial audio data of the i-th channel, retaining or enhancing the amplitude on the i-th frequency band, attenuating the amplitude in other frequency bands, to obtain the target audio data of the i-th channel, and the target audio data of the i-th channel is adapted to the audio characteristics of the sound-emitting unit corresponding to the i-th channel in the headset.
  • the spectrum of the initial audio data of each channel is adjusted so that the adjusted target audio data adapts to the audio characteristics of one of the corresponding K sound units in the headset, including: for the initial audio data of the i-th channel, retaining the audio data of the i-th frequency band, filtering out the audio data of other frequency bands, and obtaining the target audio data of the i-th channel, wherein the target audio data of the i-th channel adapts to the audio characteristics of the sound unit corresponding to the i-th channel in the headset.
  • the target audio data of the K channels are respectively input into the corresponding K sound-emitting units so that the K sound-emitting units output reverberation sounds, including: converting the target audio data of each channel into an analog electrical signal; and inputting the analog electrical signal into the corresponding sound-emitting units so that each sound-emitting unit converts the corresponding analog electrical signal into the target audio to form reverberation.
  • inputting the target audio data of the K channels into the corresponding K sound emitting units respectively so that the K sound emitting units output reverberation sounds includes: combining the target audio data of each channel to obtain integrated audio data; and inputting the integrated audio data into the corresponding sound emitting units respectively so that the K sound emitting units output reverberation sounds.
  • inputting the integrated audio data respectively into corresponding sound emitting units so that the K sound emitting units output reverberation sound comprises: inputting the integrated audio data respectively into the K sound emitting units through a digital audio interface so that the K sound emitting units output reverberation sound.
  • the present specification also provides an audio processing system, comprising: at least one storage medium storing at least one instruction set for performing audio processing; and at least one processor communicatively connected to the at least one storage medium, wherein when the audio processing system is running, the at least one processor reads the at least one instruction set and executes the above-mentioned audio processing method according to the instructions of the at least one instruction set.
  • the audio processing device includes a sound module and an audio processing circuit, wherein the sound module includes K sound units, each sound unit has different audio characteristics, and K is an integer greater than 1; after obtaining the initial audio data, the audio processing circuit converts the initial audio data into target audio data of K channels, the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit, and the target audio data of the K channels are respectively input to the corresponding K sound units, so that each sound unit converts the corresponding target audio data into target audio to form a reverberation sound; because the scheme can convert the initial audio data into target audio data of K channels through the audio processing circuit and input them into the corresponding K sound units respectively, and the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit, thereby ensuring that the sound emitted by each sound unit in the sound module has the best sound effect, and therefore, the sound quality effect of audio processing can be improved.
  • the sound module includes K sound units, each sound unit has different audio characteristics, and K is an integer greater than
  • the audio processing method and system provided in the present specification, after obtaining the initial audio data, convert the initial audio data into target audio data of K channels, the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit, wherein K is an integer greater than 1; and the target audio data of the K channels are respectively input into the corresponding K sound units, so that the K sound units output reverberation sounds, wherein the audio characteristics of each sound unit are different; since the scheme can convert the initial audio data into target audio data of K channels, and then the target audio data of the K channels are respectively input into the corresponding K sound units, and the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit, it is ensured that the sound emitted by each sound unit in the sound module has the best sound effect, and therefore the sound quality effect of the audio processing can be improved.
  • FIG1 is a schematic diagram showing an application scenario of an audio processing device provided according to an embodiment of this specification.
  • FIG2 shows a hardware structure diagram of an audio processing device provided according to an embodiment of this specification
  • FIG3 shows a schematic diagram of the structure of an audio processing circuit provided according to an embodiment of this specification
  • FIG4 shows another structural schematic diagram of an audio processing circuit provided according to an embodiment of this specification.
  • FIG5 shows another structural schematic diagram of an audio processing circuit provided according to an embodiment of this specification
  • FIG6 shows a schematic structural diagram of a sound generating unit provided according to an embodiment of this specification
  • FIG7 shows a schematic diagram of a device of an audio processing system provided according to an embodiment of this specification.
  • FIG8 is a schematic flow chart of an audio processing method provided according to an embodiment of this specification.
  • FIG9 is a schematic flow chart of an audio processing method using a DAC module according to an embodiment of this specification.
  • FIG10 is a flow chart of another audio processing method using a DAC module according to an embodiment of this specification.
  • FIG. 11 is a schematic flow chart showing an audio processing method using a digital audio interface according to an embodiment of this specification.
  • An exemplary use scenario is as follows: after the audio processing device obtains the initial audio data output by the target device, it processes the initial audio data and plays the target audio corresponding to the processed audio data through the sound module, thereby forming a reverberant sound.
  • Fig. 1 shows a schematic diagram of an application scenario of an audio processing device provided according to an embodiment of this specification.
  • the application scenario 001 may include an audio processing device 10 , a target device 20 , and a network 30 .
  • the audio processing device 10 may include a sound module 100 and an audio processing circuit 200.
  • the audio processing circuit 200 may obtain initial audio data from the target device 20, process the initial audio data, and send the processed target audio data to the sound module 100 for sounding, thereby forming a reverberant sound.
  • the audio processing device 10 may store data or instructions for executing the audio processing method described in this specification, and may execute or be used to execute the data or instructions.
  • the audio processing device 10 may include a hardware device with a data information processing function and a necessary program required to drive the hardware device to work.
  • the audio processing device 10 may be a headset, a large home or commercial sound system, etc. The above audio processing method will be introduced in the subsequent content of this article.
  • the target device 20 may be an electronic device with an audio data output function.
  • the target device 20 may include a mobile device, a tablet computer, a laptop computer, a built-in device of a motor vehicle, or the like, or any combination thereof.
  • the mobile device may include a smart home device, a smart mobile device, a virtual reality device, an augmented reality device, or the like, or any combination thereof.
  • the smart home device may include a smart TV, a desktop computer, a smart speaker, or the like, or any combination thereof.
  • the smart mobile device may include a smart phone, a personal digital assistant, a gaming device, a navigation device, or the like, or any combination thereof.
  • the virtual reality device or the augmented reality device may include a virtual reality helmet, virtual reality glasses, a virtual reality handle, an augmented reality helmet, augmented reality glasses, an augmented reality handle, or the like, or any combination thereof.
  • the virtual reality device or the augmented reality device may include Google Glass, a head-mounted display, VR, or the like.
  • the built-in device in the motor vehicle may include an onboard computer, an onboard TV, or the like.
  • the target device 20 may include an audio acquisition device for acquiring audio data in the target space, thereby acquiring initial audio data. In some embodiments, the target device 20 may also receive initial audio data from other devices.
  • the target device 20 may be installed with one or more application programs (APPs).
  • APP application programs
  • the APP can provide users with the ability and interface to interact with the outside world.
  • the APP includes but is not limited to: web browser APP programs, search APP programs, chat APP programs, shopping APP programs, video APP programs, financial management APP programs, instant messaging tools, email clients, social platform software, etc.
  • a target APP may be installed on the target device 20.
  • the target APP can generate or obtain initial audio data, or the target APP can receive initial audio data from other devices.
  • the network 30 is used to provide a medium for communication connection between the audio processing device 10 and the target device 20.
  • the network 30 can promote the exchange of information or data.
  • the audio processing device 10 and the target device 20 can be connected to the network 30, and transmit information or data to each other through the network 30.
  • the network 30 can be any type of wireless network.
  • the network 30 can include a telecommunications network, an intranet, the Internet, a local area network (LAN), a wide area network (WAN), a wireless local area network (WLAN), a metropolitan area network (MAN), a public switched telephone network (PSTN), a Bluetooth TM network, a ZigBee TM network, a near field communication (NFC) network or a similar network.
  • the above network 30 can be a Bluetooth TM network.
  • the audio processing device 10 can be an audio processing device supporting the Bluetooth TM protocol; the target device 20 can be an audio data output device supporting the Bluetooth TM protocol.
  • the audio processing device 10 can communicate with the target device 20 based on the Bluetooth TM protocol .
  • the audio processing device 10 can also transmit data with the target device 20 through a wired network or a local area network.
  • Fig. 1 the number of audio processing devices 10, target devices 20 and networks 30 in Fig. 1 is only illustrative. Any number of audio processing devices 10, target devices 20 and networks 30 may be provided according to implementation requirements.
  • Fig. 2 shows a hardware structure diagram of an audio processing device 10 provided according to some embodiments of this specification.
  • the audio processing device 10 may include: a sound module 100 and an audio processing circuit 200 .
  • the sound module 100 may include K sound units 110, where K is an integer greater than 1, and each sound unit 110 has different audio characteristics.
  • the sound unit 110 may include one or more devices that can emit sound, for example, the sound unit 110 may include one or more speakers of various types. Different speakers may have different audio characteristics.
  • the audio characteristics may be understood as different sound effects produced for the same audio signal, that is, different sound units 110 have different frequency response functions for the audio signal input of the same frequency band, and therefore have different sound effects in terms of sound quality.
  • the audio characteristics of some sound generating units 110 are better restoration of intermediate frequency audio data and/or thicker and/or butter-like sound quality; the audio characteristics of some sound generating units are better restoration of high frequency audio data and/or can produce pure and clear sound quality.
  • the bone conduction sound unit 110 has a better sound effect for medium and high frequencies
  • the air conduction sound unit has a better sound effect for low-frequency audio.
  • low frequency may refer to a frequency band of approximately 20Hz to 150Hz
  • medium frequency may refer to a frequency band of approximately 150Hz to 5KHz
  • high frequency may refer to a frequency band of approximately 5KHz to 20KHz
  • medium-low frequency may refer to a frequency band of approximately 150Hz to 500Hz
  • medium-high frequency refers to a frequency band of 500Hz to 5KHz.
  • low frequency refers to the frequency band of roughly 20Hz to 80Hz
  • mid-low frequency may refer to the frequency band of roughly 80Hz-160Hz
  • mid-frequency may refer to the frequency band of roughly 160Hz to 1280Hz
  • mid-high frequency may refer to the frequency band of roughly 1280Hz-2560Hz
  • high frequency band may refer to the frequency band of roughly 2560Hz to 20KHz.
  • the audio processing circuit 200 is configured to obtain initial audio data and convert the initial audio data into target audio data of K channels.
  • the target audio data of each channel is adapted to the audio characteristics of the corresponding sound-emitting unit 110. For example, if the audio characteristics of the sound-emitting unit 110 are that it can produce clear sound quality in the high-frequency region, but the sound performance in the mid-frequency and low-frequency regions is ordinary, then the target audio data of the channel corresponding to the sound-emitting unit 110 has more high-frequency audio data, and the amount of audio data in the mid-frequency and low-frequency parts is less or even absent.
  • the audio processing circuit 200 can input the target audio data of the K channels into the corresponding K sound-emitting units respectively, so that each sound-emitting unit converts the corresponding target audio data into target audio to form a reverberant sound.
  • the initial audio data may be digitized audio data of one or more channels.
  • the initial audio data may include pulse code modulated audio data (PCM audio data) or other types of digitized audio data.
  • PCM audio data pulse code modulated audio data
  • the initial audio data may be a binary sequence directly formed by converting an analog signal through digital-to-analog conversion.
  • the audio processing circuit 200 may directly receive the initial audio data sent by the target device 20, or may receive the initial audio data sent by the audio acquisition device, or may obtain at least one audio data as the initial audio data in a preset audio data set, or may also obtain the original audio data.
  • the audio processing circuit 200 may select the audio data of a channel from the original audio data as the initial audio data, or the audio processing circuit 200 may further include an audio acquisition circuit.
  • the audio processing circuit 200 may collect audio data through the audio acquisition circuit, and perform digital-to-analog conversion on the collected original audio data, thereby obtaining the initial audio data.
  • the audio processing circuit 200 can convert the initial audio data into target audio data of K channels, and the target audio data of each channel is adapted to the audio characteristics of the corresponding sound-emitting unit 110.
  • a channel can be understood as a channel of audio data, and one channel can correspond to one sound-emitting unit 110.
  • the audio processing circuit 200 can copy the initial audio data into initial audio data of K channels, and perform spectrum adjustment on the initial audio data of each channel, so that the adjusted target audio data is adapted to the audio characteristics of the corresponding sound-emitting unit 110.
  • the audio processing circuit 200 may include a copy circuit, through which the initial audio data is copied into the initial audio data of K channels.
  • the initial audio data may also be copied into the initial audio data of K channels by a processor integrated in the audio processing circuit 200 or an independent processor.
  • the spectrum of the initial audio data may include K frequency bands, and the i-th sounding unit 110 among the K sounding units 110 has a desired sound effect in the i-th frequency band, where i is any integer in [1, K].
  • the K frequency bands may be any frequency band in the spectrum, and the K frequency bands may completely cover the entire spectrum, or may cover a pre-set high frequency, medium frequency, and low frequency spectrum range.
  • the desired sound effect may be a preset sound effect.
  • the sound effect may include concepts in the field of music such as the desired thickness, smoothness, purity, etc. for the sound quality, and may also include one, more or any combination of scientific concepts such as the degree of restoration exceeding a preset value.
  • the degree of restoration may be understood as the degree of restoration of the target audio being played and the original audio corresponding to the target audio, or may be understood as the degree of similarity between the target audio and the original audio.
  • the preset sound effect may be understood as the sound effect when the sound effect parameters of the reverberation sound emitted by the sound-emitting unit are preset to meet the preset sound parameters.
  • the target sound effect may be understood as a specific sound effect that is expected to be achieved.
  • the i-th sound-emitting unit 110 has the desired sound effect in the i-th frequency band, and the desired sound effects of different sound-emitting units 110 in different frequency bands may be the same or different.
  • different sound-emitting units 110 may correspond to different frequency bands according to their own audio characteristics. There may be overlapping frequency bands or no overlapping frequency bands in the different frequency bands corresponding to different sound-emitting units 110.
  • spectrum adjustment can be understood as adjusting the audio data of different frequency bands of the initial audio data.
  • the audio processing circuit 200 can use a spectrum adjustment algorithm to retain or enhance the amplitude on the i-th frequency band for the initial audio data of the i-th channel, and attenuate the amplitude in other frequency bands to obtain the target audio of the i-th channel, and the target audio data of the i-th channel is adapted to the audio characteristics of the sound unit 110 corresponding to the i-th channel; or, the audio processing circuit 200 can also use a spectrum adjustment algorithm to retain the audio data of the i-th frequency band for the initial audio data of the i-th channel, filter out the audio data of other frequency bands, and obtain the target audio of the i-th channel, and the target audio data of the i-th channel is adapted to the audio characteristics of the sound unit 110 corresponding to the
  • the spectrum adjustment algorithm can be understood as a software algorithm that performs spectrum adjustment on the initial audio data.
  • spectrum adjustment algorithms for example, at least one of an effect equalization algorithm (EQ), a frequency division algorithm, and a filter algorithm.
  • EQ effect equalization algorithm
  • filter algorithm a filter algorithm
  • the algorithm parameters of the spectrum adjustment algorithm can be dynamically adjusted according to demand, thereby achieving a richer listening experience.
  • the spectrum adjustment can also be performed by hardware, so the audio processing circuit 200 can also include K spectrum adjustment circuits 210, as shown in Figure 3, wherein the i-th spectrum adjustment circuit performs spectrum adjustment on the initial audio data of the i-th channel when working.
  • the spectrum adjustment method is similar to the adjustment method of the spectrum adjustment algorithm, as described above, and will not be repeated here.
  • spectrum adjustment circuit 210 may include at least one of a frequency division circuit and a DSP processing circuit.
  • the audio processing circuit 200 After the audio processing circuit 200 converts the initial audio data into target audio data of K channels, it can input the target audio data of the K channels into the corresponding K sounding units 110 respectively, and each sounding unit 110 converts the corresponding target audio data into target audio to form a reverberant sound. There are many ways for the audio processing circuit 200 to input the target audio data of K channels into the corresponding K sounding units 110 respectively. For example, the audio processing device 200 can directly input the target audio data of K channels into the corresponding K sounding units 110 respectively; or the audio processing circuit 200 can also combine the target audio of each channel to obtain integrated audio data, and input the integrated audio data into the corresponding K sounding units 110 respectively.
  • the audio processing circuit 200 may further include a DAC (digital-to-analog converter) module 220, and the DAC module 220 may include at least one DAC 221.
  • the audio processing circuit 200 may input the target audio data of the K channels to the corresponding K sounding units 110 through the DAC 221 in the DAC module.
  • the DAC module 220 When the DAC module 220 is working, it receives the target audio data of the K channels, converts the target audio data of the K channels into K analog electrical signals, and inputs the K analog electrical signals into the corresponding sounding units 110.
  • the audio processing circuit 200 can combine the target audio data of each channel into one frame of audio data through a combination operation, thereby obtaining integrated audio data.
  • the integrated audio data includes K segments of sub-data, wherein the i-th segment of sub-data includes the target audio data of the i-th channel and the corresponding i-th identifier, where i is any integer in [1, K].
  • the target audio data corresponding to the i-th channel can be identified in the integrated audio data. Therefore, the digital audio interface 230 or the i-th sounding unit 110 can identify the target audio data of the i-th channel based on the i-th identifier.
  • the audio processing circuit 200 can input the integrated audio data formed by the combination to the corresponding K sounding units 110. Since the integrated audio data is still digital audio data at this time, the audio processing circuit 200 can also include a digital audio interface 230, as shown in FIG5. When the audio processing circuit 200 is working, the integrated audio data can be input to the corresponding K sounding units through the digital audio interface 230.
  • the audio processing circuit 200 can input the integrated audio data into the corresponding K sound units through the digital audio interface 230.
  • the digital audio interface 230 directly sends the integrated audio data to the corresponding K sound units 110, and the K sound units 110 identify the corresponding target audio data in the integrated audio data.
  • the digital audio interface 230 identifies the target audio data corresponding to each sound unit 110 in the integrated audio data and sends the target audio data to the corresponding sound unit 110.
  • the digital audio interface (Digital Audio Interfaces, DAI) 230 can be understood as an interface for transmitting digital audio signals at the board level or between boards. Compared with the analog interface, the digital audio interface 230 has stronger anti-interference ability and simple hardware design. There can be many types of digital audio interface 230, for example, it can include at least one of I2S (a digital audio interface), TDM (a digital audio interface), PCM (a digital audio interface) and PDM (a digital audio interface).
  • I2S a digital audio interface
  • TDM a digital audio interface
  • PCM a digital audio interface
  • PDM digital audio interface
  • the sound unit 110 can also include an identification circuit 111 and at least one speaker 112, as shown in Figure 6.
  • the i-th identification circuit 111 When the i-th identification circuit 111 is working: it receives integrated audio data, identifies the corresponding i-th identifier in the integrated audio data, filters out sub-data corresponding to other identifiers, and converts the target audio data corresponding to the i-th identifier into target audio covering the i-th frequency band, and sends the target audio to a speaker 112.
  • the digital audio interface 230 identifies the target audio data corresponding to each sound unit 110 in the integrated audio data and sends the target audio data to the corresponding sound unit 110
  • the digital audio interface 230 needs to have the ability to identify the target audio data corresponding to each sound unit 110. Therefore, the digital audio interface 230 is now a digital audio interface 230 with identification and distribution functions.
  • the digital audio interface 230 is working: receiving the integrated audio data, identifying the identifier corresponding to each sound unit 110 and the target audio data corresponding to the identifier in the integrated audio data, and sending the target audio data to the corresponding sound unit 110.
  • the K sound-emitting units can convert the target audio data into target audio, thereby forming a reverberant sound.
  • the target audio can be understood as a sound that can be heard by the human ear, and the electrical signal corresponding to the audio data sound, or the electrical signal carrying the sound information, the electrical signal here can be a digital signal or an analog electrical signal, etc.
  • the target audio can be sent to at least one speaker 112, and the target audio can be played through at least one speaker 112.
  • the corresponding target audio is played through at least one speaker 112 in the K sound-emitting units 110, thereby forming a reverberant sound.
  • the K sound-emitting units 110 can use the same phase to play the corresponding target audio at the same time when working, so as to avoid the mutual influence between different target audios in the reverberation sound, thereby affecting the sound quality of the reverberation sound.
  • the K sound-emitting units include at least high-frequency, medium-frequency and low-frequency speakers 112, and the high-frequency, medium-frequency and low-frequency can be set according to the actual spectrum of the initial audio data.
  • the spectrum of the initial audio data includes K frequency bands, and the K frequency bands can cover high frequency, medium frequency and low frequency.
  • low frequency can refer to a frequency band of approximately 20Hz to 150Hz
  • medium frequency can refer to a frequency band of approximately 150Hz to 5KHz
  • high frequency can refer to a frequency band of approximately 5KHz to 20KHz
  • medium-low frequency can refer to a frequency band of approximately 150Hz to 500Hz
  • medium-high frequency refers to a frequency band of 500Hz to 5KHz.
  • low frequency refers to the frequency band of roughly 20Hz to 80Hz
  • mid-low frequency may refer to the frequency band of roughly 80Hz-160Hz
  • mid-frequency may refer to the frequency band of roughly 160Hz to 1280Hz
  • mid-high frequency may refer to the frequency band of roughly 1280Hz-2560Hz
  • high frequency band may refer to the frequency band of roughly 2560Hz to 20KHz.
  • speakers 112 there can be many types of speakers 112, for example, they can include air speakers and vibration conduction speakers, etc.
  • the air speaker can be understood as a speaker that outputs air-conducted sound waves
  • the vibration speaker can be understood as a speaker that outputs sound waves conducted by solid media (such as bone-conducted sound waves).
  • the vibration speaker and the air conduction speaker can be two independent functional devices, or they can be part of a single device that realizes multiple functions.
  • Each of the K sound-emitting units 110 can include at least one of an air speaker and a vibration speaker.
  • the audio processing device 10 may be a headset.
  • headsets There may be many types of headsets, for example, wired headsets, wireless headsets, or Bluetooth headsets, etc.
  • the audio processing device 10 may also include an audio playback device that performs audio processing, for example, a hearing aid, a speaker, or other audio playback devices, and so on.
  • an audio playback device that performs audio processing, for example, a hearing aid, a speaker, or other audio playback devices, and so on.
  • the correspondence between the channels and the sound units 110 may include one of a one-to-one correspondence, a many-to-one relationship, or a one-to-many relationship.
  • the one-to-one correspondence may be that each channel corresponds to a sound unit 110.
  • the i-th channel may correspond to the i-th sound unit 110, so that the audio processing circuit 200 or the processor 400 sends the target audio data of each channel to the sound unit 110, where i is in the range of [1, K].
  • the many-to-one relationship may be that multiple channels correspond to the same sound unit 110.
  • the multiple channels may correspond to the same sound unit, so that the audio processing circuit 200 or the processor 400 sends the target audio data of M channels to N sound units 110.
  • the one-to-many relationship can be that one channel can correspond to multiple sound units 110. For example, taking M channels and N sound units, where M is less than N, in order to input the target audio data of each channel into the sound unit, one channel can be corresponded to multiple sound units, so that the audio processing circuit 200 or the processor 400 sends the target audio data of M channels to N sound units 110.
  • the audio processing circuit 200 adjusts the correspondence between the channel and the sounding unit.
  • the sounding unit 110 changes for example, it may include the sounding unit failing to sound, the sounding unit being abnormal, or the sounding performance of the sounding unit changing (for example, the sounding unit changes from bone conduction sounding to air conduction sounding unit, etc.) or other changes, etc.
  • the sound effect adjustment request may be a request to adjust the sound effect of the currently played audio.
  • the sound effect may be the playback effect of the currently played audio, and the types of sound effects may be various, for example, it may include heavy metal, light music, electronic music, classical music, pop music or jazz, etc.
  • the audio processing circuit 200 or the processor 400 may adjust the correspondence between the channel and the sounding unit.
  • the adjustment can be made in the same correspondence or in different correspondences.
  • the so-called adjustment in the same correspondence refers to adjusting the sound units corresponding to one or more channels without changing the type of the correspondence.
  • the i-th channel corresponding to the i-th sound unit can be adjusted to the i-th channel corresponding to the n-th sound unit.
  • the so-called adjustment to a different correspondence refers to adjusting the current correspondence to another correspondence.
  • the audio processing circuit 200 or the processor 400 can adjust the one-to-one correspondence to a many-to-one relationship, or the one-to-one correspondence to a one-to-many relationship.
  • the sound unit 110 corresponding to each channel can be the same as or different from that before the adjustment.
  • the i-th channel before adjustment corresponds to the i-th sound unit 110
  • the i+1-th channel corresponds to the i+1-th sound unit 110
  • the i-th channel after adjustment can correspond to the i-th sound unit 110
  • the i+1-th channel can also correspond to the i-th sound unit 110, and so on.
  • the audio processing circuit 200 in the audio processing device 10 and other electronic devices cooperating therewith can be circuits or electronic devices integrated on one or more circuit boards and electrically connected to each other.
  • the above audio processing device 10 may also include a processor and a storage medium, and then the processor is used to complete all or part of the functions of the above audio processing circuit 200 and some electronic devices.
  • FIG7 shows a schematic diagram of an audio processing system 001 (hereinafter referred to as system 001) according to some embodiments of the present application.
  • the audio processing device 10 may also include at least one storage medium 300 and at least one processor 400.
  • the audio processing device 10 may also include a communication port 500 and an internal communication bus 600.
  • the internal communication bus 600 may connect various system components, including the storage medium 300 , the processor 400 , and the communication port 500 .
  • the data communication between the audio processing device 10 and the outside world can be completed through the communication port 500.
  • the audio processing device 10 can obtain initial audio data from the target device 20 through the communication port 500.
  • At least one storage medium 300 may include a data storage device.
  • the data storage device may be a non-temporary storage medium or a temporary storage medium.
  • the data storage device may include one or more of a disk, a read-only storage medium (ROM), or a random access storage medium (RAM).
  • the storage medium 300 may also include at least one instruction set stored in the data storage device for acquiring initial audio data and processing the initial audio data.
  • the instruction is a computer program code
  • the computer program code may include a program, a routine, an object, a component, a data structure, a process, a module, etc. for executing the audio processing method provided in this specification.
  • At least one processor 400 can be connected to at least one storage medium 300 through an internal communication bus 600.
  • the communication connection refers to any form of connection that can directly or indirectly receive information.
  • At least one processor 400 is used to execute the above-mentioned at least one instruction set.
  • the audio processing device 10 is running, at least one processor 400 reads the at least one instruction set and executes the audio processing method provided in this specification according to the instructions of the at least one instruction set.
  • the processor 400 can perform all the steps included in the audio processing method.
  • the processor 242 can be in the form of one or more processors.
  • the processor 242 can include one or more hardware processors, such as a microcontroller, a microprocessor, a reduced instruction set computer (RISC), an application-specific integrated circuit (ASIC), an application-specific instruction set processor (ASIP), a central processing unit (CPU), a graphics processing unit (GPU), a physical processing unit (PPU), a microcontroller unit, a digital signal processor (DSP), a field programmable gate array (FPGA), an advanced RISC machine (ARM), a programmable logic device (PLD), any circuit or processor capable of performing one or more functions, etc., or any combination thereof.
  • RISC reduced instruction set computer
  • ASIC application-specific integrated circuit
  • ASIP application-specific instruction set processor
  • CPU central processing unit
  • GPU graphics processing unit
  • PPU physical processing unit
  • DSP digital signal processor
  • FPGA field programmable gate array
  • ARM advanced RISC machine
  • PLD programmable logic device
  • the audio processing device 10 in this specification may also include multiple processors 400, and therefore, the operations and/or method steps disclosed in this specification may be performed by one processor as described in this specification, or may be performed jointly by multiple processors.
  • the processor 400 of the audio processing device 10 in this specification performs step A and step B, it should be understood that step A and step B may also be performed jointly or separately by two different processors 400 (e.g., the first processor performs step A, the second processor performs step B, or the first and second processors perform steps A and B together).
  • all audio processing steps may be performed by the audio processing circuit 200 , or may be completed jointly by the audio processing circuit 200 , the storage medium 300 , and the processor 400 .
  • the audio processing circuit 200 obtains audio data, converts the initial audio data into target audio data of K channels, adapts the target audio data of each channel to the audio characteristics of the corresponding sound unit, and inputs the target audio data of the K channels into the corresponding K sound units 110 respectively, and each sound unit converts the corresponding target audio data into target audio to form a reverberation sound.
  • the audio processing circuit 200 or the processor 400 obtains the initial audio processing
  • the processor 400 obtains the control instruction in the storage medium 300, performs the copy operation based on the control instruction, copies the initial audio data of K channels, and uses the spectrum adjustment algorithm to perform spectrum adjustment on the initial audio data of K channels to obtain the target audio data of each channel
  • the processor 400 sends the target audio data to the DAC module 220 in the audio processing circuit 200
  • the DAC module 220 can convert the target audio data into an analog electrical signal after receiving the target audio data of each channel, and send the analog electrical signal to the corresponding sound unit 100, each sound unit converts the corresponding target audio data into the target audio to form a reverberation sound
  • the audio processing circuit 200 or the processor 400 The initial audio processing is obtained, the processor 400 obtains the control instruction in the storage medium 300, and performs a copy operation based on the control instruction to copy the initial audio data
  • the processor 400 outputs the initial audio data of K channels, and uses a spectrum adjustment algorithm to perform spectrum adjustment on the initial audio data of K channels to obtain target audio data of each channel.
  • the processor 400 combines the target audio data of each channel to obtain integrated audio data, and sends the integrated audio data to the digital audio interface 230 in the audio processing circuit 200.
  • the digital audio interface 230 identifies the target audio data corresponding to each sound unit 110 in the integrated audio data, and sends the target audio data to the corresponding sound unit 110.
  • Each sound unit 110 converts the corresponding target audio data into target audio to form reverberation sound, and so on.
  • FIG8 shows a flow chart of an audio processing method for headphones provided according to an embodiment of the present specification.
  • the audio processing device 10 can execute the audio processing method P700 of the present specification.
  • the audio processing circuit 200 and/or the processor 400 in the audio processing device 10 can read the instruction set stored in the local storage medium, and then, according to the instruction of the instruction set, execute the audio processing method P700 described in the present specification.
  • the method P700 may include:
  • the audio processing circuit 200 or the processor 400 may obtain initial audio data.
  • the manner of obtaining the initial audio data may refer to the above description, which will not be described in detail here.
  • S720 Convert the initial audio data into target audio data of K channels, where the target audio data of each channel is adapted to the audio characteristics of a corresponding sound unit among the K sound units in the earphone.
  • the audio processing circuit 200 or the processor 400 can copy the initial audio data into initial audio data of K channels, and perform spectrum adjustment on the initial audio of each channel so that the adjusted target audio data adapts to the audio characteristics of a corresponding one of the K sound units 110 in the earphone.
  • the initial audio data includes K frequency bands, which can be specifically described above.
  • the i-th sounding unit among the K sounding units 110 has a desired sound effect in the i-th frequency band, which can be specifically described above and will not be described here one by one.
  • the audio processing circuit 200 or the processor 400 can use a spectrum adjustment algorithm to perform spectrum adjustment on the initial audio data of each channel.
  • the audio processing circuit 200 or the processor 400 can retain or enhance the amplitude of the i-th frequency band for the i-th initial audio data, and attenuate the amplitudes in other frequency bands to obtain the target audio data of the i-th channel, and the target audio data of the i-th channel is adapted to the i-th channel in the earphone.
  • the audio characteristics of the corresponding sound unit 110 or, for the initial audio data of the i-th channel, retain the audio data of the i-th frequency band, filter out the audio data of other frequency bands, and obtain the target audio data of the i-th channel, the target audio data of the i-th channel is adapted to the audio characteristics of the sound unit 110 corresponding to the i-th channel in the earphone; or, the audio processing circuit 200 can also use the spectrum adjustment circuit 210 to perform spectrum adjustment on the initial audio data of each channel.
  • the specific adjustment method is similar to the adjustment method using the spectrum adjustment algorithm. See the above for details, and will not be repeated here.
  • S730 Input the target audio data of the K channels into the corresponding K sound-emitting units respectively, so that the K sound-emitting units output reverberation sounds.
  • the audio characteristics of each sound unit 110 are different.
  • the audio processing circuit 200 or the processor 400 can use the DAC module 220 to convert the target audio data of each channel into an analog electrical signal, and input the analog electrical signal into the corresponding sound unit 110, so that each sound unit converts the corresponding analog electrical signal into the target audio to form a reverberation sound, or obtain integrated audio data from the target audio data of each channel, and input the integrated audio data into the corresponding sound unit 110 through the digital audio interface 230, so that K sound units 110 output reverberation sounds.
  • the DAC module 220 to convert the target audio data of each channel into an analog electrical signal, and input the analog electrical signal into the corresponding sound unit 110, so that each sound unit converts the corresponding analog electrical signal into the target audio to form a reverberation sound, or obtain integrated audio data from the target audio data of each channel, and input the integrated audio data into the corresponding sound unit 110 through the digital audio interface 230, so that K sound units 110 output reverberation sounds.
  • the earphone may include at least one of an air conduction sounding unit, a bone conduction sounding unit and other types of sounding units.
  • an air conduction sounding unit a bone conduction sounding unit
  • other types of sounding units can have the desired sound effects for the full spectrum of audio data, then for other types of sounding units, there is no need to perform spectrum adjustment on the initial audio data.
  • the process of the audio processing method can be shown in Figures 9, 10 and 11, and can be specifically as follows:
  • Fig. 9 is a schematic diagram of a process of performing audio processing using the DAC module 220.
  • the audio processing circuit 200 or the processor 400 performs a copy operation on the initial audio data of one channel to obtain initial audio data corresponding to K channels, and performs spectrum adjustment on the initial audio data corresponding to C channels using a spectrum adjustment algorithm.
  • each channel corresponds to a DAC221, and each DAC221 corresponds to a sound unit 110.
  • the corresponding DAC221 in the DAC module 220 can be used to input the target audio data corresponding to the C channels into the sound unit 110 (1 ⁇ C), and then the initial audio data corresponding to the other (K-C) channels that have not been spectrally adjusted are input as target audio data into the sound unit 110 (C+1 ⁇ K). Then, the K sound units 110 can convert the received target audio data into target audio, and make sound through at least one speaker in the sound unit 110, thereby forming a reverberant sound.
  • the specific process can be found in the above description, and will not be repeated here.
  • the corresponding relationship between the channel and the sound unit can also include a many-to-one and one-to-many relationship. Therefore, the relationship between the DAC221 in the DAC module 220 and the sound unit 110 may not be a one-to-one corresponding relationship.
  • the DAC221 in the DAC module 220 can be regarded as a whole, as shown in Figure 10.
  • the audio processing circuit 200 or the processor 400 can send the target audio data to the sound unit 110 through the DAC module 220. Therefore, the DAC221 in the DAC module 220 can adaptively send the corresponding target audio data to the sound unit according to the performance of the sound unit 110.
  • the sound unit m has a better sounding performance for the target audio data of the nth channel that has been spectrally processed.
  • the audio processing circuit 200 or the processor 400 can send the target audio data corresponding to the nth channel to the sound unit, where m is any integer in [1, K] and n is any integer in [1, C].
  • the DAC221 in the DAC module 220 can also switch the target audio data corresponding to one or more channels to different sounding units.
  • the target audio data of the i-th channel currently corresponds to the sounding unit i, but at this time, as the performance of the sounding unit i changes, the sounding fails, or a sound effect adjustment request is received, the target audio data of the i-th channel can be sent to other sounding units that meet the preset sounding performance, so that other sounding units can continue to make sounds based on the target audio data, i is any integer in [1, K]. If other sounding units currently also have corresponding target audio data, the current target audio data can be discarded, or the two target audio data can be fused, and the sound can be made based on the fused audio data, and so on.
  • the DAC221 in the DAC module 220 can also switch the target audio data corresponding to multiple channels to the same sounding unit.
  • the audio processing circuit 200 or the processor 400 can input the target audio data of the i-th channel and the target audio data of the (i+1)-th channel to the same sound unit i through one or more DAC221 in the DAC module 220, where i is any integer in [1, K], etc.
  • the DAC221 in the DAC module 220 can also switch the target audio data of a single channel to multiple sound units.
  • the audio processing circuit 200 or the processor 400 can send the target audio data of the i-th channel to multiple sound units through one or more DAC221 in the DAC module 220, where i is any integer in [1, K].
  • FIG11 shows a schematic flow chart of audio processing using a digital audio interface 230.
  • the audio processing circuit 200 or the processor 200 performs a copy operation on the initial audio data of a channel to obtain the initial audio data corresponding to K channels, and the initial audio data corresponding to C channels thereof are spectrally adjusted using a spectrum adjustment algorithm, and then the target audio data corresponding to the C channels after spectrum processing and the initial audio data corresponding to the unprocessed (K-C) channels are combined as target audio data to obtain a frame of integrated audio data.
  • the integrated audio data is respectively input to the K sounding units 110 through the digital audio interface 230, so that the K sounding units can obtain the corresponding target audio data, and then the target audio data is converted into the target audio, and sound is emitted through at least one speaker in the sounding unit 110, thereby forming a reverberant sound.
  • the specific process can be referred to as described above, and will not be repeated here.
  • the audio data of one channel is sent to multiple channels, and the initial audio data corresponding to each channel can be adjusted individually to adapt to the audio characteristics of different sound units 110.
  • the algorithm parameters of the spectrum adjustment algorithm can be dynamically adjusted to achieve a richer listening experience.
  • the audio processing device includes a sound module and an audio processing circuit, wherein the sound module includes K sound units, each sound unit has different audio characteristics, and K is an integer greater than 1; after obtaining the initial audio data, the audio processing circuit converts the initial audio data into target audio data of K channels, the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit, and the target audio data of the K channels are respectively input to the corresponding K sound units, so that each sound unit converts the corresponding target audio data into target audio to form a reverberation sound; because the scheme can convert the initial audio data into target audio data of K channels through the audio processing circuit and input them into the corresponding K sound units respectively, and the target audio data of each channel is adapted to the audio characteristics of the corresponding sound unit, thereby ensuring that the sound emitted by each sound unit in the sound module has the best sound effect, and therefore, the sound quality effect of audio processing can be improved.
  • the sound module includes K sound units, each sound unit has different audio characteristics, and K is an integer greater than
  • the present specification provides a non-transitory storage medium storing at least one set of executable instructions for audio processing.
  • the executable instructions When the executable instructions are executed by the processor, the executable instructions instruct the processor to implement the steps of the audio processing method P700 described in the present specification.
  • various aspects of the present specification can also be implemented in the form of a program product, which includes a program code.
  • the program product When the program product is run on the audio processing device 10, the program code is used to make the audio processing device 10 perform the steps of the audio processing method P700 described in the present specification.
  • the program product for implementing the above method can use a portable compact disk read-only memory (CD-ROM) to include program code and can be run on the audio processing device 10.
  • CD-ROM portable compact disk read-only memory
  • the program product of the present specification is not limited to this.
  • the readable storage medium can be any tangible medium containing or storing a program, which can be used by the instruction execution system or used in combination with it.
  • the program product can use any combination of one or more readable media.
  • the readable medium can be a readable signal medium or a readable storage medium.
  • the readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination of the above.
  • readable storage media include: an electrical connection with one or more conductors, a portable disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the above.
  • the computer-readable storage medium may include a data signal propagated in a baseband or as part of a carrier wave, wherein the readable program code is carried. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above.
  • the readable storage medium may also be any readable medium other than a readable storage medium, which may send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, device, or device.
  • the program code contained on the readable storage medium may be transmitted using any suitable medium, including but not limited to wireless, wired, optical cable, RF, etc., or any suitable combination of the above.
  • Program code for performing the operations of the present specification may be written in any combination of one or more programming languages, including object-oriented programming languages such as Java, C++, etc., and conventional procedural programming languages such as "C" or similar programming languages.
  • the program code may be executed entirely on the audio processing device 10, partially on the audio processing device 10, as a stand-alone software package, partially on the audio processing device 10 and partially on a remote computing device, or entirely on a remote computing device.

Landscapes

  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Otolaryngology (AREA)
  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

本说明书提供的音频处理设备和方法,音频处理设备包括发声模组和音频处理电路,其中,发声模组包括K个发声单元,每个发声单元的音频特性均不同,K为大于1的整数;音频处理电路在获取初始音频数据后,将初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元的音频特性,以及将K路声道的目标音频数据分别输入至对应的K个发声单元,使每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音;该方案可以提升音频处理的音质效果。

Description

音频处理设备和方法 技术领域
本说明书涉及音频处理领域,尤其涉及一种音频处理设备和方法。
背景技术
随着音频播放的智能化,用户的音频的音质效果提出了更高的要求。在音频播放中,使用低音、中音和高音等多个发声单元是一种满足高音质需求的解决方案。在采用多个发声单元播放音频时,现有的音频处理方法是将一声道的音频数据直接发送至多个发声单元进行播放。
在对现有技术的研究和实践过程中,本申请的发明人发现由于不同发声单元具有不同的音频特性,对于低音、中音和高音等不同发声单元直接播放相同的音频数据,在播放过程中并不能发挥每个发声单元的最佳播放效果,从而影响了播放时的音质效果,因此,导致音频处理的音质效果较差。
因此,需要提供一种音质效果更好的音频处理设备和方法。
发明内容
本说明书提供一种音质效果更好的音频处理设备和方法。
第一方面,本说明书提供一种音频处理设备,包括:发声模组,包括K个发声单元,每个所述发声单元的音频特性均不同,其中K为大于1的整数;以及音频处理电路,被配置为:获取初始音频数据,将所述初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的所述发声单元的音频特性,以及将所述K路声道的目标音频数据分别输入至对应的所述K个发声单元,使每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音。
在一些实施例中,为了将所述初始音频数据转换成K路声道的目标音频数据,且每路声道的目标音频数据适配对应的所述发声单元的音频特性,所述音频处理电路:将所 述初始音频数据复制成K路声道的所述初始音频数据;以及对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述对应的发声单元的音频特性。
在一些实施例中,所述初始音频数据的频谱包括K个频段;以及所述K个发声单元中的第i个发声单元在第i个频段具有期望的音响效果,所述i为[1,K]中的任意整数,其中,为了对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述对应的发声单元的音频特性,所述音频处理电路:对第i路声道的所述初始音频数据,保留或增强第i个频段上的幅值,衰减其他频段下的幅值,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述第i路声道对应的发声单元的音频特性。
在一些实施例中,所述期望的音响效果包括还原度超过预设值、预设音响效果或目标音响效果中的至少一种。
在一些实施例中,为了对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述对应的发声单元的音频特性,音频处理电路:对第i声道的所述初始音频数据,保留第i个频段的音频数据,过滤掉其他频段的音频数据,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述第i路声道对应的发声单元的音频特性。
在一些实施例中,所述音频处理电路还包括K个频谱调节电路,其中,第i个频谱调节电路工作时对所述第i路声道的所述初始音频数据进行所述频谱调节。
在一些实施例中,所述发声单元至少包括高频、中频和低频的扬声器,以及所述K个频段覆盖所述高频、所述中频和所述低频。
在一些实施例中,所述音频处理电路还包括DAC模组,所述DAC模组工作时:接收所述K路声道的目标音频数据;将所述K路声道的目标音频数据转换为K路模拟电信号;以及将所述K路模拟电信号分别输入对应的发声单元。
在一些实施例中,为了将所述K路声道的目标音频数据分别输入至对应的所述K个发声单元,所述音频处理电路:将每路声道的目标音频进行组合,得到整合音频数据;以及将所述整合音频数据分别输入至对应的所述K个发声单元。
在一些实施例中,所述整合音频数据包括K段子数据,其中,第i段子数据包括第i路声道的目标音频数据和对应的第i个标识,所述i为[1,K]中的任意整数。
在一些实施例中,所述音频处理电路还包括数字音频接口,工作时通过所述数字音频接口将所述整合音频数据分别输入至对应的所述K个发声单元,其中,所述K个发 声单元中的每个所述发声单元包括识别电路和至少一个扬声器,当第i个识别电路工作时:接收到所述整合音频数据,在所述整合音频数据中识别出对应的第i个标识,并过滤掉其他标识对应的子数据,以及将所述第i个标识对应的目标音频数据转换为覆盖第i个频段的目标音频,并将所述目标音频发送至所述至少一个扬声器。
在一些实施例中,所述音频处理电路还包括数字音频接口,所述音频数字接口在工作时:接收所述整合音频数据;在所述整合音频数据中识别出每个发声单元对应的标识和所述标识对应的目标音频数据;以及将所述目标音频数据发送至对应的发声单元。
在一些实施例中,所述K个发声单元工作时以相同的相位同时播放对应的所述目标音频。
在一些实施例中,所述音频处理设备为耳机。
在一些实施例中,所述声道与所述发声单元之间对应关系包括一一对应的关系、多对一的关系或一对多的关系中的一种。
在一些实施例中,在所述K个发声单元中至少一个发声单元发生变化或者接收到音效调整请求时,所述音频处理电路调整所述声道与所述发声单元之间的对应关系。
第二方面,本说明书还提供一种用于耳机的音频处理方法,包括,通过所述耳机的音频处理电路:获取初始音频数据;将所述初始音频数据转换为K路声道的目标音频数据,每路声道的目标音频数据适配所述耳机内对应的K发声单元中的一个发声单元的音频特性,其中K为大于1的整数,且每个发声单元的音频特性均不同;以及将所述K路声道的目标音频数据分别输入对应的所述K个发声单元,使得所述K个发声单元输出混响声音。
在一些实施例中,所述将所述初始音频数据转换为K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元的音频特性,包括:将所述初始音频数据复制成K路声道的所述初始音频数据;以及对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述耳机内对应的K个发声单元中的一个发声单元的音频特性。
在一些实施例中,所述初始音频数据包括K个频段;以及所述K个发生单元中的第i个发声单元在第i个频段具有期望的音响效果,所述i为[1,K]中的任意整数,其中,所述对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述耳机内对应的K个发声单元中的一个发声单元的音频特性,包括:对第i路声道的所述初始音频数据,保留或增强第i个频段上的幅值,衰减其他频段下的幅值,得到第i 路声道的目标音频数据,所述第i路声道的目标音频数据适配所述耳机内所述第i路声道对应的发声单元的音频特性。
在一些实施例中,所述对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述耳机内对应的K发声单元中的一个发声单元的音频特性,包括:对第i路声道的所述初始音频数据,保留第i个频段的音频数据,滤掉其他频段的音频数据,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述耳机内所述第i路声道对应的发声单元的音频特性。
在一些实施例中,所述将所述K路声道的目标音频数据分别输入对应的所述K个发声单元,使得所述K个发声单元输出混响声音,包括:将每路声道的目标音频数据转换为模拟电信号;以及将所述模拟电信号分别输入所述对应的发声单元,使每个发声单元将对应的模拟电信号转换为目标音频,形成混响。
在一些实施例中,所述将所述K路声道的目标音频数据分别输入对应的所述K个发声单元,使得所述K个发声单元输出混响声音,包括:将每路声道的目标音频数据进行组合,得到整合音频数据;以及将所述整合音频数据分别输入至对应的发声单元,使得所述K个发声单元输出混响声音。
在一些实施例中,所述将所述整合音频数据分别输入至对应的发声单元,使得所述K个发声单元输出混响声音,包括:通过数字音频接口将所述整合音频数据分别输入至所述K个发声单元,使得所述K个发声单元输出混响声音。
第三方面,本说明书还提供一种音频处理系统,包括:至少一个存储介质,存储有至少一个指令集,用于进行音频处理;以及至少一个处理器,同所述至少一个存储介质通信连接,其中,当所述音频处理系统运行时,所述至少一个处理器读取所述至少一个指令集,并且根据所述至少一个指令集的指示执行上述的音频处理方法。
由以上技术方案可知,本说明书提供的音频处理设备,包括发声模组和音频处理电路,其中,发声模组包括K个发声单元,每个发声单元的音频特性均不同,K为大于1的整数;音频处理电路在获取初始音频数据后,将初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元的音频特性,以及将K路声道的目标音频数据分别输入至对应的K个发声单元,使每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音;由于该方案可以通过音频处理电路将初始音频数据转换为K路声道的目标音频数据并分别输入至对应的K个发声单元,而且,每路声道的目标音频数据适配对应的发声单元的音频特性,从而保证发声模组中每个发声单元发 出的声音具有最佳的发声效果,因此,可以提升音频处理的音质效果。
本说明书提供的音频处理方法和系统,在获取初始音频数据后,将初始音频数据转换为K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元的音频特性,其中K为大于1的整数;以及将K路声道的目标音频数据分别输入对应的K个发声单元,使得K个发声单元输出混响声音,其中,每个发声单元的音频特性均不同;由于该方案可以将初始音频数据转换为K路声道的目标音频数据,然后,分别将K路声道的目标音频数据输入至对应的K个发声单元,而且,每路声道的目标音频数据适配对应的发声单元的音频特性,从而保证发声模组中每个发声单元发出的声音具有最佳的发声效果,因此,可以提升音频处理的音质效果。
本说明书提供的音频处理设备和方法的其他功能将在以下说明中部分列出。根据描述,以下数字和示例介绍的内容将对那些本领域的普通技术人员显而易见。本说明书提供的音频处理设备和方法的创造性方面可以通过实践或使用下面详细示例中所述的方法、装置和组合得到充分解释。
附图说明
为了更清楚地说明本说明书实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本说明书的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1示出了根据本说明书的实施例提供的一种音频处理设备的应用场景示意图;
图2示出了根据本说明书的实施例提供的一种音频处理设备的硬件结构图;
图3示出了根据本说明书的实施例提供的一种音频处理电路的结构示意图;
图4示出了根据本说明书的实施例提供的一种音频处理电路的另一个结构示意图;
图5示出了根据本说明书的实施例提供的一种音频处理电路的另一结构示意图;
图6示出了根据本说明书的实施例提供的一种发声单元的结构示意图;
图7示出了根据本说明书的实施例提供的一种音频处理系统的设备示意图;
图8示出了根据本说明书的实施例提供的一种音频处理方法的流程示意图;
图9示出了根据本说明书的实施例提供的一种采用DAC模组的音频处理方法的流程示意图;
图10出了根据本说明书的实施例提供的另一种采用DAC模组的音频处理方法的流程示意图;以及
图11示出了根据本说明书的实施例提供的一种采用数字音频接口的音频处理方法的流程示意图。
具体实施方式
以下描述提供了本说明书的特定应用场景和要求,目的是使本领域技术人员能够制造和使用本说明书中的内容。对于本领域技术人员来说,对所公开的实施例的各种局部修改是显而易见的,并且在不脱离本说明书的精神和范围的情况下,可以将这里定义的一般原理应用于其他实施例和应用。因此,本说明书不限于所示的实施例,而是与权利要求一致的最宽范围。
这里使用的术语仅用于描述特定示例实施例的目的,而不是限制性的。比如,除非上下文另有明确说明,这里所使用的,单数形式“一”,“一个”和“该”也可以包括复数形式。当在本说明书中使用时,术语“包括”、“包含”和/或“含有”意思是指所关联的整数,步骤、操作、元素和/或组件存在,但不排除一个或多个其他特征、整数、步骤、操作、元素、组件和/或组的存在或在该系统/方法中可以添加其他特征、整数、步骤、操作、元素、组件和/或组。
考虑到以下描述,本说明书的这些特征和其他特征、以及结构的相关元件的操作和功能、以及部件的组合和制造的经济性可以得到明显提高。参考附图,所有这些形成本说明书的一部分。然而,应该清楚地理解,附图仅用于说明和描述的目的,并不旨在限制本说明书的范围。还应理解,附图未按比例绘制。
本说明书中使用的流程图示出了根据本说明书中的一些实施例的系统实现的操作。应该清楚地理解,流程图的操作可以不按顺序实现。相反,操作可以以反转顺序或同时实现。此外,可以向流程图添加一个或多个其他操作。可以从流程图中移除一个或多个操作。
在对本说明书具体实施例说明之前,先对本说明书的应用场景进行如下介绍。
本说明书涉及音频处理设备的使用场景。一个示例性的使用场景如下:音频处理设备获取到目标设备输出的初始音频数据之后,对初始音频数据进行处理,并通过发声模组播放处理后的音频数据对应的目标音频,从而形成混响声音。
图1示出了根据本说明书的实施例提供的一种音频处理设备的应用场景示意图。如图1所示,应用场景001可以包括音频处理设备10、目标设备20以及网络30。
所述音频处理设备10可以包括发声模组100和音频处理电路200。在一些实施例中,音频处理电路200可以从目标设备20获取初始音频数据,并对初始音频数据进行处理,并将处理好的目标音频数据发送至发声模组100进行发声,从而形成混响声音。在一些实施例中,音频处理设备10可以存储有执行本说明书描述的音频处理方法的数据或指令,并可以执行或用于执行所述数据或指令。在一些实施例中,音频处理设备10可以包括具有数据信息处理功能的硬件设备和驱动该硬件设备工作所需必要的程序。比如,音频处理设备10可以为耳机、大型家用或商用的音响系统等。上述音频处理方法将在本文后续内容中进行介绍。
目标设备20可以为具有音频数据输出功能的电子设备。在一些实施例中,目标设备20可以包括移动设备、平板电脑、笔记本电脑、机动车辆的内置设备或类似内容,或其任意组合。在一些实施例中,所述移动设备可包括智能家居设备、智能移动设备、虚拟现实设备、增强现实设备或类似设备,或其任意组合。在一些实施例中,所述智能家居装置可包括智能电视、台式电脑、智能音箱等,或任意组合。在一些实施例中,所述智能移动设备可包括智能手机、个人数字辅助、游戏设备、导航设备等,或其任意组合。在一些实施例中,所述虚拟现实设备或增强现实设备可能包括虚拟现实头盔、虚拟现实眼镜、虚拟现实手柄、增强现实头盔、增强现实眼镜、增强现实手柄或类似内容,或其中的任何组合。例如,所述虚拟现实设备或所述增强现实设备可能包括谷歌眼镜、头戴式显示器、VR等。在一些实施例中,所述机动车中的内置装置可包括车载计算机、车载电视等。在一些实施例中,目标设备20可以包括音频采集设备,用于采集目标空间内的音频数据,从而获取初始音频数据。在一些实施例中,目标设备20还可以从其他设备接收初始音频数据。
在一些实施例中,目标设备20可以安装有一个或多个应用程序(APP)。所述APP能够为用户提供通过同外界交互的能力以及界面。所述APP包括但不限于:网页浏览器类APP程序、搜索类APP程序、聊天类APP程序、购物类APP程序、视频类APP程序、理财类APP程序、即时通信工具、邮箱客户端、社交平台软件等等。在一些实施例中,目标设备20上可以安装有目标APP。所述目标APP能够生成或获取初始音频数据,或者所述目标APP能够从其他设备接收初始音频数据。
网络30用以在音频处理设备10和目标设备20之间提供通信连接的介质。网络30可以促进信息或数据的交换。如图1所示,音频处理设备10和目标设备20可以同网络30连接,并且通过网络30互相传输信息或数据。在一些实施例中,网络30可以是任何类型的无线网络。比如,网络30可以包括电信通信网络、内联网、互联网、局域网(LAN)、广域网(WAN)、无线局域网(WLAN)、大都市市区网(MAN)、公用电话交换网(PSTN)、蓝牙 TM网络、ZigBee TM网络、近场通信(NFC)网络或类似网络。比如,上述网络30可以为蓝牙 TM网络,则该情况下,音频处理设备10可以为支持蓝牙 TM(bluetooth TM)协议的音频处理设备;目标设备20可以为支持蓝牙 TM协议的音频数据输出设备。音频处理设备10可以与目标设备20之间基于蓝牙 TM协议进行通信。在一些实施例中,音频处理设备10还可以与目标设备20之间通过有线网络或者局域网络进行数据传输。
应该理解,图1中的音频处理设备10、目标设备20和网络30的数目仅仅是示意性的。根据实现需要,可以具有任意数目的音频处理设备10、目标设备20和网络30。
图2示出了根据本说明书的一些实施例提供的一种音频处理设备10的硬件结构图。如图2所示,音频处理设备10可以包括:发声模组100和音频处理电路200。
在一些实施例中,发声模组100可以包括K个发声单元110,K为大于1的整数,另外,每个发声单元110的音频特性均不同。所述发声单元110可以包括一个或多个可以发出声音的设备,比如,所述发声单元110可以包括一个或多个各种类型的扬声器。不同扬声器可以有不同的音频特性。所述音频特性可以理解为针对相同的音频信号而产生不同的音响效果,也就是说,不同的发声单元110对相同频段的音频信号输入其频响函数不同的,因此在音质上具有不同的音响效果。比如,有些发生单元110的音频特性为对中频音频数据具有更好的还原度和/或发出更浑厚和/或更丝滑的音响效果(thick and/or butter-like sound quality);有些发生单元的音频特性对在高频的音频数据有更好的还原度和/或能产生纯净清澈的音响效果(pure and clear sound quality)。再比如,以发声单元110包括骨传导发声单元和气传导发声单元为例,骨传导发声单元针对中、高频具有更好的音响效果,而气传导发声单元针对低频的音频具有更好的发声效果。其中,对人耳来说,低频可以指的是大体上20Hz至150Hz的频段,中频可以指的是大体上150Hz至5KHz的频段,高频段可以指的是大体上5KHz至20KHz的频段,中低频可以指的是大体上150Hz至500Hz的频段,中高频指的是500Hz至5KHz的频段。本领域普通技术人员将会理解,上述频段的区分只是作为一个例子大概给出区间。上述频段的定义可 以随着不同行业、不同的应用场景和不同分类标准而改变。比如在另外一些应用场景下,低频指的是大体上20Hz至80Hz的频段,中低频可以指大体上80Hz-160Hz之间的频段,中频可以指大体上160Hz至1280Hz的频段,中高频可以指大体上1280Hz-2560Hz的频段,高频段可以指大体上2560Hz至20KHz的频段。
音频处理电路200被配置为可以获取初始音频数据,将初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元110的音频特性,比如,如果发声单元110的音频特性是在高频区可以发出清澈的音质,但是在中频和低频区的音响表现力普通,则该发声单元110对应的那路声道的目标音频数据便具有更多的高频音频数据,而在中频和低频部分的音频数据量较少甚至没有。将初始音频数据转换之后,音频处理电路200可以将K路声道的目标音频数据分别输入至对应的K个发声单元,使每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音。
其中,初始音频数据可以为一个或多个声道的数字化音频数据。该初始音频数据可以包括脉冲编码调制的音频数据(PCM音频数据)或者其他类型的数字化音频数据。以初始音频数据为PCM音频数据为例,该初始音频数据就可以为将一种模拟信号经数模转换直接形成的二进制序列。音频处理电路200获取初始音频数据的方式可以有多种,比如,音频处理电路200可以直接接收目标设备20发送的初始音频数据,或者,可以接收音频采集设备发送的初始音频数据,或者,可以在预设音频数据集合中获取至少一个音频数据作为初始音频数据,或者,还可以获取原始音频数据。然后,音频处理电路200可以在原始音频数据中选取出一个声道的音频数据作为初始音频数据,或者,音频处理电路200上还可以包括音频采集电路。所述音频处理电路200可以通过音频采集电路采集音频数据,并对采集到的原始音频数据进行数模转换,从而得到初始音频数据。
音频处理电路200在获取初始音频之后,便可以将初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元110的音频特性。声道可以理解为音频数据的通道,一个声道可以对应一个发声单元110。音频处理电路200将初始音频数据转换成K路声道的目标音频数据的方式可以有多种,比如,音频处理电路200可以将初始音频数据复制成K路声道的初始音频数据,以及对每路声道的初始音频数据进行频谱调节,使得调节后的目标音频数据适配对应的发声单元110的音频特性。
其中,音频处理电路200可以包括复制电路,通过复制电路将初始音频数据复制成K路声道的初始音频数据。在一些实施例中,还可以通过集成在音频处理电路200上的 处理器或者独立处理器将初始音频数据复制成K路声道的初始音频数据。
其中,初始音频数据的频谱可以包括K个频段,K个发声单元110中的第i个发声单元110在第i个频段具有期望的音响效果,i为[1,K]中的任意整数。以初始音频数据的频谱分布在[20Hz,20KHz]为例,则K个频段可以为该频谱中的任意一个频段,K个频段可以完整覆盖整个频谱,也可以覆盖预先设定的高频、中频和低频等频谱范围,K个频段之间可以存在重合的频段,也可以是完全独立的频段,等等。
在一些实施例中,期望的音响效果可以为预设的音响效果。所述音响效果可以包括对音质所期望的浑厚度、丝滑度、纯净度等音乐领域的概念,也可以包括还原度超过预设值等科学概念中一个、多个或者任意组合的。比如,所述还原度可以理解为在播放的目标音频与该目标音频对应的原始音频的还原程度,也可以理解为目标音频与原始音频之间的相似程度。所述预设音响效果可以理解为预先设定的针对发声单元发出的混响声音的音效参数符合预设音响参数时的音响效果。所述目标音响效果可以理解为期望达到的一个特定音响效果。在这里需要说明的是,第i个发声单元110在第i个频段具有期望的音响效果,不同发声单元110在不同频段上的期望的音响效果可以相同,也可以不相同。另外,不同发声单元110根据自身的音频特性,可以对应不同的频段。不同发声单元110对应不同的频段中可以存在重合的频段,也可以不存在重合的频段。
其中,频谱调节可以理解为针对初始音频数据的不同频段的音频数据进行调节。音频处理电路200对每路声道的初始音频数据进行频谱调节的方式可以有多种。比如,对任意在[1,K]之间的整数i,音频处理电路200可以采用频谱调节算法对第i路声道的初始音频数据,保留或增强第i个频段上的幅值,衰减其他频段下的幅值,得到第i路声道的目标音频,该第i路声道的目标音频数据适配第i路声道对应的发声单元110的音频特性;或者,音频处理电路200还可以采用频谱调节算法对第i路声道的初始音频数据,保留第i个频段的音频数据,过滤掉其他频段的音频数据,得到第i路声道的目标音频,该第i路声道的目标音频数据适配第i路声道对应的发声单元110的音频特性。
其中,频谱调节算法可以理解为一种对初始音频数据进行频谱调节的软件算法。频谱调节算法的类型可以有多种,比如,可以包括效果均衡算法(EQ)、分频算法和滤波器算法中的至少一种。另外,频谱调节算法的算法参数可以根据需求动态调节,从而实现更加丰富的听感。
在一些实施例中,还可以通过硬件方式进行频谱调节,因此,音频处理电路200还 可以包括K个频谱调节电路210,如图3所示,其中,第i个频谱调节电路工作时对第i路声道的初始音频数据进行频谱调节。频谱调节的方式与频谱调节算法的调节方式类似,详见上文所述,在此就不再一一赘述。
其中,频谱调节电路210的类型可以有多种,比如,可以包括分频电路和DSP处理电路等的至少一种。
音频处理电路200在将初始音频数据转换为K路声道的目标音频数据之后,便可以将K路声道的目标音频数据分别输入至对应的K个发声单元110,每个发声单元110将对应的目标音频数据转换为目标音频,形成混响声音。音频处理电路200将K路声道的目标音频数据分别输入至对应的K个发声单元110的方式可以有多种,比如,音频处理设备200可以直接将K路声道的目标音频数据分别输入至对应的K个发声单元110;或者,音频处理电路200还可以将每路声道的目标音频进行组合,得到整合音频数据,将整合音频数据分别输入至对应的K个发声单元110。
其中,如图4所示,音频处理电路200还可以包括DAC(数模转换器)模组220,DAC模组220可以包括至少一个DAC221。音频处理电路200就可以通过DAC模组中的DAC221将将K路声道的目标音频数据分别输入至对应的K个发声单元110。DAC模组220在工作时,接收K路声道的目标音频数据,将K路声道的目标音频数据转换为K路模拟电信号,以及将K路模拟电信号分别输入对应的发声单元110。DAC模组220中的DAC221与K个发声单元之间的对应关系可以有多种,比如,可以包括一对一,一对多,或者多对多,等等。
其中,音频处理电路200将每路声道的目标音频数据进行组合的方式可以有多种,比如,音频处理电路200可以通过组合操作将每路声道的目标音频数据组成一帧音频数据,从而得到整合音频数据。
其中,整合音频数据包括K段子数据,其中,第i段子数据包括第i路声道的目标音频数据和对应的第i个标识,所述i为[1,K]中的任意整数。通过第i个标识,就可以在整合音频数据中识别出第i路声道对应的目标音频数据。因此,数字音频接口230或者第i个发声单元110就可以基于第i个标识,识别出第i路声道的目标音频数据。
音频处理电路200在将每路声道的目标音频数据进行组合之后,便可以将组合后形成的整合音频数据分别输入至对应的K个发声单元110。由于整合音频数据此时依然是数字化的音频数据,因此,音频处理电路200还可以包括数字音频接口230,如图5所 示。音频处理电路200工作时就可以通过数字音频接口230将整合音频数据分别输入至对应的K个发声单元。
其中,音频处理电路200就可以通过数字音频接口230将整合音频数据分别输入至对应的K个发声单元的方式可以有多种,比如,数字音频接口230在接收到整合音频数据之后,直接将整合音频数据分别发送至对应的K个发声单元110,K个发声单元110在整合音频数据中识别出对应的目标音频数据,或者,数字音频接口230在整合音频数据中识别出每个发声单元110对应的目标音频数据,将目标音频数据发送至对应的发声单元110。
其中,数字音频接口(Digital Audio Interfaces,DAI)230可以理解为板级或板间传输数字音频信号的接口,相比于模拟接口,数字音频接口230抗干扰能力更强,硬件设计简单。数字音频接口230的类型可以有多种,比如,可以包括I2S(一种数字音频接口)、TDM(一种数字音频接口)、PCM(一种数字音频接口)和PDM(一种数字音频接口)中的至少一种。在数字音频接口230直接将整合音频数据发送至K个发声单元110时,K个发声单元110中的每个发声单元110需要在整合音频数据中识别出对应的目标音频数据,因此,发声单元110还可以包括识别电路111和至少一个扬声器112,如图6所示。当第i个识别电路111工作时:接收整合音频数据,在整合音频数据中识别出对应的第i个标识,并滤掉其他标识对应的子数据,以及将第i个标识对应的目标音频数据转换为覆盖第i个频段的目标音频,并将目标音频发送至至一个扬声器112。
其中,在数字音频接口230在整合音频数据中识别出每个发声单元110对应的目标音频数据,并将目标音频数据发送至对应的发声单元110时,数字音频接口230就需要具有识别每个发声单元110对应的目标音频数据的能力,因此,数字音频接口230此时就是一个具备识别和分发功能的数字音频接口230。数字音频接口230在工作时:接收整合音频数据,在整合音频数据中识别出每个发声单元110对应的标识和该标识对应的目标音频数据,以及将目标音频数据发送至对应的发声单元110。
K个发声单元在接收到对应的目标音频数据之后,便可以将目标音频数据转换为目标音频,从而形成混响声音。所述目标音频就可以理解为人耳可以听见的声音,而音频数据声音对应的电信号,或者说载有声音信息的电信号,这里的电信号可以为数字信号或模拟电信号,等等。发声单元110将目标音频数据转换为目标音频之后,便可以将目标音频发送至至少一个扬声器112,通过至少一个扬声器112播放目标音频。通过K个 发声单元110中的至少一个扬声器112播放对应的目标音频,从而就可以形成混响声音。
其中,K个发声单元110工作时可以采用相同的相位同时播放对应的目标音频,就可以避免混响声音中不同的目标音频之间互相影响,从而影响了混响声音的音质。
其中,K个发声单元中至少包括高频、中频和低频的扬声器112,高频、中频和低频可以根据初始音频数据的实际频谱进行设定。初始音频数据的频谱包括K个频段,K个频段可以覆盖高频、中频和低频。如前所述,在一些应用场景中,低频可以指的是大体上20Hz至150Hz的频段,中频可以指的是大体上150Hz至5KHz的频段,高频段可以指的是大体上5KHz至20KHz的频段,中低频可以指的是大体上150Hz至500Hz的频段,中高频指的是500Hz至5KHz的频段。本领域普通技术人员将会理解,上述频段的区分只是作为一个例子大概给出区间。上述频段的定义可以随着不同行业、不同的应用场景和不同分类标准而改变。比如在另外一些应用场景下,低频指的是大体上20Hz至80Hz的频段,中低频可以指大体上80Hz-160Hz之间的频段,中频可以指大体上160Hz至1280Hz的频段,中高频可以指大体上1280Hz-2560Hz的频段,高频段可以指大体上2560Hz至20KHz的频段。
其中,扬声器112的类型可以有多种,比如,可以包括空气扬声器和振动传导扬声器,等等。所述空气扬声器可以理解为输出空气传导声波的扬声器,而振动扬声器可以理解为输出固体媒介传导的声波(如骨传导声波)的扬声器。振动扬声器和空气传导扬声器可以是两个独立的功能器件,也可以是实现多个功能的单个器件的一部分。K个发声单元110中每个发声单元110中可以包括空气扬声器和振动扬声器中的至少一种。
在一些实施例中,音频处理设备10可以为耳机。耳机的类型可以有多种,比如,可以包括有线耳机、无线耳机或蓝牙耳机,等等。
在一些实施例中,音频处理设备10还可以包括进行音频处理的音频播放设备,比如,可以包括助听器、音箱或者其他音频播放设备,等等。
在一些实施例中,声道与发声单元110之间的对应关系可以包括一一对应的关系、多对一的关系或一对多的关系中的一种。所述一一对应的关系就可以为每一路声道元对应一个发声单元110,比如,以发声单元110为K个,声道为K路为例,则第i路声道就可以对应第i个发声单元110,从而使得音频处理电路200或处理器400将每路声道的目标音频数据发送至发声单元110其中,i的取值范围为[1,K]。所述多对一的关系就可以为多路声道对应同一个发声单元110,比如,以声道为M路,发声单元的数量为N 个,M大于N为例,为了将每路声道的目标音频数据输入至发声单元110,就可以将多路声道对应同一发声单元,从而使得音频处理电路200或处理器400将M路声道的目标音频数据发送至N个发声单元110。所述一对多的关系就可以为将一路声道可以对应多个发声单元110,比如,以声道为M路,发声单元的数量为N个,M小于N为例,为了将每路声道的目标音频数据输入至发声单元,就可以将一路声道对应多个发声单元,从而使得音频处理电路200或处理器400将M路声道的目标音频数据发送至N个发声单元110。
在一些实施例中,在K个发声单元110中至少一个发声单元110发生变化或者接收到音效调整请求时,所述音频处理电路200调整声道与发声单元之间的对应关系。发声单元110发生变化的情况可以有多种,比如,可以包括发声单元发声失败、发声单元出现异常或者发声单元的发声性能发生变化(譬如,发声单元从骨导发声变为气导发声单元,等等)或者其他变化的情况,等等。所述音效调整请求可以为对当前播放的音频的音效进行调整的请求。所述音效可以为当前播放的音频的播放效果,音效的类型可以有多种,比如,可以包括重金属、轻音乐、电音、古典、流行音乐或爵士,等等。在K个发声单元110中至少一个发声单元110发声变化或者接收到音效调整请求时,音频处理电路200或者处理器400均可以调整声道与发声单元之间的对应关系。调整声道与发声单元之间的对应关系的方式可以有多种,比如,可以在同一种对应关系中进行调整,也可以调整为不同的对应关系。所谓在同一种对应关系中调整是指在不改变对应关系的类型的情况下调整一路或多路声道对应的发声单元,比如,以对应关系为一一对应的关系为例,就可以将第i路声道对应第i个发声单元调整为第i路声道对应第n个发声单元。所谓调整为不同的对应关系是指将当前的对应关系调整为另一种对应关系,比如,一当前的声道与发声单元110的对应关系为一一对应的关系为例,音频处理电路200或者处理器400可以将一一对应的关系调整为多对一的关系,或者,还可以将一一对应的关系调整为一对多的关系。在调整后的对应关系中,每路声道所对应的发声单元110与调整之前可以相同,也可以不同。比如,以当前对应的关系为为一一对应的关系,调整后的关系为多对一的关系为例,调整前的第i路声道对应第i个发声单元110,第i+1路声道对应第i+1个发声单元110,调整后的第i路声道可以对应第i个发声单元110,第i+1路声道也可以对应第i个发声单元110,等等。
在上述描述中,所述音频处理设备10中的音频处理电路200以及与其配合的其他电子器件,比如DAC、频谱调节设备等,都可以是集成在一块或者多块电路板上相互 电连接的电路或者电子器件。上述音频处理设备10还可以包括处理器和存储介质,然后通过处理器来完成上述音频处理电路200和一些电子器件的全部或者部分功能。
图7示出了根据本申请一些实施例中的音频处理系统001(以下简称系统001)的设备示意图。除了前面描述的硬件之外,音频处理设备10还可以包括至少一个存储介质300和至少一个处理器400。为了满足内部和外部通信需求,音频处理设备10还可以包括通信端口500和内部通信总线600。
内部通信总线600可以连接不同的系统组件,包括存储介质300、处理器400和通信端口500。
音频处理设备10同外界的数据通信可以通过通信端口500来完成。比如,音频处理设备10可以通过通信端口500从目标设备20中获取初始音频数据。
至少一个存储介质300可以包括数据存储装置。所述数据存储装置可以是非暂时性存储介质,也可以是暂时性存储介质。比如,所述数据存储装置可以包括磁盘、只读存储介质(ROM)或随机存取存储介质(RAM)中的一种或多种。当音频处理设备10运行时,存储介质300还可以包括存储在所述数据存储装置中的至少一个指令集,用于获取初始音频数据,并对初始音频数据进行处理。所述指令是计算机程序代码,所述计算机程序代码可以包括执行本说明书提供的音频处理方法的程序、例程、对象、组件、数据结构、过程、模块等等。
至少一个处理器400可以同至少一个存储介质300通过内部通信总线600通信连接。所述通信连接是指能够直接地或者间接地接收信息的任何形式的连接。至少一个处理器400用以执行上述至少一个指令集。当音频处理设备10运行时,至少一个处理器400读取所述至少一个指令集,并且根据所述至少一个指令集的指示执行本说明书提供的音频处理方法。处理器400可以执行音频处理方法包含的所有步骤。处理器242可以是一个或多个处理器的形式,在一些实施例中,处理器242可以包括一个或多个硬件处理器,例如微控制器,微处理器,精简指令集计算机(RISC),专用集成电路(ASIC),特定于应用的指令集处理器(ASIP),中央处理单元(CPU),图形处理单元(GPU),物理处理单元(PPU),微控制器单元,数字信号处理器(DSP),现场可编程门阵列(FPGA),高级RISC机器(ARM),可编程逻辑器件(PLD),能够执行一个或多个功能的任何电路或处理器等,或其任何组合。仅仅为了说明问题,在本说明书中音频处理设备10中仅描述了一个处理器400。然而,应当注意,本说明书中音频处理设备10还可以包括多 个处理器400,因此,本说明书中披露的操作和/或方法步骤可以如本说明书所述的由一个处理器执行,也可以由多个处理器联合执行。例如,如果在本说明书中音频处理设备10的处理器400执行步骤A和步骤B,则应该理解,步骤A和步骤B也可以由两个不同处理器400联合或分开执行(例如,第一处理器执行步骤A,第二处理器执行步骤B,或者第一和第二处理器共同执行步骤A和B)。
在一些实施例中,音频处理设备10对初始音频数据进行处理时,可以由音频处理电路200执行全部的音频处理步骤,或者,还可以由音频处理电路200、存储介质300和处理器400共同完成。
其中,音频处理电路200执行全部的音频处理步骤的方式可以有多种,比如,音频处理电路200获取音频数据,将初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元的音频特性,以及将K路声道的目标音频数据分别输入至对应的K个发声单元110,每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音。
其中,音频处理电路200、存储介质300和处理器400共同完成音频处理的步骤可以有多种,比如,音频处理电路200或处理器400获取初始音频处理,处理器400在存储介质300中获取控制指令,基于该控制指令执行复制操作,复制出K路声道的初始音频数据,并采用频谱调节算法对K路声道的初始音频数据进行频谱调节,得到每路声道的目标音频数据,处理器400将目标音频数据发送至音频处理电路200中的DAC模组200,DAC模组220在接收到每路声道的目标音频数据之后,便可以将目标音频数据转换为模拟电信号,并将模拟电信号发送至对应的发声单元100,每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音;或者,音频处理电路200或处理器400获取初始音频处理,处理器400在存储介质300中获取控制指令,基于该控制指令执行复制操作,复制出K路声道的初始音频数据,并采用频谱调节算法对K路声道的初始音频数据进行频谱调节,得到每路声道的目标音频数据,处理器400对每路声道的目标音频数据进行组合操作,得到整合音频数据,并将整合音频数据发送至音频处理电路200中的数字音频接口230,数字音频接口230将整合音频数据直接发送至每路发声单元110,每路发声单元110在整合音频数据中识别出对应的目标音频数据,并将目标音频数据转换为目标音频,形成混响声音;或者,音频处理电路200或处理器400获取初始音频处理,处理器400在存储介质300中获取控制指令,基于该控制指令执行复制操作,复制 出K路声道的初始音频数据,并采用频谱调节算法对K路声道的初始音频数据进行频谱调节,得到每路声道的目标音频数据,处理器400对每路声道的目标音频数据进行组合操作,得到整合音频数据,并将整合音频数据发送至音频处理电路200中的数字音频接口230,数字音频接口230在整合音频数据中识别出每路发声单元110对应的目标音频数据,并将目标音频数据发送至对应的发声单元110,每个发声单元110将对应的目标音频数据转换为目标音频,形成混响声音,等等。
下面对用于耳机的音频处理方法进行介绍。
图8示出了根据本说明书实施例提供的用于耳机的音频处理方法的流程示意图。音频处理设备10可以执行本说明书的音频处理方法P700。具体的,音频处理设备10中的音频处理电路200和/或处理器400可以读取存储在本地存储介质中的指令集,然后,根据指令集的指示,执行本说明书描述的音频处理方法P700。如图8所示,方法P700可以包括:
S710:获取初始音频数据。
例如,音频处理电路200或处理器400可以获取初始音频数据,获取初始音频数据的方式可以参见上文所述,在此就不再一一赘述。
S720:将初始音频数据转换为K路声道的目标音频数据,每路声道的目标音频数据适配耳机内对应的K发声单元中一个发声单元的音频特性。
例如,音频处理电路200或处理器400可以将初始音频数据复制成K路声道的初始音频数据,以及对每路声道的初始进行频谱调节,使得调节后的目标音频数据适配耳机内对应的K发声单元110中的一个发声单元的音频特性。
其中。初始音频数据包括K个频段,具体可以详见上文所述。K个发声单元110中第i个发声单元在第i个频段具有期望的音响效果,所述期望的音响效果可以参见上文所述,在此就不再一一赘述。
其中,对每路声道的初始音频数据进行频谱调节,使得调节后的目标音频数据适配耳机内对应的K个发声单元110中的一个发声单元的音频特性的方式可以有多种,比如,音频处理电路200或者处理器400可以采用频谱调节算法对每路声道的初始音频数据进行频谱调节,具体的调节方式可以有多种,譬如,音频处理电路200或者处理器400可以对第i路的初始音频数据,保留或增强第i个频段上的幅值,衰减其他频段下的幅值,得到第i路声道的目标音频数据,该第i路声道的目标音频数据适配耳机内第i路声道 对应的发声单元110的音频特性,或者,对第i路声道的初始音频数据,保留第i个频段的音频数据,滤掉其他频段的音频数据,得到第i路声道的目标音频数据,该第i路声道的目标音频数据适配耳机内第i路声道对应的发声单元110的音频特性;或者,音频处理电路200还可以采用频谱调节电路210对每路声道的初始音频数据进行频谱调节,具体的调节方式与采用频谱调节算法的调节方式类似,详见上文所述,在此就不再一一赘述。
其中,频谱调节算法的类型可以有多种,具体可以参见上文所述,在此就不再一一赘述。
S730:将K路声道的目标音频数据分别输入对应的K个发声单元,使得K个发声单元输出混响声音。
其中,每个发声单元110的音频特性均不同。
其中,将K路声道的目标音频数据分别输入对应的K个发声单元110,使得K个发声单元110输出混响声音的方式可以有多种,具体可以如下:
例如,音频处理电路200或处理器400可以采用DAC模组220将每路声道的目标音频数据转换为模拟电信号,以及将模拟电信号分别输入对应的发声单元110,使每个发声单元将对应的模拟电信号转换为目标音频,形成混响声音,或者,将每路声道的目标音频数据,得到整合音频数据,以及通过数字音频接口230将整合音频数据分别输入至对应的发声单元110,使得K个发声单元110输出混响声音,具体可以参见上文所述,在此就不再一一赘述。
其中,以音频处理设备10为耳机为例,耳机中可以包括气导发声单元、骨导发声单元和其他类型的发声单元中的至少一种。以其他类型的发声单元可以针对全频谱的音频数据均有期望的音响效果为例,则针对其他类型的发声单元,就无需对初始音频数据进行频谱调节。因此,以初始音频数据为一声道的初始音频数据,K个发声单元110中C个发声单元110对应的初始音频数据需要进行频谱调节,(K-C)个发声单元110无需进行频谱调节为例,音频处理方法的流程可以如图9、10和11所示,具体可以如下:
图9示出了采用DAC模组220进行音频处理的流程示意图。在图9中,音频处理电路200或处理器400对一声道的初始音频数据进行复制操作,得到K路声道对应的初始音频数据,对其中的C个声道对应的初始音频数据采用频谱调节算法进行频谱调节。以声道和发声单元110之间的对应关系为一一对应为例,每一路声道对应一个DAC221, 每一个DAC221对应一个发声单元110,就可以采用DAC模组220中对应的DAC221将其中的C路声道对应的目标音频数据输入至发声单元110(1~C),然后,将其他未进行频谱调节的(K-C)路声道对应的初始音频数据作为目标音频数据输入至发声单元110(C+1~K),然后,K个发声单元110就可以将接收到的目标音频数据转换为目标音频,并通过发声单元110中的至少一个扬声器发声,从而形成混响声音,具体过程可以参见上文所述,在此就不再一一赘述。
其中,需要说明的是,声道与发声单元之间的对应关系还可以包括多对一和一对多的关系,因此,DAC模组220中的DAC221与发声单元110的之间的关系也可以不为一一对应的关系,此时就可以将DAC模组220中的DAC221看作一个整体,如图10所示,此时,音频处理电路200或处理器400就可以通过DAC模组220统筹将目标音频数据发送至发声单元110。因此,DAC模组220中的DAC221就可以根据发声单元110的性能,自适应的将对应的目标音频数据发送至发声单元。比如,发声单元m对进行频谱处理过的第n路声道的目标音频数据具有更好的发声性能,此时,音频处理电路200或处理器400就可以将第n路声道对应的目标音频数据发送至发声单元,m为[1,K]中的任意整数,n为[1,C]中的任意整数。另外,DAC模组220中的DAC221还可以对一路或多路声道对应的目标音频数据切换输入不同的发声单元。比如,第i路声道的目标音频数据当前对应的发声单元为发声单元i,但是此时随着发声单元i性能发生变化、发声失败或者接收到音效调整请求时,就可以将第i路声道的目标音频数据发送至符合预设发声性能的其他发声单元,从而使得其他发声单元可以继续基于该目标音频数据进行发声,i为[1,K]中的任意整数。如果其他发声单元当前也有对应的目标音频数据时,就可以将当前的目标音频数据进行丢弃,或者,还可以将两个目标音频数据进行融合,并基于融合后的音频数据进行发声,等等。DAC模组220中的DAC221还可以将多路声道对应的目标音频数据切换输入同一发声单元。比如,音频处理电路200或处理器400可以通过DAC模组220中的一个或多个DAC221,将第i路声道的目标音频数据和第(i+1)路声道的目标音频数据输入至同一发声单元i,i为[1,K]中的任意整数,等等。DAC模组220中的DAC221还可以将单一声道的目标音频数据切换输入至多个发声单元,比如,音频处理电路200或处理器400可以通过DAC模组220中的一个或多个DAC221将第i路声道的目标音频数据发送至多个发声单元,i为[1,K]中的任意整数。
图11示出了采用数字音频接口230进行音频处理的流程示意图。在图11中,音频处理电路200或处理器200对一声道的初始音频数据进行复制操作,得到K个声道对应 的初始音频数据,对其中的C个声道对应的初始音频数据采用频谱调节算法进行频谱调节,然后,将频谱处理后的C个声道对应的目标音频数据和未处理的(K-C)个声道对应的初始音频数据作为目标音频数据进行组合操作,得到一帧的整合音频数据。通过数字音频接口230将整合音频数据分别输入至K个发声单元110,使得K个发声单元可以获取对应的目标音频数据,然后,将目标音频数据转换为目标音频,并通过发声单元110中的至少一个扬声器发声,从而形成混响声音,具体过程可以参见上文所述,在此就不再一一赘述。
其中,在音频处理方法的流程中,实现了将一个声道的音频数据发送至多个声道,而且,每个声道对应的初始音频数据还可以单独进行调整,以适应不同发声单元110的音频特性,另外,在采用频谱调节算法对初始音频数据进行频谱调节时,频谱调节算法的算法参数可以进行动态调节,从而实现更加丰富的听感。
综上所述,本说明书提供的音频处理设备和音频处理方法,音频处理设备包括发声模组和音频处理电路,其中,发声模组包括K个发声单元,每个发声单元的音频特性均不同,K为大于1的整数;音频处理电路在获取初始音频数据后,将初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的发声单元的音频特性,以及将K路声道的目标音频数据分别输入至对应的K个发声单元,使每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音;由于该方案可以通过音频处理电路将初始音频数据转换为K路声道的目标音频数据并分别输入至对应的K个发声单元,而且,每路声道的目标音频数据适配对应的发声单元的音频特性,从而保证发声模组中每个发声单元发出的声音具有最佳的发声效果,因此,可以提升音频处理的音质效果。
本说明书另一方面提供一种非暂时性存储介质,存储有至少一组用来进行音频处理的可执行指令。当所述可执行指令被处理器执行时,所述可执行指令指导所述处理器实施本说明书所述音频处理方法P700的步骤。在一些可能的实施方式中,本说明书的各个方面还可以实现为一种程序产品的形式,其包括程序代码。当所述程序产品在音频处理设备10上运行时,所述程序代码用于使音频处理设备10执行本说明书描述的音频处理方法P700的步骤。用于实现上述方法的程序产品可以采用便携式紧凑盘只读存储器(CD-ROM)包括程序代码,并可以在音频处理设备10上运行。然而,本说明书的程序产品不限于此,在本说明书中,可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统使用或者与其结合使用。所述程序产品可以采用一个或多个可读介质的任意组合。可读介质可以是可读信号介质或者可读存储介质。可读存储介质 例如可以为但不限于电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。可读存储介质的更具体的例子包括:具有一个或多个导线的电连接、便携式盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。所述计算机可读存储介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了可读程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。可读存储介质还可以是可读存储介质以外的任何可读介质,该可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。可读存储介质上包含的程序代码可以用任何适当的介质传输,包括但不限于无线、有线、光缆、RF等等,或者上述的任意合适的组合。可以以一种或多种程序设计语言的任意组合来编写用于执行本说明书操作的程序代码,所述程序设计语言包括面向对象的程序设计语言—诸如Java、C++等,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在音频处理设备10上执行、部分地在音频处理设备10上执行、作为一个独立的软件包执行、部分在音频处理设备10上部分在远程计算设备上执行、或者完全在远程计算设备上执行。
上述对本说明书中的一些实施例进行了描述。其他实施例在所附权利要求书的范围内。在一些情况下,在权利要求书中记载的动作或步骤可以按照不同于实施例中的顺序来执行并且仍然可以实现期望的结果。另外,在附图中描绘的过程不一定要求示出特定顺序或者连续顺序才能实现期望的结果。在某些实施方式中,多任务处理和并行处理也是可以的或者是可能有利的。
综上所述,在阅读本详细公开内容之后,本领域技术人员可以明白,前述详细公开内容可以仅以示例的方式呈现,并且可以不是限制性的。尽管这里没有明确说明,本领域技术人员可以理解本说明书需求囊括对实施例的各种合理改变,改进和修改。这些改变,改进和修改旨在由本说明书提出,并且在本说明书的示例性实施例的精神和范围内。
此外,本说明书中的某些术语已被用于描述本说明书的实施例。例如,“一个实施例”,“实施例”和/或“一些实施例”意味着结合该实施例描述的特定特征,结构或特性可以包括在本说明书的至少一个实施例中。因此,可以强调并且应当理解,在本说明书的各个部分中对“实施例”或“一个实施例”或“替代实施例”的两个或更多个引用不一定都指 代相同的实施例。此外,特定特征,结构或特性可以在本说明书的一个或多个实施例中适当地组合。
应当理解,在本说明书的实施例的前述描述中,为了帮助理解一个特征,出于简化本说明书的目的,本说明书将各种特征组合在单个实施例、附图或其描述中。然而,这并不是说这些特征的组合是必须的,本领域技术人员在阅读本说明书的时候完全有可能将其中一部分设备标注出来作为单独的实施例来理解。也就是说,本说明书中的实施例也可以理解为多个次级实施例的整合。而每个次级实施例的内容在于少于单个前述公开实施例的所有特征的时候也是成立的。
本文引用的每个专利,专利申请,专利申请的出版物和其他材料,例如文章,书籍,说明书,出版物,文件,物品等,可以通过引用结合于此。用于所有目的全部内容,除了与其相关的任何起诉文件历史,可能与本文件不一致或相冲突的任何相同的,或者任何可能对权利要求的最宽范围具有限制性影响的任何相同的起诉文件历史。现在或以后与本文件相关联。举例来说,如果在与任何所包含的材料相关联的术语的描述、定义和/或使用与本文档相关的术语、描述、定义和/或之间存在任何不一致或冲突时,使用本文件中的术语为准。
最后,应理解,本文公开的申请的实施方案是对本说明书的实施方案的原理的说明。其他修改后的实施例也在本说明书的范围内。因此,本说明书披露的实施例仅仅作为示例而非限制。本领域技术人员可以根据本说明书中的实施例采取替代配置来实现本说明书中的申请。因此,本说明书的实施例不限于申请中被精确地描述过的实施例。

Claims (23)

  1. 一种音频处理设备,其特征在于,包括:
    发声模组,包括K个发声单元,每个所述发声单元的音频特性均不同,其中K为大于1的整数;以及
    音频处理电路,被配置为:
    获取初始音频数据,
    将所述初始音频数据转换成K路声道的目标音频数据,每路声道的目标音频数据适配对应的所述发声单元的音频特性,以及
    将所述K路声道的目标音频数据分别输入至对应的所述K个发声单元,使每个发声单元将对应的目标音频数据转换为目标音频,形成混响声音。
  2. 根据权利要求1所述的音频处理设备,其特征在于,为了将所述初始音频数据转换成K路声道的目标音频数据,且每路声道的目标音频数据适配对应的所述发声单元的音频特性,所述音频处理电路:
    将所述初始音频数据复制成K路声道的所述初始音频数据;以及
    对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述对应的发声单元的音频特性。
  3. 根据权利要求1所述的音频处理设备,其特征在于,
    所述初始音频数据的频谱包括K个频段;以及
    所述K个发声单元中的第i个发声单元在第i个频段具有期望的音响效果,所述i为[1,K]中的任意整数,
    其中,为了对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述对应的发声单元的音频特性,所述音频处理电路:
    对第i路声道的所述初始音频数据,保留或增强第i个频段上的幅值,衰减其他频段下的幅值,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述第i路声道对应的发声单元的音频特性。
  4. 根据权利要求3所述的音频处理设备,其特征在于,所述期望的音响效果包括还原度超过预设值、预设音响效果或目标音响效果中的至少一种。
  5. 根据权利要求3所述的音频处理设备,其特征在于,为了对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述对应的发声单元的音频特性,所述音频处理电路:
    对第i声道的所述初始音频数据,保留第i个频段的音频数据,过滤掉其他频段的音频数据,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述第i路声道对应的发声单元的音频特性。
  6. 根据权利要求3所述的音频处理设备,其特征在于,所述音频处理电路还包括K个频谱调节电路,其中,第i个频谱调节电路工作时对所述第i路声道的所述初始音频数据进行所述频谱调节。
  7. 根据权利3所述的音频处理设备,其特征在于,所述发声单元至少包括高频、中频和低频的扬声器,以及
    所述K个频段覆盖所述高频、所述中频和所述低频。
  8. 根据权利要求1所述的音频处理设备,其特征在于,所述音频处理电路还包括DAC模组,所述DAC模组工作时:
    接收所述K路声道的目标音频数据;
    将所述K路声道的目标音频数据转换为K路模拟电信号;以及
    将所述K路模拟电信号分别输入对应的发声单元。
  9. 根据权利要求1所述的音频处理设备,其特征在于,为了将所述K路声道的目标音频数据分别输入至对应的所述K个发声单元,所述音频处理电路:
    将每路声道的目标音频进行组合,得到整合音频数据;以及
    将所述整合音频数据分别输入至对应的所述K个发声单元。
  10. 根据权利要求9所述的音频处理设备,其特征在于,所述整合音频数据包括K段子数据,其中,第i段子数据包括第i路声道的目标音频数据和对应的第i个标识,所述i为[1,K]中的任意整数。
  11. 根据权利要求9所述的音频处理设备,其特征在于,所述音频处理电路还包括数字音频接口,工作时通过所述数字音频接口将所述整合音频数据分别输入至对应的所述K个发声单元,
    其中,所述K个发声单元中的每个所述发声单元包括识别电路和至少一个扬声器,当第i个识别电路工作时:
    接收到所述整合音频数据,
    在所述整合音频数据中识别出对应的第i个标识,并过滤掉其他标识对应的子数据,以及
    将所述第i个标识对应的目标音频数据转换为覆盖第i个频段的目标音频,并将所述目标音频发送至所述至少一个扬声器。
  12. 根据权利要求9所述的音频处理设备,其特征在于,所述音频处理电路还包括数字音频接口,所述音频数字接口在工作时:
    接收所述整合音频数据;
    在所述整合音频数据中识别出每个发声单元对应的标识和所述标识对应的目标音频数据;以及
    将所述目标音频数据发送至对应的发声单元。
  13. 根据权利要求1所述的音频处理设备,其特征在于,所述K个发声单元工作时以相同的相位同时播放对应的所述目标音频。
  14. 根据权利要求1所述的音频处理设备,其特征在于,所述音频处理设备为耳机。
  15. 根据权利要求1所述的音频处理设备,其特征在于,所述声道与所述发声单元之间对应关系包括一一对应的关系、多对一的关系或一对多的关系中的一种。
  16. 根据权利要求1所述的音频处理设备,其特征在于,在所述K个发声单元中至少一个发声单元发生变化或者接收到音效调整请求时,所述音频处理电路调整所述声道与所述发声单元之间的对应关系。
  17. 一种用于耳机的音频处理方法,其特征在于,包括,通过所述耳机的音频处理电路:
    获取初始音频数据;
    将所述初始音频数据转换为K路声道的目标音频数据,每路声道的目标音频数据适配所述耳机内对应的K个发声单元中一个发声单元的音频特性,其中K为大于1的整数,且每个所述发声单元的音频特性均不同;以及
    将所述K路声道的目标音频数据分别输入对应的所述K个发声单元,使得所述K个发声单元输出混响声音。
  18. 根据权利要求17所述的音频处理方法,其特征在于,所述将所述初始音频数据转换为K路声道的目标音频数据,每路声道的目标音频数据适配所述耳机内对应的K个发声单元中的一个发声单元的音频特性,包括:
    将所述初始音频数据复制成K路声道的所述初始音频数据;以及
    对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述耳机内对应的K发声单元中的一个发声单元的音频特性。
  19. 根据权利要求18所述的音频处理方法,其特征在于,
    所述初始音频数据包括K个频段;以及
    所述K个发生单元中的第i个发声单元在第i个频段具有期望的音响效果,所述i为[1,K]中的任意整数,
    其中,所述对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述耳机内对应的K个发声单元中的一个发声单元的音频特性,包括:
    对第i路声道的所述初始音频数据,保留或增强第i个频段上的幅值,衰减其他频段下的幅值,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述耳机内所述第i路声道对应的发声单元的音频特性。
  20. 根据权利要求19所述的音频处理方法,其特征在于,所述对每路声道的所述初始音频数据进行频谱调节,使得调节后的目标音频数据适配所述耳机内对应的K发声单元中的一个发声单元的音频特性,包括:
    对第i路声道的所述初始音频数据,保留第i个频段的音频数据,滤掉其他频段的音频数据,得到第i路声道的目标音频数据,所述第i路声道的目标音频数据适配所述耳机内所述第i路声道对应的发声单元的音频特性。
  21. 根据权利要求17所述的音频处理方法,其特征在于,所述将所述K路声道的目标音频数据分别输入对应的所述K个发声单元,使得所述K个发声单元输出混响声音,包括:
    将每路声道的目标音频数据转换为模拟电信号;以及
    将所述模拟电信号分别输入所述对应的发声单元,使每个发声单元将对应的模拟电信号转换为目标音频,形成混响声音。
  22. 根据权利要求17所述的音频处理方法,其特征在于,所述将所述K路声道的目标音频数据分别输入对应的所述K个发声单元,使得所述K个发声单元输出混响声音,包括:
    将每路声道的目标音频数据进行组合,得到整合音频数据;以及
    将所述整合音频数据分别输入至对应的发声单元,使得所述K个发声单元输出混响声音。
  23. 根据权利要求22所述的音频处理方法,其中,所述将所述整合音频数据分别输入至对应的发声单元,使得所述K个发声单元输出混响声音,包括:
    通过数字音频接口将所述整合音频数据分别输入至所述K个发声单元,使得所述K个发声单元输出混响声音。
PCT/CN2022/128860 2022-11-01 2022-11-01 音频处理设备和方法 WO2024092478A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/128860 WO2024092478A1 (zh) 2022-11-01 2022-11-01 音频处理设备和方法

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/128860 WO2024092478A1 (zh) 2022-11-01 2022-11-01 音频处理设备和方法

Publications (1)

Publication Number Publication Date
WO2024092478A1 true WO2024092478A1 (zh) 2024-05-10

Family

ID=90929070

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/128860 WO2024092478A1 (zh) 2022-11-01 2022-11-01 音频处理设备和方法

Country Status (1)

Country Link
WO (1) WO2024092478A1 (zh)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103974167A (zh) * 2013-02-06 2014-08-06 辉达公司 平板电子设备及其音频播放装置
CN108632714A (zh) * 2017-03-23 2018-10-09 展讯通信(上海)有限公司 扬声器的声音处理方法、装置及移动终端
US20180352365A1 (en) * 2016-11-12 2018-12-06 Ryan Ingebritsen SYSTEMS, Devices, AND METHODS FOR RECONFIGURING AND ROUTING A MULTICHANNEL AUDIO FILE
CN113507633A (zh) * 2021-05-26 2021-10-15 海信视像科技股份有限公司 一种声音数据处理方法和装置
CN115250397A (zh) * 2021-04-28 2022-10-28 华为技术有限公司 Tws耳机和tws耳机的播放方法及装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103974167A (zh) * 2013-02-06 2014-08-06 辉达公司 平板电子设备及其音频播放装置
US20180352365A1 (en) * 2016-11-12 2018-12-06 Ryan Ingebritsen SYSTEMS, Devices, AND METHODS FOR RECONFIGURING AND ROUTING A MULTICHANNEL AUDIO FILE
CN108632714A (zh) * 2017-03-23 2018-10-09 展讯通信(上海)有限公司 扬声器的声音处理方法、装置及移动终端
CN115250397A (zh) * 2021-04-28 2022-10-28 华为技术有限公司 Tws耳机和tws耳机的播放方法及装置
CN113507633A (zh) * 2021-05-26 2021-10-15 海信视像科技股份有限公司 一种声音数据处理方法和装置

Similar Documents

Publication Publication Date Title
JP5820806B2 (ja) スペクトル管理システム
WO2015102117A1 (ja) 音響ヘッドホンの触覚装置でオーディオ信号を再生するための方法および装置
JP2011010183A (ja) 楽曲再生システム、携帯端末装置および楽曲再生プログラム
US10104470B2 (en) Audio processing device, audio processing method, recording medium, and program
CN1658709A (zh) 声音再现设备和声音再现方法
WO2015062649A1 (en) Method and mobile device for processing an audio signal
CN115442709B (zh) 音频处理方法、虚拟低音增强系统、设备和存储介质
KR20200085226A (ko) 사용자-특정 및 하드웨어-특정 오디오 정보에 기반한 커스터마이즈화된 오디오 처리
JP2011128617A (ja) 信号処理装置及び方法
US10547927B1 (en) Systems and methods for processing an audio signal for replay on stereo and multi-channel audio devices
US11716065B2 (en) Systems and methods of volume limiting
TW202245483A (zh) 通過使用線性化及/或頻寬擴展產生第一控制信號和第二控制信號的裝置和方法
WO2024092478A1 (zh) 音频处理设备和方法
WO2024037189A9 (zh) 声像校准方法和装置
WO1999026454A1 (en) Low-frequency audio simulation system
EP3599775B1 (en) Systems and methods for processing an audio signal for replay on stereo and multi-channel audio devices
TWI501657B (zh) 電子音訊裝置
CN113645531B (zh) 一种耳机虚拟空间声回放方法、装置、存储介质及耳机
CN117998250A (zh) 音频处理设备和方法
CN211352437U (zh) 一种音频输出装置及电子设备
CN115804106A (zh) 声学输出装置和声学输出装置的控制方法
US20140376726A1 (en) Stereo headphone audio process
TWI828041B (zh) 用以控制包含差分信號的合成生成之聲音產生器的裝置及方法
US20210274300A1 (en) Acoustic radiation reproduction
CN213880236U (zh) 一种音效转换装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22963797

Country of ref document: EP

Kind code of ref document: A1