WO2020238000A1 - 一种音频处理方法、装置、终端及计算机可读存储介质 - Google Patents
一种音频处理方法、装置、终端及计算机可读存储介质 Download PDFInfo
- Publication number
- WO2020238000A1 WO2020238000A1 PCT/CN2019/113685 CN2019113685W WO2020238000A1 WO 2020238000 A1 WO2020238000 A1 WO 2020238000A1 CN 2019113685 W CN2019113685 W CN 2019113685W WO 2020238000 A1 WO2020238000 A1 WO 2020238000A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- audio
- matched
- frequency
- response curve
- frequency response
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 23
- 238000001228 spectrum Methods 0.000 claims abstract description 58
- 230000000694 effects Effects 0.000 claims abstract description 44
- 238000000034 method Methods 0.000 claims abstract description 35
- 230000004044 response Effects 0.000 claims description 98
- 230000003595 spectral effect Effects 0.000 claims description 20
- 238000004590 computer program Methods 0.000 claims description 12
- 230000003044 adaptive effect Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 4
- 238000010183 spectrum analysis Methods 0.000 description 4
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/04—Circuits for transducers, loudspeakers or microphones for correcting frequency response
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/26—Pre-filtering or post-filtering
-
- H—ELECTRICITY
- H03—ELECTRONIC CIRCUITRY
- H03G—CONTROL OF AMPLIFICATION
- H03G5/00—Tone control or bandwidth control in amplifiers
- H03G5/16—Automatic control
- H03G5/165—Equalizers; Volume or gain control in limited frequency bands
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2430/00—Signal processing covered by H04R, not provided for in its groups
Definitions
- the present invention relates to the field of audio processing technology, and in particular to an audio processing method, device, terminal and computer readable storage medium.
- the audio playback effect of the audio playback device can be adjusted by adjusting the equalizer. Specifically, in response to the user's adjustment operation on the equalizer, the corresponding parameters of the equalizer can be adjusted to adjust the audio playback effect of the audio playback device.
- the above adjustment method needs to listen to the audio playback effect of the audio playback device every time the equalizer is adjusted. If you are not satisfied with the audio playback effect, you need to adjust the equalizer again. It can be seen that the existing method of adjusting the audio playback effect takes time Longer and less efficient.
- the audio processing method, device, terminal, and computer-readable storage medium provided by the embodiments of the present invention can adopt matching rules from the audio to be matched to the reference audio to adaptively adjust the audio playback effect of the audio playback device, thereby improving audio playback The adjustment efficiency of the effect.
- an audio processing method including:
- target filter set as a matching rule from the audio to be matched to the reference audio
- an audio processing device including:
- An obtaining unit configured to obtain the audio to be matched and the reference audio, and obtain the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio;
- the determining unit is configured to determine the spectral difference between the to-be-matched audio and the reference audio according to the spectral distribution of the to-be-matched audio and the spectral distribution of the reference audio, and use the spectral difference to determine
- the target filter set of the audio to be matched matched to the reference audio, and the target filter set is used as a matching rule from the audio to be matched to the reference audio;
- a processing unit configured to compensate the audio playback device by using the matching rule to adjust the audio playback effect of the audio playback device
- the playback unit is used to play audio through the compensated audio playback device.
- an embodiment of the present invention provides a terminal including a processor and a memory, the processor and the memory are connected to each other, wherein the memory is used to store a computer program, and the computer program includes program instructions.
- the processor is configured to call the program instructions to execute the audio processing method as described in the first aspect.
- an embodiment of the present invention provides a computer-readable storage medium that stores a computer program, and the computer program includes program instructions that, when executed by a processor, cause all The processor executes the audio processing method as described in the first aspect.
- the terminal can determine the matching rule of the audio to be matched to the reference audio according to the spectral difference between the audio to be matched and the reference audio (that is, to match the audio to be matched to the target filter of the reference audio)
- the matching rule can be used to compensate the audio playback device to adjust the audio playback effect of the audio playback device, and play audio through the compensated audio playback device; this method realizes the audio playback effect of the audio playback device
- the adaptive adjustment process effectively improves the adjustment efficiency of the audio playback effect, so that the audio playback device can present a better audio playback effect.
- Figure 1 is a schematic diagram of an audio processing scenario provided by an embodiment of the present invention.
- FIG. 2 is a schematic diagram of an audio processing interface provided by an embodiment of the present invention.
- FIG. 3 is a schematic flowchart of an audio processing method provided by an embodiment of the present invention.
- FIG. 4 is a schematic diagram of an absolute error frequency response curve provided by an embodiment of the present invention.
- FIG. 5 is a schematic diagram of another absolute error frequency response curve provided by an embodiment of the present invention.
- FIG. 6 is a schematic structural diagram of an audio processing device provided by an embodiment of the present invention.
- FIG. 7 is a schematic structural diagram of a terminal provided by an embodiment of the present invention.
- an audio processing method can obtain a matching rule between the audio to be matched and the reference audio, and use the matching rule to compensate the audio playback device to adjust the audio.
- the audio playback effect of the playback device so as to play audio through the compensated audio playback device.
- the audio to be matched refers to the recorded audio played by the first audio playback device, and the audio playback effect of the first audio device is poor or non-standard.
- the reference audio refers to the recorded audio played by the second audio playback device, and the audio playback effect of the second audio playback device is better or more standard.
- the first audio playback device and the second audio playback device are audio playback devices of the same type.
- the audio played by the first audio playback device and the audio played by the second audio playback device correspond to the same sound source, for example, correspond to the same segment of the same song.
- the aforementioned compensated audio playback device and the first audio playback device may be the same audio playback device or not, and the audio playback effect of the audio playback device is poor or non-standard.
- the audio processing method may be applied to the terminal to adjust the audio playback effect of the connected audio playback device.
- the connection can be wired or wireless.
- FIG. 1 shows a smart phone 10 connected to a pair of earphones 20.
- the smart phone can use the matching rule to compensate the earphone 20, so that the earphone 20 can present a better audio playback effect.
- the audio processing method can also be applied to the adjustment of the audio playback effect of the audio playback device.
- the earphone 20 can use the matching rule to compensate itself, so as to present a better audio playback effect.
- the terminal may obtain the matching rule from the audio to be matched to the reference audio in any of the following ways.
- the terminal can calculate the matching rule in real time.
- the terminal obtaining the matching rule between the audio to be matched and the reference audio may include: the terminal obtains the audio to be matched and the reference audio; the terminal determines that the audio to be matched is to be matched according to the spectral difference between the audio to be matched and the reference audio The matching rule of the reference audio.
- the terminal may obtain the matching rule from the server.
- obtaining the matching rule from the audio to be matched to the reference audio by the terminal may include: the terminal sends a matching rule obtaining request to the server, and receiving the matching rule from the audio to be matched to the reference audio returned by the server according to the matching rule obtaining request.
- the server can also calculate the matching rule in the above-mentioned manner.
- the matching rule acquisition request may carry the to-be-matched audio and the reference audio, so that the server can calculate the matching rule from the to-be-matched audio to the reference audio according to the to-be-matched audio and the reference audio.
- the matching rule obtaining request is used to trigger the server to obtain the matching rule from the audio to be matched to the reference audio.
- the matching rule acquisition request may also carry the identifier of the audio playback device, so that the server can query the matching rule from the audio to be matched corresponding to the audio playback device to the reference audio according to the identifier of the audio playback device.
- the identification is the name and/or model.
- the terminal may calculate the matching rule in advance before acquiring the audio to be matched to the reference audio. Specifically, the terminal can obtain the audio to be matched and the reference audio before obtaining the audio to be matched to the reference audio, and determine the difference between the audio to be matched and the reference audio according to the spectral difference between the audio to be matched and the reference audio. Matching rules.
- the terminal can respond to the user's adjustment operation on the audio playback effect of the audio playback device, and obtain a matching rule from the audio to be matched to the reference audio.
- the adjustment operation of the audio playback effect may include a touch operation of an audio playback effect adjustment button or an audio playback device adaptation button, such as a click operation.
- Fig. 2 shows an audio processing interface 30, which includes an audio playback effect adjustment button 01.
- the terminal can respond to the user's click operation on the audio playback effect adjustment button 01 to execute the audio described in the embodiment of the present invention. Approach.
- the terminal can obtain the matching rule from the audio to be matched to the reference audio when detecting that the audio playing device is in the audio playing state. Or, when detecting that the audio playback device is in the audio playback state, the terminal can respond to the user's adjustment operation on the audio playback effect of the audio playback device, and obtain a matching rule from the audio to be matched to the reference audio.
- the terminal may also obtain the matching rule from the audio to be matched to the reference audio when other preset conditions are met, which are not listed here in this embodiment of the present invention.
- the embodiment of the present invention provides a schematic flowchart of an audio processing method in FIG. 3.
- the method described in the embodiment of the present invention may be implemented by a terminal, and the terminal may be a smart terminal or an audio playback device.
- the terminal includes, but is not limited to, smart phones, tablet computers, notebook computers, desktop computers, MP3, MP4 and other smart terminals.
- the audio playback device includes, but is not limited to, earphones, such as wired earphones, wireless earphones, such as Bluetooth/Wireless Fidelity (WIFI) earphones, wireless speakers, such as Bluetooth/WIFI speakers, and other devices that can be used for audio playback.
- WIFI Bluetooth/Wireless Fidelity
- earphones can also be divided into ordinary earphones and high-fidelity (High Fidelity, HIFI) earphones, which are not limited in the embodiment of the present invention.
- the method includes steps S301-S303:
- the terminal can obtain the matching rule from the audio to be matched to the reference audio.
- the terminal can obtain the audio to be matched and the reference audio, and determine the matching rule from the audio to be matched to the reference audio according to the spectrum difference between the audio to be matched and the reference audio.
- the terminal determines the matching rule from the to-be-matched audio to the reference audio according to the frequency spectrum difference between the to-be-matched audio and the reference audio, refer to steps S302-S305.
- the terminal in order to determine the matching rule from the audio to be matched to the reference audio, the terminal may obtain the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio.
- the terminal can perform spectrum analysis on the audio to be matched to obtain the spectrum distribution of the audio to be matched, and perform spectrum analysis on the reference audio to obtain the spectrum distribution of the reference audio.
- the methods used for spectrum analysis include but are not limited to any one of the following: periodogram method, autocorrelation method, Bartlett method, and Welch method.
- Spectrum distribution refers to the frequency distribution curve, such as frequency response curve, hereinafter referred to as frequency response curve.
- the frequency response curve may be an amplitude-frequency response curve.
- the abscissa of the frequency response curve represents the frequency point, hereinafter referred to as the frequency point
- the ordinate represents the amplitude value at the corresponding frequency point.
- the amplitude value in the embodiment of the present invention may also be referred to as an amplitude-frequency response value.
- the terminal can also perform spectrum analysis on the audio to be matched to obtain the frequency spectrum distribution of the audio to be matched, and download the frequency spectrum distribution of the reference audio from the server.
- the server stores the frequency spectrum distribution of the reference audio, which effectively saves the storage space of the terminal.
- the terminal can also download the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio from the server.
- the server stores the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio, which effectively saves the storage space of the terminal.
- S303 Determine a frequency spectrum difference between the audio to be matched and the reference audio according to the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio.
- the frequency spectrum difference refers to the difference between the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio.
- the spectral difference refers to an error frequency response curve between the audio to be matched and the reference audio, and the error frequency response curve is to measure the spectral distribution of the audio to be matched and the reference audio. Obtained after the difference operation.
- the error frequency response curve between the audio to be matched and the reference audio is Calculated by the following formula:
- S304 Determine a target filter set for matching the audio to be matched to the reference audio by using the frequency spectrum difference
- the target filter set may include one or more filters.
- the determining the target filter set for matching the audio to be matched to the reference audio by using the spectral difference may include: determining filter parameters according to the error frequency response curve;
- the filter parameter determines a target filter set for matching the audio to be matched to the reference audio from the filter set to be selected, the set of to be selected filters includes M filters, where M is A positive integer greater than or equal to 1.
- the filter parameters include cut-off frequency, gain value and quality factor.
- the candidate filter set includes any one or more of the following types of filters: a low-pass filter, a first-type band-pass filter, a second-type band-pass filter, and a high-pass filter.
- Each type of filter includes any one or more of the following orders: 1st order filter, 2nd order filter, 3rd order filter, 4th order filter.
- the low-pass filter includes a first-order low-pass filter, a second-order low-pass filter, a third-order low-pass filter, and a fourth-order low-pass filter.
- the determining the filter parameters according to the error frequency response curve may include: obtaining the maximum amplitude value and the target frequency point on the error frequency response curve, and the amplitude value corresponding to the target frequency point is One half of the maximum amplitude value, or the target frequency point is a preset frequency point; the maximum amplitude value and the target frequency point are used to determine filter parameters, and the filter parameters include cutoff frequency, Gain value and quality factor.
- the target frequency point may include a low-side frequency and a high-side frequency.
- the low-side frequency described in the embodiment of the present invention may be referred to as the low-side cut-off frequency
- the high-side frequency may be referred to as the high-side cut-off frequency.
- the maximum amplitude value on the error frequency response curve refers to the maximum amplitude value of the absolute error frequency response curve.
- the absolute error frequency response curve is obtained after taking the absolute value of the amplitude value of the error frequency response curve.
- the error frequency response curve is The absolute error frequency response curve is Calculated by the following formula:
- the target frequency can be found directly on the absolute error frequency response curve. Specifically, find the absolute error frequency response curve The frequency point corresponding to the maximum amplitude value of, and set the index number of the frequency point as m, and the absolute error frequency response curve The maximum amplitude value of is expressed as Frequency response curve from absolute error , Find the absolute error frequency response curve The maximum amplitude value point to the left of the amplitude value is The frequency point and the amplitude on the right are Frequency point and set the amplitude value The frequency point of is determined as the target frequency point.
- the low-side frequency is the maximum amplitude value and the left side amplitude value is Frequency point
- the high-side frequency is the maximum amplitude value and the right-side amplitude value is Frequency point. among them, It is one-half of the maximum amplitude value.
- the target frequency can be found indirectly through the absolute error frequency response curve.
- Frequency response curve due to absolute error Is composed of some discrete values, so there may be no way to determine the absolute error frequency response curve Directly find the target frequency point in, then use formula 1.3 to calculate the target frequency point.
- the absolute error frequency response curve To the left of the maximum amplitude point the amplitude frequency point FG k-1 corresponds to less than The amplitude frequency point FG k corresponding to the amplitude value is greater than The amplitude value on the left side of the maximum amplitude value point is calculated from these two amplitude frequency points Calculate the low-side frequency f l .
- f l is calculated by the following formula:
- f k represents the k-th frequency point, and f k is the frequency point corresponding to FG k , Is the amplitude value corresponding to FG k ;
- f k-1 represents the k-1 frequency point, f k-1 is the frequency point corresponding to FG k-1 , Is the amplitude value corresponding to FG k-1 .
- the absolute error frequency response curve The amplitude value on the right side of the maximum amplitude point is The frequency point of, that is, the high-side frequency f h , can also be obtained by calculation in a similar manner, which is not described in detail in the embodiment of the present invention.
- the preset frequency can be set as the target frequency.
- the preset frequency point may specifically include a low frequency threshold (denoted as f l_thr ) and a high frequency threshold (denoted as f h_thr ).
- f l_thr a low frequency threshold
- f h_thr a high frequency threshold
- the target frequency can be found directly or indirectly as shown in the upper left corner of the picture in Figure 5; or, it may also appear that the maximum value cannot be found on the left as shown in the upper right corner of the picture in Figure 5.
- the frequency point corresponding to one-half of the amplitude value; or, the frequency point corresponding to one-half of the maximum amplitude value may not be found on the right side of the picture shown in the lower left corner of Figure 5; or, also It may appear that the frequency point corresponding to one-half of the maximum amplitude value cannot be found on the left and right sides of the picture in the lower right corner of Figure 5.
- the low-side frequency can be set as the low-frequency threshold in the preset frequency point. As the set f l f l_thr.
- the high-side frequency can be set as the high-frequency threshold in the preset frequency point. For example, set f h to f h_thr .
- the low frequency threshold and the high frequency threshold are set according to empirical values.
- the maximum amplitude value and the target frequency point can be used to determine the filter parameters.
- the cutoff frequency can be calculated from the target frequency. Specifically, the product of the low-side frequency and the high-side frequency can be square rooted to obtain the cutoff frequency.
- the filter parameters can be used to determine a target filter set for matching the audio to be matched to the reference audio from the set of filters to be selected.
- using the filter parameters to determine a target filter set for matching the audio to be matched to the reference audio from the filter set to be selected may include: using the filter parameters to obtain N filters are determined from the M filters included in the filter set to be selected, where N is a positive integer less than or equal to the M; the frequency response curve of each filter in the N filters is obtained The average absolute error value between the error frequency response curves; the first filter corresponding to the smallest average absolute error value among the N filters is added to match the audio to be matched to the reference Audio target filter set.
- the determining N filters from the M filters included in the filter set to be selected by using the filter parameters may include: combining the cutoff frequency in the filter parameters with The preset frequency points are compared to obtain a comparison result, and N filters are determined from the M filters included in the filter set to be selected according to the comparison result.
- N filters including low-pass filters are determined from the M filters included in the filter set to be selected; when f c If f h_thr is greater than or equal to f h_thr , N filters including a high-pass filter are determined from M filters included in the filter set to be selected.
- the error frequency response curve may be updated for error judgment.
- the frequency response curve of the first filter can be used to update the error frequency response curve to obtain the updated error frequency response curve; if the updated error frequency response curve corresponds to the average absolute error value and the maximum If the absolute error value does not meet the preset conditions, the filter parameters are re-determined according to the updated error frequency response curve; the re-determined filter parameters are used to add filters to the target filter set until the updated error frequency response
- the average absolute error value and the maximum absolute error value corresponding to the curve satisfy the preset condition, and the preset condition includes that the average absolute error value is less than or equal to the first preset threshold, and the maximum absolute error value is less than or equal to the second preset Threshold, after meeting the preset condition, it means that the matching error of the audio to be matched to the reference audio has tended to converge stably, and the target filter set at this time can be used as the matching rule of the
- aver_value is the average absolute error value corresponding to the updated error frequency response curve
- aver_value_thr is the first preset threshold
- max_value is the maximum average absolute error value corresponding to the updated error frequency response curve
- max_value_thr is the second preset threshold .
- using the frequency response curve of the first filter to update the error frequency response curve to obtain the updated error frequency response curve may include: using the error frequency response curve and the first filter The frequency response curve performs difference calculation to obtain the updated error frequency response curve. For example, if the frequency response curve of the first filter is g ji , the updated error frequency response curve When the updated error frequency response curve When the corresponding aver_value and max_value do not meet the preset conditions, it needs to be based on the updated error frequency response curve Re-determine the filter parameters; use the re-determined filter parameters to add filters to the target filter set until the updated error frequency response curve The corresponding aver_value and max_value satisfy Equation 1.5.
- S306 Use the matching rule to compensate the audio playback device to adjust the audio playback effect of the audio playback device.
- the process of using the matching rule to compensate the audio playback device can be understood as the process of using the matching rule to equalize the audio played by the audio playback device.
- the embodiment of the present invention can obtain a target filter set that simulates the equalization process of the audio to be matched to the reference audio through an equalization matching algorithm, and then use the target filter set to compensate the audio playback device.
- the equalization matching algorithm can simulate the original equalization process through existing filters.
- the terminal may obtain the matching rule from the audio to be matched to the reference audio, and the matching rule is obtained according to the spectral difference between the audio to be matched and the reference audio;
- the playback device compensates to adjust the audio playback effect of the audio playback device; the compensated audio playback device plays the audio, thereby effectively improving the adjustment efficiency of the audio playback effect, so that the audio playback device can present better audio playback effect.
- FIG. 6 is a schematic structural diagram of an audio processing device according to an embodiment of the present invention.
- the device can be applied to terminals. Specifically, the device can run the following units:
- the obtaining unit 601 is configured to obtain the audio to be matched and the reference audio, and obtain the frequency spectrum distribution of the audio to be matched and the frequency spectrum distribution of the reference audio;
- the determining unit 602 is configured to determine the spectral difference between the to-be-matched audio and the reference audio according to the spectral distribution of the to-be-matched audio and the spectral distribution of the reference audio, and use the spectral difference to determine Matching the audio to be matched to a target filter set of the reference audio, and using the target filter set as a matching rule from the audio to be matched to the reference audio;
- the processing unit 603 is configured to compensate the audio playback device by using the matching rule to adjust the audio playback effect of the audio playback device;
- the playing unit 604 is used for playing audio through the compensated audio playing device.
- the frequency spectrum difference is an error frequency response curve between the audio to be matched and the reference audio
- the determining unit 602 uses the frequency spectrum difference to determine a method for matching the audio to be matched to the
- the target filter set of the reference audio is specifically determining filter parameters according to the error frequency response curve; using the filter parameters to determine from the set of filters to be selected for matching the audio to be matched to the reference
- An audio target filter set, the candidate filter set includes M filters, and the M is a positive integer greater than or equal to 1.
- the determining unit 602 determines the filter parameters according to the error frequency response curve, specifically obtaining the maximum amplitude value and the target frequency point on the error frequency response curve, and the amplitude value corresponding to the target frequency point Is one half of the maximum amplitude value, or the target frequency point is a preset frequency point; the maximum amplitude value and the target frequency point are used to determine filter parameters, and the filter parameters include a cutoff frequency , Gain value and quality factor.
- the determining unit 602 uses the filter parameters to determine a target filter set for matching the audio to be matched to the reference audio from the filter set to be selected, specifically using the The filter parameters determine N filters from the M filters included in the set of filters to be selected, where N is a positive integer less than or equal to the M; and obtain the value of each of the N filters The average absolute error value between the frequency response curve and the error frequency response curve; adding the first filter corresponding to the smallest average absolute error value among the N filters to be used for matching the audio to be matched Into the target filter set of the reference audio.
- the determining unit 602 is further configured to use the frequency response curve of the first filter to update the error frequency response curve to obtain the updated error frequency response curve; if the updated error frequency response curve is The average absolute error value and the maximum absolute error value corresponding to the response curve do not meet the preset conditions, the filter parameters are re-determined according to the updated error frequency response curve; the re-determined filter parameters are used to add to the target filter set Filter until the average absolute error value and the maximum absolute error value corresponding to the updated error frequency response curve meet the preset condition, and the preset condition includes that the average absolute error value is less than or equal to the first preset threshold, and the maximum The absolute error value is less than or equal to the second preset threshold.
- steps S301-S307 involved in the audio processing method shown in FIG. 3 may be executed by the units in the embodiment in FIG. 6.
- step S301 and step S302 shown in FIG. 3 can be executed by the acquiring unit 601 shown in FIG. 6
- steps S303-S305 can be executed by the determining unit 602 shown in FIG. 6
- step S306 and step S307 can be respectively processed by Unit 603 and playing unit 604.
- the units in the audio processing device shown in FIG. 6 can be separately or completely combined into one or several additional units to form, or some of the units can be disassembled. It is divided into multiple functionally smaller units to form, which can realize the same operation without affecting the realization of the technical effect of the embodiment of the present invention.
- the above-mentioned units are divided based on logical functions. In practical applications, the function of one unit may also be realized by multiple units, or the functions of multiple units may be realized by one unit. In other embodiments of the present invention, the image processing apparatus may also include other units. In practical applications, these functions may also be implemented with the assistance of other units, and may be implemented by multiple units in cooperation.
- a general-purpose computing device such as a computer including a central processing unit (CPU), a random access storage medium (RAM), a read-only storage medium (ROM) and other processing elements and storage elements
- CPU central processing unit
- RAM random access storage medium
- ROM read-only storage medium
- Run a computer program capable of executing the steps involved in the audio processing method as shown in FIG. 3 to construct the audio processing device as shown in FIG. 6 and to implement the audio processing method of the embodiment of the present invention .
- the computer program can be recorded on, for example, a computer-readable recording medium, and loaded into the aforementioned computing device via the computer-readable recording medium, and run in it.
- the audio processing device may obtain the matching rule from the audio to be matched to the reference audio, the matching rule is obtained based on the spectral difference between the audio to be matched and the reference audio; the matching rule is used Compensate the audio playback device to adjust the audio playback effect of the audio playback device; play audio through the compensated audio playback device, thereby effectively improving the adjustment efficiency of the audio playback effect, so that the audio playback device can present better Audio playback effect.
- an embodiment of the present invention also provides a terminal.
- the internal structure of the terminal includes at least one or more processors 701 and a memory 702.
- the processor 701 and the memory 702 in the terminal may be connected through a bus or in other ways.
- the connection through a bus is taken as an example.
- the internal structure of the terminal may further include one or more input devices 703 and one or more output devices 704.
- the processor 701, the storage device 702, the input device 703, and the output device 704 are connected by a bus or other means.
- the connection by a bus is taken as an example.
- the memory 702 is configured to store a computer program including program instructions
- the processor 701 is configured to execute the program instructions stored in the memory 702.
- the processor 701 loads and executes one or more instructions stored in the computer-readable storage medium to implement the corresponding steps of the method in the above-mentioned corresponding embodiment; in specific implementation, the computer-readable storage medium At least one instruction in is loaded by the processor 701 and executes the following steps:
- target filter set as a matching rule from the audio to be matched to the reference audio
- the frequency spectrum difference is an error frequency response curve between the audio to be matched and the reference audio. Accordingly, the at least one program instruction is loaded by the processor 701 and is also used to execute:
- the filter parameters are used to determine a target filter set for matching the audio to be matched to the reference audio from a set of to-be-selected filters, the set of to-be-selected filters includes M filters, and M is a positive integer greater than or equal to 1.
- the at least one program instruction is loaded by the processor 701 and is also used to execute:
- the amplitude value corresponding to the target frequency point is one-half of the maximum amplitude value, or the target frequency point is a preset frequency point ;
- the filter parameters are determined by using the maximum amplitude value and the target frequency point, and the filter parameters include a cut-off frequency, a gain value and a quality factor.
- the at least one program instruction is loaded by the processor 701 and is also used to execute:
- N filters from M filters included in the filter set to be selected by using the filter parameters, where N is a positive integer less than or equal to the M;
- the first filter corresponding to the smallest average absolute error value among the N filters is added to the target filter set for matching the audio to be matched to the reference audio.
- the at least one program instruction is loaded by the processor 701 and is also used to execute:
- N filters are determined from the M filters included in the filter set to be selected.
- the at least one program instruction is loaded by the processor 701 and is also used to execute:
- the filter parameters are re-determined according to the updated error frequency response curve
- the preset frequency point includes a low frequency threshold and a high frequency threshold.
- the processor 701 may be a central processing unit (Central Processing Unit, CPU), and the processor may also be another general-purpose processor, that is, a microprocessor or any conventional processor.
- the memory 702 may include a read-only memory and a random access memory, and provides instructions and data to the processor 701. Therefore, the processor 701 and the memory 702 are not limited here.
- the input device 703 may include an audio collection device, such as a recorder, or may also include a touch screen.
- the output device 704 may include an audio playback device, such as a speaker, or may also include a display screen. Among them, the touch screen and the display screen can be integrated into a touch display screen.
- the input device 703 and the output device 704 may also include standard wired interfaces and/or wireless interfaces.
- the program can be stored in a computer-readable storage medium. When executed, it may include the processes of the above-mentioned method embodiments.
- the storage medium may be a magnetic disk, an optical disc, a read-only memory (Read-Only Memory, ROM), or a random access memory (Random Access Memory, RAM), etc.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Acoustics & Sound (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
Description
Claims (10)
- 一种音频处理方法,其特征在于,包括:获取待匹配音频和参考音频;获取所述待匹配音频的频谱分布以及所述参考音频的频谱分布;根据所述待匹配音频的频谱分布以及所述参考音频的频谱分布确定所述待匹配音频和所述参考音频之间的频谱差异;利用所述频谱差异确定出用于将所述待匹配音频匹配到所述参考音频的目标滤波器集合;将所述目标滤波器集合作为所述待匹配音频到所述参考音频的匹配规则;利用所述匹配规则对音频播放设备进行补偿,以调节所述音频播放设备的音频播放效果;通过补偿后的音频播放设备播放音频。
- 根据权利要求1所述的方法,其特征在于,所述频谱差异为所述待匹配音频和所述参考音频之间的误差频响曲线,所述利用频谱差异确定出用于将所述待匹配音频匹配到所述参考音频的目标滤波器集合,包括:根据所述误差频响曲线确定滤波器参数;利用所述滤波器参数从待选滤波器集合中确定出用于将所述待匹配音频匹配到所述参考音频的目标滤波器集合,所述待选滤波器集合包括M个滤波器,所述M为大于或等于1的正整数。
- 根据权利要求2所述的方法,其特征在于,所述根据所述误差频响曲线确定滤波器参数,包括:获取所述误差频响曲线上的最大幅度值和目标频点,所述目标频点对应的幅度值为所述最大幅度值的二分之一,或者,所述目标频点为预设频点;利用所述最大幅度值和所述目标频点确定滤波器参数,所述滤波器参数包括截止频率、增益值和品质因数。
- 根据权利要求3所述的方法,其特征在于,所述利用所述滤波器参数 从待选滤波器集合中确定出用于将所述待匹配音频匹配到所述参考音频的目标滤波器集合,包括:利用所述滤波器参数从待选滤波器集合包括的M个滤波器中确定出N个滤波器,所述N为小于或等于所述M的正整数;获取所述N个滤波器中每一个滤波器的频响曲线与所述误差频响曲线之间的平均绝对误差值;将所述N个滤波器中对应的平均绝对误差值最小的第一滤波器,添加到用于将所述待匹配音频匹配到所述参考音频的目标滤波器集合中。
- 根据权利要求1所述的方法,其特征在于,所述利用所述滤波器参数从待选滤波器集合包括的M个滤波器中确定出N个滤波器,包括:将所述滤波器参数中的截止频率与所述预设频点进行比较,得到比较结果;根据所述比较结果从待选滤波器集合包括的M个滤波器中确定出N个滤波器。
- 根据权利要求4或5所述的方法,其特征在于,所述方法还包括:利用所述第一滤波器的频响曲线对所述误差频响曲线进行更新,得到更新后的误差频响曲线;若更新后的误差频响曲线对应的平均绝对误差值和最大绝对误差值不满足预设条件,则根据更新后的误差频响曲线重新确定滤波器参数;利用重新确定的滤波器参数向所述目标滤波器集合中添加滤波器,直到更新后的误差频响曲线对应的平均绝对误差值和最大绝对误差值满足所述预设条件,所述预设条件包括平均绝对误差值小于或等于第一预设阈值,且最大绝对误差值小于或等于第二预设阈值。
- 根据权利要求3-5任一项所述的方法,其特征在于,所述预设频点包括低频阈值和高频阈值。
- 一种音频处理装置,其特征在于,包括:获取单元,用于获取待匹配音频和参考音频,并获取所述待匹配音频的频谱分布以及所述参考音频的频谱分布;确定单元,用于根据所述待匹配音频的频谱分布以及所述参考音频的频谱分布确定所述待匹配音频和所述参考音频之间的频谱差异,并利用所述频谱差异确定出用于将所述待匹配音频匹配到所述参考音频的目标滤波器集合,将所述目标滤波器集合作为所述待匹配音频到所述参考音频的匹配规则;处理单元,用于利用所述匹配规则对音频播放设备进行补偿,以调节所述音频播放设备的音频播放效果;播放单元,用于通过补偿后的音频播放设备播放音频。
- 一种终端,其特征在于,包括处理器和存储器,所述处理器和存储器相互连接,其中,所述存储器用于存储计算机程序,所述计算机程序包括程序指令,所述处理器被配置用于调用所述程序指令,执行如权利要求1~7任意一项权利要求所述的音频处理方法。
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质存储有计算机程序,所述计算机程序包括程序指令,所述程序指令当被处理器执行时使所述处理器执行如权利要求1~7任意一项所述的音频处理方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/629,357 US12058499B2 (en) | 2019-05-24 | 2019-10-28 | Audio processing method and device, terminal, and computer-readable storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910439413.2A CN110191396B (zh) | 2019-05-24 | 2019-05-24 | 一种音频处理方法、装置、终端及计算机可读存储介质 |
CN201910439413.2 | 2019-05-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020238000A1 true WO2020238000A1 (zh) | 2020-12-03 |
Family
ID=67717801
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/113685 WO2020238000A1 (zh) | 2019-05-24 | 2019-10-28 | 一种音频处理方法、装置、终端及计算机可读存储介质 |
Country Status (3)
Country | Link |
---|---|
US (1) | US12058499B2 (zh) |
CN (1) | CN110191396B (zh) |
WO (1) | WO2020238000A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112769410A (zh) * | 2020-12-25 | 2021-05-07 | 西安讯飞超脑信息科技有限公司 | 滤波器构建方法、音频处理方法及电子设备、存储装置 |
CN113949968A (zh) * | 2021-09-07 | 2022-01-18 | 万魔声学股份有限公司 | 一种频响校正方法、电子设备及信号处理方法 |
CN117241170A (zh) * | 2023-11-16 | 2023-12-15 | 武汉海微科技有限公司 | 基于二分频音箱的音频播放方法、装置、设备及存储介质 |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110191396B (zh) | 2019-05-24 | 2022-05-27 | 腾讯音乐娱乐科技(深圳)有限公司 | 一种音频处理方法、装置、终端及计算机可读存储介质 |
CN112767971B (zh) * | 2019-10-21 | 2022-02-01 | 华为技术有限公司 | 音频播放方法、模型获取方法、装置、终端及服务器 |
CN111933161B (zh) * | 2020-07-16 | 2024-09-03 | 腾讯音乐娱乐科技(深圳)有限公司 | 均衡器滤波参数的生成方法、音频信号滤波方法及均衡器 |
CN112905833A (zh) * | 2021-01-19 | 2021-06-04 | 腾讯音乐娱乐科技(深圳)有限公司 | 一种音频重放设备预热方法、装置、设备及介质 |
US20240212704A1 (en) * | 2021-09-22 | 2024-06-27 | Boe Technology Group Co., Ltd. | Audio adjusting method, device and apparatus, and storage medium |
CN115331683A (zh) * | 2022-08-08 | 2022-11-11 | 北京达佳互联信息技术有限公司 | 音频处理方法、装置、电子设备及存储介质 |
CN118053436A (zh) * | 2022-11-15 | 2024-05-17 | 抖音视界有限公司 | 音频处理方法、装置及电子设备 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150243271A1 (en) * | 2014-02-22 | 2015-08-27 | Apple Inc. | Active noise control with compensation for acoustic leak in personal listening devices |
CN106601268A (zh) * | 2016-12-26 | 2017-04-26 | 腾讯音乐娱乐(深圳)有限公司 | 一种多媒体数据处理方法及装置 |
CN109658942A (zh) * | 2018-12-27 | 2019-04-19 | 腾讯音乐娱乐科技(深圳)有限公司 | 一种音频数据处理方法、装置以及相关设备 |
CN110191396A (zh) * | 2019-05-24 | 2019-08-30 | 腾讯音乐娱乐科技(深圳)有限公司 | 一种音频处理方法、装置、终端及计算机可读存储介质 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4773094A (en) * | 1985-12-23 | 1988-09-20 | Dolby Ray Milton | Apparatus and method for calibrating recording and transmission systems |
US20040013272A1 (en) * | 2001-09-07 | 2004-01-22 | Reams Robert W | System and method for processing audio data |
JP3932957B2 (ja) * | 2001-05-17 | 2007-06-20 | 三菱電機株式会社 | クラッタ抑圧装置およびクラッタ抑圧方法 |
JP3869823B2 (ja) * | 2003-08-06 | 2007-01-17 | 株式会社国際電気通信基礎技術研究所 | 音声の周波数特性の等化装置 |
US8565449B2 (en) * | 2006-02-07 | 2013-10-22 | Bongiovi Acoustics Llc. | System and method for digital signal processing |
CN101048935B (zh) * | 2004-10-26 | 2011-03-23 | 杜比实验室特许公司 | 控制音频信号的单位响度或部分单位响度的方法和设备 |
KR100636213B1 (ko) * | 2004-12-28 | 2006-10-19 | 삼성전자주식회사 | 실시간 주파수 특성 보정 방법 및 그를 적용한 사운드재생 장치 |
EP1722360B1 (en) * | 2005-05-13 | 2014-03-19 | Harman Becker Automotive Systems GmbH | Audio enhancement system and method |
US8566086B2 (en) * | 2005-06-28 | 2013-10-22 | Qnx Software Systems Limited | System for adaptive enhancement of speech signals |
US8077880B2 (en) * | 2007-05-11 | 2011-12-13 | Audyssey Laboratories, Inc. | Combined multirate-based and fir-based filtering technique for room acoustic equalization |
CA2679953C (en) * | 2007-06-19 | 2014-01-21 | Dolby Laboratories Licensing Corporation | Loudness measurement with spectral modifications |
US8301676B2 (en) * | 2007-08-23 | 2012-10-30 | Fisher-Rosemount Systems, Inc. | Field device with capability of calculating digital filter coefficients |
US8965756B2 (en) * | 2011-03-14 | 2015-02-24 | Adobe Systems Incorporated | Automatic equalization of coloration in speech recordings |
EP2770635A1 (en) * | 2013-02-25 | 2014-08-27 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Equalization filter coefficient determinator, apparatus, equalization filter coefficient processor, system and methods |
CN103987001A (zh) * | 2014-05-28 | 2014-08-13 | 深圳市金立通信设备有限公司 | 一种音频修正的方法及装置 |
CN104661153B (zh) * | 2014-12-31 | 2018-02-02 | 歌尔股份有限公司 | 一种耳机音效补偿方法、装置及耳机 |
CN104811155B (zh) * | 2015-04-20 | 2017-08-04 | 深圳市冠旭电子股份有限公司 | 一种均衡器调整方法及装置 |
CN105792072B (zh) * | 2016-03-25 | 2020-10-09 | 腾讯科技(深圳)有限公司 | 一种音效处理方法、装置及终端 |
CN107992711B (zh) * | 2018-01-18 | 2021-04-13 | 桂林电子科技大学 | M通道过采样调制图滤波器组的优化设计方法 |
-
2019
- 2019-05-24 CN CN201910439413.2A patent/CN110191396B/zh active Active
- 2019-10-28 US US17/629,357 patent/US12058499B2/en active Active
- 2019-10-28 WO PCT/CN2019/113685 patent/WO2020238000A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150243271A1 (en) * | 2014-02-22 | 2015-08-27 | Apple Inc. | Active noise control with compensation for acoustic leak in personal listening devices |
CN106601268A (zh) * | 2016-12-26 | 2017-04-26 | 腾讯音乐娱乐(深圳)有限公司 | 一种多媒体数据处理方法及装置 |
CN109658942A (zh) * | 2018-12-27 | 2019-04-19 | 腾讯音乐娱乐科技(深圳)有限公司 | 一种音频数据处理方法、装置以及相关设备 |
CN110191396A (zh) * | 2019-05-24 | 2019-08-30 | 腾讯音乐娱乐科技(深圳)有限公司 | 一种音频处理方法、装置、终端及计算机可读存储介质 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112769410A (zh) * | 2020-12-25 | 2021-05-07 | 西安讯飞超脑信息科技有限公司 | 滤波器构建方法、音频处理方法及电子设备、存储装置 |
CN112769410B (zh) * | 2020-12-25 | 2024-06-11 | 西安讯飞超脑信息科技有限公司 | 滤波器构建方法、音频处理方法及电子设备、存储装置 |
CN113949968A (zh) * | 2021-09-07 | 2022-01-18 | 万魔声学股份有限公司 | 一种频响校正方法、电子设备及信号处理方法 |
CN117241170A (zh) * | 2023-11-16 | 2023-12-15 | 武汉海微科技有限公司 | 基于二分频音箱的音频播放方法、装置、设备及存储介质 |
CN117241170B (zh) * | 2023-11-16 | 2024-01-19 | 武汉海微科技有限公司 | 基于二分频音箱的音频播放方法、装置、设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
US12058499B2 (en) | 2024-08-06 |
CN110191396B (zh) | 2022-05-27 |
CN110191396A (zh) | 2019-08-30 |
US20220345817A1 (en) | 2022-10-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020238000A1 (zh) | 一种音频处理方法、装置、终端及计算机可读存储介质 | |
US9131298B2 (en) | Constrained dynamic amplitude panning in collaborative sound systems | |
US10638245B2 (en) | Dynamic multi-speaker optimization | |
KR20210020751A (ko) | 복수의 소비자 장치에 개인화된 오디오 재생 기능을 제공하기 위한 시스템 및 방법 | |
CN110580141B (zh) | 移动终端 | |
CN109686347A (zh) | 音效处理方法、音效处理装置、电子设备和可读介质 | |
CN108365827A (zh) | 具有动态阈值的频带压缩 | |
CN111415673A (zh) | 基于用户特定和硬件特定音频信息的定制的音频处理 | |
US10812031B2 (en) | Electronic device and method for adjusting gain of digital audio signal based on hearing recognition characteristics | |
JP2010500806A (ja) | オーディオ信号を処理する装置及び方法 | |
WO2023284438A1 (zh) | 音频数据处理方法和装置、电子设备 | |
WO2023071425A1 (zh) | 音量控制方法及装置、电子设备及存储介质 | |
WO2021203906A1 (zh) | 自动音量调整方法、装置、介质和设备 | |
US20240163623A1 (en) | In-sync digital waveform comparison to determine pass/fail results of a device under test (dut) | |
WO2020097824A1 (zh) | 音频处理方法、装置、存储介质及电子设备 | |
US10405114B2 (en) | Automated detection of an active audio output | |
US20240015438A1 (en) | Managing low frequencies of an output signal | |
WO2021042538A1 (zh) | 一种音频处理方法、装置及计算机存储介质 | |
CN116156413A (zh) | 音频播放方法、装置、电子设备和存储介质 | |
US10902864B2 (en) | Mixed-reality audio intelligibility control | |
WO2021103262A1 (zh) | 耳机的控制方法、耳机及可读存储介质 | |
WO2020107385A1 (zh) | 增益处理方法及其装置、电子设备、信号采集方法及其系统 | |
US9197958B2 (en) | Method and apparatus for defining audio signals for headphones | |
WO2020107192A1 (zh) | 立体声播放方法、装置、存储介质及电子设备 | |
WO2024036957A1 (zh) | 一种音频数据处理方法及装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19930788 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19930788 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 22.03.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19930788 Country of ref document: EP Kind code of ref document: A1 |