US20170294890A1 - System and method for dynamic equalization of audio data - Google Patents
System and method for dynamic equalization of audio data Download PDFInfo
- Publication number
- US20170294890A1 US20170294890A1 US15/634,586 US201715634586A US2017294890A1 US 20170294890 A1 US20170294890 A1 US 20170294890A1 US 201715634586 A US201715634586 A US 201715634586A US 2017294890 A1 US2017294890 A1 US 2017294890A1
- Authority
- US
- United States
- Prior art keywords
- audio data
- signal
- coupled
- generate
- audio
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims description 17
- 230000001052 transient effect Effects 0.000 claims abstract description 11
- 230000005236 sound signal Effects 0.000 claims description 28
- 230000008878 coupling Effects 0.000 claims 3
- 238000010168 coupling process Methods 0.000 claims 3
- 238000005859 coupling reaction Methods 0.000 claims 3
- 238000010586 diagram Methods 0.000 description 12
- 230000008569 process Effects 0.000 description 8
- 230000000873 masking effect Effects 0.000 description 7
- 230000008901 benefit Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 230000003044 adaptive effect Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 210000002985 organ of corti Anatomy 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 210000005036 nerve Anatomy 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 208000016354 hearing loss disease Diseases 0.000 description 1
- 238000003032 molecular docking Methods 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
-
- H—ELECTRICITY
- H03—ELECTRONIC CIRCUITRY
- H03G—CONTROL OF AMPLIFICATION
- H03G5/00—Tone control or bandwidth control in amplifiers
- H03G5/16—Automatic control
- H03G5/165—Equalizers; Volume or gain control in limited frequency bands
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/02—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
- G10L19/028—Noise substitution, i.e. substituting non-tonal spectral components by noisy source
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
-
- G10L21/0205—
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0316—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
- G10L21/0324—Details of processing therefor
- G10L21/034—Automatic adjustment
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0316—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude
- G10L21/0364—Speech enhancement, e.g. noise reduction or echo cancellation by changing the amplitude for improving intelligibility
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
- G10L25/06—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being correlation coefficients
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/52—Automatic gain control
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/60—Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/35—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using translation techniques
- H04R25/356—Amplitude, e.g. amplitude shift or compression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/50—Customised settings for obtaining desired overall acoustical characteristics
- H04R25/505—Customised settings for obtaining desired overall acoustical characteristics using digital signal processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/12—Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
- H04R3/14—Cross-over networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
- H04S3/02—Systems employing more than two channels, e.g. quadraphonic of the matrix type, i.e. in which input signals are combined algebraically, e.g. after having been phase shifted with respect to each other
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2225/00—Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
- H04R2225/43—Signal processing in hearing aids to enhance the speech intelligibility
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2420/00—Details of connection covered by H04R, not provided for in its groups
- H04R2420/01—Input selection or mixing for amplifiers or loudspeakers
Definitions
- the present disclosure relates generally to audio data processing, and more specifically to a system and method for dynamic equalization of audio data that reduces audio energy processing consumption.
- Equalization of audio data is used to control the relative gain of frequency components of the audio data, such as to boost low frequency components, middle frequency components or high frequency components.
- a system for processing audio data includes a plurality of gain adjustment devices, each gain adjustment device having an associated audio input frequency band.
- a plurality of control signal processing systems are configured to receive audio input data for one of the associated audio input frequency bands and to generate a gain adjustment device control signal.
- the gain adjustment device control signal is configured to decrease a gain setting of an associated gain adjustment device for a predetermined period of time as a function of a transient in the associated audio input frequency band.
- FIG. 1 is a diagram of a system for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure
- FIG. 2 is a diagram of a system for controlling dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure
- FIG. 3 is a diagram of an algorithm for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure
- FIG. 4 is a diagram of a system for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure.
- FIG. 5 is a diagram of an algorithm for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure.
- Audio data can include events that are louder than other events, such as gunshots, cymbal crashes, drum beats and so forth. When these events occur, they mask audio data that is 13 dB lower in gain for a period of time (typically around 200 milliseconds), such as audio data that has the same frequency components as the frequency components of the event. This masking occurs as a result of the psychoacoustic processes related to hearing. However, even though the masked audio signals cannot be perceived, the nerve cells in the organ of Corti are still receiving the masked audio signals, and are using energy to process them. This additional energy use results in a loss of hearing sensitivity. As such, the audio processing system that amplifies such signals is not only wasting energy on amplification of signals that are not perceived by the listener, it is also wasting that energy to create an inferior listening experience.
- the amount of energy consumed by the audio processing system can be reduced, which can result in longer battery life.
- the effect of such masked audio signals on the nerves in the organ of Corti can be reduced or eliminated, which results in an improved audio experience for the listener.
- FIG. 1 is a diagram of a system 100 for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure.
- System 100 can be implemented in hardware or a suitable combination of hardware and software.
- “hardware” can include a combination of discrete components, an integrated circuit, an application-specific integrated circuit, a field programmable gate array, or other suitable hardware.
- “software” can include one or more objects, agents, threads, lines of code, subroutines, separate software applications, two or more lines of code or other suitable software structures operating in two or more software applications, on one or more processors (where a processor includes a microcomputer or other suitable controller, memory devices, input-output devices, displays, data input devices such as a keyboard or a mouse, peripherals such as printers and speakers, associated drivers, control cards, power sources, network devices, docking station devices, or other suitable devices operating under control of software systems in conjunction with the processor or other devices), or other suitable software structures.
- software can include one or more lines of code or other suitable software structures operating in a general purpose software application, such as an operating system, and one or more lines of code or other suitable software structures operating in a specific purpose software application.
- the term “couple” and its cognate terms, such as “couples” and “coupled,” can include a physical connection (such as a copper conductor), a virtual connection (such as through randomly assigned memory locations of a data memory device), a logical connection (such as through logical gates of a semiconducting device), other suitable connections, or a suitable combination of such connections.
- System 100 includes crossover 102 , which receives audio data and processes the audio data to generate separate frequency bands of audio data.
- crossover 102 can generate a first band having a frequency range of 0-50 Hz, a second band having a frequency range of 50-500 Hz, a third band having a frequency range of 500-4500 Hz and a fourth band having a frequency range of 4500 Hz and above, or other suitable numbers of bands and associated frequency ranges can also or alternatively be used.
- the input to crossover 102 can be an unprocessed audio signal, a normalized audio signal or other suitable audio signals.
- crossover 102 The outputs of crossover 102 are further filtered using associated filters, such as low pass filter 104 , low mid pass filter 106 , mid pass filter 108 and high pass filter 110 , or other suitable filters.
- the high frequency band can be further processed to add harmonic components, such as to compensate for lossy compression processing of the audio data that can result in audio data having a narrow image width and sparse frequency components.
- the harmonic components can be added using clipping circuit 112 , which generates harmonic components by clipping the high frequency components of the audio data.
- High pass filter 114 is used to remove lower frequency harmonic components
- scaler 116 is used to control the magnitude of the harmonic processed audio that is added to the unprocessed audio at adder 138 .
- Control of scaler 116 is provided by crossover 118 , which can generate a high frequency band output for frequencies above a predetermined level, such as 8000 Hz, and a low frequency band output for frequencies below the predetermined level.
- the RMS values of the high and low frequency bands are generated by RMS processors 120 and 122 , and the RMS values are then converted from linear values to log values by DB20 124 and DB20 126 , respectively.
- the difference between the high and low frequency components is then determined using subtractor 128 , and a value from table 130 is used to determine the amount of high frequency harmonic frequency component signal to be added to the unprocessed high frequency audio signal.
- the amount can be set to zero until there is a 6 dB difference between the low and high frequency components, and as the difference increases from 6 dB to 10 dB, the amount of high frequency harmonic frequency component signal that is added to the unprocessed high frequency audio signal can increase from 0 dB to 8 dB. As the difference between the low and high frequency components increases from 10 dB to 15 dB, the amount of high frequency harmonic frequency component signal that is added to the unprocessed high frequency audio signal can increase from 8 dB to 9 dB. Likewise, other suitable amounts of high frequency harmonic frequency component signal can be added to the unprocessed high frequency audio signal.
- Increasing the amount of high frequency harmonic frequency component signal that is added to the unprocessed high frequency audio signal as a function of the change in the relative content of low and high frequency components of the high frequency band can be used to improve audio quality, because the difference is indicative of a sparse audio signal.
- the additional harmonic content helps to improve a sparse audio signal by providing additional frequency components that are complementary to the audio data.
- the high frequency harmonic components are then added to the unprocessed high frequency components by adder 138 .
- Equalization of the low pass frequency component is accomplished using scaler 132 under control of an input A
- equalization of the mid pass frequency component is accomplished using scaler 136 under control of an input B
- equalization of the high pass frequency component is accomplished using scaler 140 under control of an input C.
- the outputs of the equalized audio components and the unprocessed low-mid pass output 106 are combined using adder 142 to generate an output.
- system 100 performs dynamic equalization to reduce power loss and also improves audio signal quality.
- the psychoacoustic masking processes result in a 150 to 200 millisecond loss in perception when a masking input having a crest of 13 dB or more is generated, due to the reaction of kinocilia to such audio inputs.
- When such transients occur then maintain or increasing audio during the dead zone that follows the transient only serves to increase the power consumed by the audio processing system without increasing audio quality.
- processing of that audio energy still requires work to be done by kinocilia in the organ of Corti, and can also increase the amount of additional energy that is required in order to generate a perceptible response.
- system 100 helps to reduce both the amount of energy required to process the audio data as well as the amount of energy required by the listener to listen to the audio data.
- adding harmonic frequency content to the high frequency audio data when the total audio data content is sparse helps to improve the perceived audio quality, by providing additional frequency components in the sparse audio data that complement the existing frequency components.
- FIG. 2 is a diagram of a system 200 for controlling dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure.
- System 200 can be implemented in hardware or a suitable combination of hardware and software.
- System 200 includes automatic gain control core 202 and automatic gain control multiplier 204 , which are configured to receive an audio signal input and to generate a normalized audio signal output.
- the normalized audio signal output of AGC multiplier 204 can also be provided to crossover 102 or other suitable systems or components.
- Filter 206 can be a band pass filter having a frequency range of 40 to 80 Hz or other suitable filters.
- the output from filter 206 is processed by RMS processor 208 to generate a signal that represents the RMS value of the output of filter 206 .
- Derivative processor 210 receives the band pass RMS value and generates an output that represents the rate of change of the band pass RMS value.
- Downward expander 212 is used to prevent dynamic equalization of the associated frequency band when there is no associated transient occurring in the frequency band.
- Filter 214 can be a band pass filter having a frequency range of 500 to 4000 Hz or other suitable filters.
- the output from filter 214 is processed by RMS processor 216 to generate a signal that represents the RMS value of the output of filter 214 .
- Derivative processor 218 receives the band pass RMS value and generates an output that represents the rate of change of the band pass RMS value.
- Downward expander 220 is used to prevent dynamic equalization of the associated frequency band when there is no associated transient occurring in the frequency band.
- Filter 222 can be a high pass filter having a frequency range of 4000 Hz and above or other suitable filters.
- the output from filter 222 is processed by RMS processor 224 to generate a signal that represents the RMS value of the output of filter 222 .
- Derivative processor 226 receives the band pass RMS value and generates an output that represents the rate of change of the band pass RMS value.
- Downward expander 228 is used to prevent dynamic equalization of the associated frequency band when there is no associated transient occurring in the frequency band.
- system 200 In operation, system 200 generates control inputs for a dynamic equalizer, by detecting transients in frequency bands associated with the dynamic equalization. System 200 thus helps to improve power consumption for an audio data processor, and also helps to improve the perceptual audio quality.
- FIG. 3 is a diagram of an algorithm 300 for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure.
- Algorithm 300 can be implemented in hardware or a suitable combination of hardware and software. Although algorithm 300 is shown as a flow chart, other suitable programming paradigms such as state diagrams and object-oriented programming can also or alternatively be used.
- Algorithm 300 begins at 302 , where audio data is received and processed, such as to generate a normalized audio signal by using a first adaptive gain control processor that is used to remove a DC signal component and a second adaptive gain control processor that receives the output of the first adaptive gain control processor and the input audio, or in other suitable manners.
- the algorithm then proceeds to 304 , where the audio data is filtered to generate different bands of audio data.
- the algorithm then proceeds in parallel to 306 and 314 .
- the high and low frequency components of one or more of the bands of audio data are analyzed, such as to determine whether there is a greater RMS value of one component compared to the other.
- the algorithm then proceeds to 308 , where it is determined whether the difference is indicative of an audio signal that is sparse or that otherwise would benefit from additional harmonic content, such as by exceeding a predetermined level. If the difference does not exceed the level, the algorithm proceeds to 318 where the unprocessed signal is dynamically equalized, otherwise the algorithm proceeds to 310 , where harmonic content is generated.
- harmonic content can be generated by clipping the audio signal and then filtering the clipped signal to remove predetermined harmonic frequency components, or in other suitable manners. The algorithm then proceeds to 312 , where the harmonic content is added to unprocessed audio signal and the combined signal is processed at 318 .
- the audio signals are processed to determine whether a transient has occurred, such as by generating a derivative of an RMS value of the frequency component or in other suitable manners.
- a downward expander or other suitable components or processes can be used to ensure that the control signal is only generated for significant transients that will cause an associated psychoacoustic masking of predetermined associated audio frequency components.
- the algorithm then proceeds to 316 , where a control signal is generated based on the transient.
- the control signal is applied to the audio signal to perform dynamic equalization at 318 , such as to reduce the gain of the audio signal frequency components when a masking transient occurs, to both reduce power consumption and listener fatigue, and to improve audio quality to the listener.
- algorithm 300 allows audio data to be dynamically equalized, by detecting masking transients and by using such masking transients to generate dynamic equalization controls. By dynamically equalizing the audio data that would otherwise not be perceptible to the listener, the amount of energy required to process the audio data can be reduced, and the perceived quality of the audio data can be improved.
- FIG. 4 is a diagram of a system 400 for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure.
- System 400 can be implemented in hardware or a suitable combination of hardware and software.
- System 400 includes time to frequency conversion system 402 , which converts frames of a time-varying audio signal into frames of frequency components, such as by performing a fast-Fourier transform or in other suitable manners.
- Bin comparison system 404 receives the frames of frequency domain data and compares the magnitude of the left channel audio data with the magnitude of the right channel audio data for each frequency bin.
- Phase adjustment system 406 receives the comparison data for each bin of frequency data from bin comparison system 404 and sets the phase of the right channel frequency bin component equal to the phase of the left channel frequency bin component if the magnitude of the left channel frequency bin component is greater than the magnitude of the right channel frequency bin component.
- the output of phase adjustment system 406 is parametric audio data.
- Surround processing system 408 receives the parametric audio data and generates surround audio data.
- surround processing system 408 can receive speaker location data and can calculate a phase angle difference for the input audio data that corresponds to the location of the speaker.
- surround processing system 408 can generate audio data for any suitable number of speakers in any suitable locations, by adjusting the phase angle of the parametric audio data to reflect the speaker location relative to other speakers.
- system 400 allows audio data to be processed to generate parametric audio data, which can then be processed based on predetermined speaker locations to generate N-dimensional audio data.
- System 400 eliminates the phase data of the input audio data, which is not needed when the input audio data is processed to be output from speakers in non-stereophonic speaker locations.
- FIG. 5 is a diagram of an algorithm 500 for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure.
- Algorithm 500 can be implemented in hardware or a suitable combination of hardware and software. Although algorithm 500 is shown as a flow chart, other suitable programming paradigms such as state diagrams and object-oriented programming can also or alternatively be used.
- Algorithm 500 begins at 502 where audio data is received, such as analog or digital audio data in the time domain. The algorithm then proceeds to 504 .
- the audio data is converted from the time domain to the frequency domain, such as by performing a fast Fourier transform on the audio data or in other suitable manners.
- the algorithm then proceeds to 506 .
- 506 it is determined whether the magnitude of a left channel frequency component of the frequency domain audio data is greater than the magnitude of the associated right channel frequency component.
- 506 can be performed on a frequency component basis for each of the frequency components of the audio data, or in other suitable manners. If it is determined that the magnitude of a left channel frequency component of the frequency domain audio data is not greater than the magnitude of the associated right channel frequency component, the algorithm proceeds to 510 , otherwise the algorithm proceeds to 508 , where the phase of the right channel frequency component is replaced with the phase of the left channel frequency component. The algorithm then proceeds to 510 .
- the audio data is processed for an N-channel surround playback environment.
- the locations of each of a plurality of speakers can be input into a system which can then determine a preferred phase relationship of the left and right channel audio data for that speaker.
- the phase and magnitude of the audio data can then be generated as a function of the speaker location, or other suitable processes can also or alternatively be used.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Quality & Reliability (AREA)
- Otolaryngology (AREA)
- General Health & Medical Sciences (AREA)
- Neurosurgery (AREA)
- Mathematical Physics (AREA)
- Pure & Applied Mathematics (AREA)
- Mathematical Optimization (AREA)
- Mathematical Analysis (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Algebra (AREA)
- Circuit For Audible Band Transducer (AREA)
- Stereophonic System (AREA)
- Tone Control, Compression And Expansion, Limiting Amplitude (AREA)
- Television Receiver Circuits (AREA)
Abstract
A system for processing audio data is disclosed that includes a plurality of gain adjustment devices, each gain adjustment device having an associated audio input frequency band. A plurality of control signal processing systems are configured to receive audio input data for one of the associated audio input frequency bands and to generate a gain adjustment device control signal. The gain adjustment device control signal is configured to decrease a gain setting of an associated gain adjustment device for a predetermined period of time as a function of a transient in the associated audio input frequency band.
Description
- The present application claims priority to and benefit of U.S. Provisional Patent Application No. 62/092,603, filed on Dec. 16, 2014, U.S. Provisional Patent Application No. 62/133,167, filed on Mar. 13, 2015, U.S. Provisional Patent Application No. 62/156,061, filed on May 1, 2015, and U.S. Provisional Patent Application No. 62/156,065, filed on May 1, 2015, each of which are hereby incorporated by reference for all purposes as if set forth herein in their entirety.
- The present disclosure relates generally to audio data processing, and more specifically to a system and method for dynamic equalization of audio data that reduces audio energy processing consumption.
- Equalization of audio data is used to control the relative gain of frequency components of the audio data, such as to boost low frequency components, middle frequency components or high frequency components.
- A system for processing audio data is disclosed that includes a plurality of gain adjustment devices, each gain adjustment device having an associated audio input frequency band. A plurality of control signal processing systems are configured to receive audio input data for one of the associated audio input frequency bands and to generate a gain adjustment device control signal. The gain adjustment device control signal is configured to decrease a gain setting of an associated gain adjustment device for a predetermined period of time as a function of a transient in the associated audio input frequency band.
- Other systems, methods, features, and advantages of the present disclosure will be or become apparent to one with skill in the art upon examination of the following drawings and detailed description. It is intended that all such additional systems, methods, features, and advantages be included within this description, be within the scope of the present disclosure, and be protected by the accompanying claims.
- Aspects of the disclosure can be better understood with reference to the following drawings. The components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present disclosure. Moreover, in the drawings, like reference numerals designate corresponding parts throughout the several views, and in which:
-
FIG. 1 is a diagram of a system for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure; -
FIG. 2 is a diagram of a system for controlling dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure; -
FIG. 3 is a diagram of an algorithm for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure; -
FIG. 4 is a diagram of a system for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure; and -
FIG. 5 is a diagram of an algorithm for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure. - In the description that follows, like parts are marked throughout the specification and drawings with the same reference numerals. The drawing figures might not be to scale and certain components can be shown in generalized or schematic form and identified by commercial designations in the interest of clarity and conciseness.
- Audio data can include events that are louder than other events, such as gunshots, cymbal crashes, drum beats and so forth. When these events occur, they mask audio data that is 13 dB lower in gain for a period of time (typically around 200 milliseconds), such as audio data that has the same frequency components as the frequency components of the event. This masking occurs as a result of the psychoacoustic processes related to hearing. However, even though the masked audio signals cannot be perceived, the nerve cells in the organ of Corti are still receiving the masked audio signals, and are using energy to process them. This additional energy use results in a loss of hearing sensitivity. As such, the audio processing system that amplifies such signals is not only wasting energy on amplification of signals that are not perceived by the listener, it is also wasting that energy to create an inferior listening experience.
- By detecting such transient events and dynamically equalizing the audio data to reduce the audio signals that will be masked, the amount of energy consumed by the audio processing system can be reduced, which can result in longer battery life. In addition, the effect of such masked audio signals on the nerves in the organ of Corti can be reduced or eliminated, which results in an improved audio experience for the listener.
-
FIG. 1 is a diagram of asystem 100 for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure.System 100 can be implemented in hardware or a suitable combination of hardware and software. - As used herein, “hardware” can include a combination of discrete components, an integrated circuit, an application-specific integrated circuit, a field programmable gate array, or other suitable hardware. As used herein, “software” can include one or more objects, agents, threads, lines of code, subroutines, separate software applications, two or more lines of code or other suitable software structures operating in two or more software applications, on one or more processors (where a processor includes a microcomputer or other suitable controller, memory devices, input-output devices, displays, data input devices such as a keyboard or a mouse, peripherals such as printers and speakers, associated drivers, control cards, power sources, network devices, docking station devices, or other suitable devices operating under control of software systems in conjunction with the processor or other devices), or other suitable software structures. In one exemplary embodiment, software can include one or more lines of code or other suitable software structures operating in a general purpose software application, such as an operating system, and one or more lines of code or other suitable software structures operating in a specific purpose software application. As used herein, the term “couple” and its cognate terms, such as “couples” and “coupled,” can include a physical connection (such as a copper conductor), a virtual connection (such as through randomly assigned memory locations of a data memory device), a logical connection (such as through logical gates of a semiconducting device), other suitable connections, or a suitable combination of such connections.
-
System 100 includescrossover 102, which receives audio data and processes the audio data to generate separate frequency bands of audio data. In one exemplary embodiment,crossover 102 can generate a first band having a frequency range of 0-50 Hz, a second band having a frequency range of 50-500 Hz, a third band having a frequency range of 500-4500 Hz and a fourth band having a frequency range of 4500 Hz and above, or other suitable numbers of bands and associated frequency ranges can also or alternatively be used. The input tocrossover 102 can be an unprocessed audio signal, a normalized audio signal or other suitable audio signals. - The outputs of
crossover 102 are further filtered using associated filters, such aslow pass filter 104, low mid pass filter 106,mid pass filter 108 andhigh pass filter 110, or other suitable filters. In addition, the high frequency band can be further processed to add harmonic components, such as to compensate for lossy compression processing of the audio data that can result in audio data having a narrow image width and sparse frequency components. In one exemplary embodiment, the harmonic components can be added usingclipping circuit 112, which generates harmonic components by clipping the high frequency components of the audio data.High pass filter 114 is used to remove lower frequency harmonic components, andscaler 116 is used to control the magnitude of the harmonic processed audio that is added to the unprocessed audio atadder 138. Control ofscaler 116 is provided bycrossover 118, which can generate a high frequency band output for frequencies above a predetermined level, such as 8000 Hz, and a low frequency band output for frequencies below the predetermined level. The RMS values of the high and low frequency bands are generated byRMS processors subtractor 128, and a value from table 130 is used to determine the amount of high frequency harmonic frequency component signal to be added to the unprocessed high frequency audio signal. In one exemplary embodiment, the amount can be set to zero until there is a 6 dB difference between the low and high frequency components, and as the difference increases from 6 dB to 10 dB, the amount of high frequency harmonic frequency component signal that is added to the unprocessed high frequency audio signal can increase from 0 dB to 8 dB. As the difference between the low and high frequency components increases from 10 dB to 15 dB, the amount of high frequency harmonic frequency component signal that is added to the unprocessed high frequency audio signal can increase from 8 dB to 9 dB. Likewise, other suitable amounts of high frequency harmonic frequency component signal can be added to the unprocessed high frequency audio signal. Increasing the amount of high frequency harmonic frequency component signal that is added to the unprocessed high frequency audio signal as a function of the change in the relative content of low and high frequency components of the high frequency band can be used to improve audio quality, because the difference is indicative of a sparse audio signal. The additional harmonic content helps to improve a sparse audio signal by providing additional frequency components that are complementary to the audio data. The high frequency harmonic components are then added to the unprocessed high frequency components byadder 138. - Equalization of the low pass frequency component is accomplished using
scaler 132 under control of an input A, equalization of the mid pass frequency component is accomplished usingscaler 136 under control of an input B, and equalization of the high pass frequency component is accomplished usingscaler 140 under control of an input C. The outputs of the equalized audio components and the unprocessed low-mid pass output 106 are combined usingadder 142 to generate an output. - In operation,
system 100 performs dynamic equalization to reduce power loss and also improves audio signal quality. The psychoacoustic masking processes result in a 150 to 200 millisecond loss in perception when a masking input having a crest of 13 dB or more is generated, due to the reaction of kinocilia to such audio inputs. When such transients occur, then maintain or increasing audio during the dead zone that follows the transient only serves to increase the power consumed by the audio processing system without increasing audio quality. In addition, while the audio input is not resulting in nerve signals that ultimately reach the listener's brain, processing of that audio energy still requires work to be done by kinocilia in the organ of Corti, and can also increase the amount of additional energy that is required in order to generate a perceptible response. By dynamically equalizing the audio data to reduce the gain during such periods,system 100 helps to reduce both the amount of energy required to process the audio data as well as the amount of energy required by the listener to listen to the audio data. In addition, adding harmonic frequency content to the high frequency audio data when the total audio data content is sparse helps to improve the perceived audio quality, by providing additional frequency components in the sparse audio data that complement the existing frequency components. -
FIG. 2 is a diagram of asystem 200 for controlling dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure. System 200 can be implemented in hardware or a suitable combination of hardware and software. -
System 200 includes automatic gain control core 202 and automaticgain control multiplier 204, which are configured to receive an audio signal input and to generate a normalized audio signal output. The normalized audio signal output ofAGC multiplier 204 can also be provided tocrossover 102 or other suitable systems or components. -
Filter 206 can be a band pass filter having a frequency range of 40 to 80 Hz or other suitable filters. The output fromfilter 206 is processed byRMS processor 208 to generate a signal that represents the RMS value of the output offilter 206.Derivative processor 210 receives the band pass RMS value and generates an output that represents the rate of change of the band pass RMS value.Downward expander 212 is used to prevent dynamic equalization of the associated frequency band when there is no associated transient occurring in the frequency band. -
Filter 214 can be a band pass filter having a frequency range of 500 to 4000 Hz or other suitable filters. The output fromfilter 214 is processed byRMS processor 216 to generate a signal that represents the RMS value of the output offilter 214.Derivative processor 218 receives the band pass RMS value and generates an output that represents the rate of change of the band pass RMS value.Downward expander 220 is used to prevent dynamic equalization of the associated frequency band when there is no associated transient occurring in the frequency band. -
Filter 222 can be a high pass filter having a frequency range of 4000 Hz and above or other suitable filters. The output fromfilter 222 is processed byRMS processor 224 to generate a signal that represents the RMS value of the output offilter 222.Derivative processor 226 receives the band pass RMS value and generates an output that represents the rate of change of the band pass RMS value.Downward expander 228 is used to prevent dynamic equalization of the associated frequency band when there is no associated transient occurring in the frequency band. - In operation,
system 200 generates control inputs for a dynamic equalizer, by detecting transients in frequency bands associated with the dynamic equalization.System 200 thus helps to improve power consumption for an audio data processor, and also helps to improve the perceptual audio quality. -
FIG. 3 is a diagram of analgorithm 300 for dynamic equalization of audio data, in accordance with an exemplary embodiment of the present disclosure.Algorithm 300 can be implemented in hardware or a suitable combination of hardware and software. Althoughalgorithm 300 is shown as a flow chart, other suitable programming paradigms such as state diagrams and object-oriented programming can also or alternatively be used. -
Algorithm 300 begins at 302, where audio data is received and processed, such as to generate a normalized audio signal by using a first adaptive gain control processor that is used to remove a DC signal component and a second adaptive gain control processor that receives the output of the first adaptive gain control processor and the input audio, or in other suitable manners. The algorithm then proceeds to 304, where the audio data is filtered to generate different bands of audio data. The algorithm then proceeds in parallel to 306 and 314. - At 306, the high and low frequency components of one or more of the bands of audio data are analyzed, such as to determine whether there is a greater RMS value of one component compared to the other. The algorithm then proceeds to 308, where it is determined whether the difference is indicative of an audio signal that is sparse or that otherwise would benefit from additional harmonic content, such as by exceeding a predetermined level. If the difference does not exceed the level, the algorithm proceeds to 318 where the unprocessed signal is dynamically equalized, otherwise the algorithm proceeds to 310, where harmonic content is generated. In one exemplary embodiment, harmonic content can be generated by clipping the audio signal and then filtering the clipped signal to remove predetermined harmonic frequency components, or in other suitable manners. The algorithm then proceeds to 312, where the harmonic content is added to unprocessed audio signal and the combined signal is processed at 318.
- At 314, the audio signals are processed to determine whether a transient has occurred, such as by generating a derivative of an RMS value of the frequency component or in other suitable manners. A downward expander or other suitable components or processes can be used to ensure that the control signal is only generated for significant transients that will cause an associated psychoacoustic masking of predetermined associated audio frequency components. The algorithm then proceeds to 316, where a control signal is generated based on the transient. The control signal is applied to the audio signal to perform dynamic equalization at 318, such as to reduce the gain of the audio signal frequency components when a masking transient occurs, to both reduce power consumption and listener fatigue, and to improve audio quality to the listener.
- In operation,
algorithm 300 allows audio data to be dynamically equalized, by detecting masking transients and by using such masking transients to generate dynamic equalization controls. By dynamically equalizing the audio data that would otherwise not be perceptible to the listener, the amount of energy required to process the audio data can be reduced, and the perceived quality of the audio data can be improved. -
FIG. 4 is a diagram of asystem 400 for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure.System 400 can be implemented in hardware or a suitable combination of hardware and software. -
System 400 includes time tofrequency conversion system 402, which converts frames of a time-varying audio signal into frames of frequency components, such as by performing a fast-Fourier transform or in other suitable manners. -
Bin comparison system 404 receives the frames of frequency domain data and compares the magnitude of the left channel audio data with the magnitude of the right channel audio data for each frequency bin. -
Phase adjustment system 406 receives the comparison data for each bin of frequency data frombin comparison system 404 and sets the phase of the right channel frequency bin component equal to the phase of the left channel frequency bin component if the magnitude of the left channel frequency bin component is greater than the magnitude of the right channel frequency bin component. The output ofphase adjustment system 406 is parametric audio data. -
Surround processing system 408 receives the parametric audio data and generates surround audio data. In one exemplary embodiment,surround processing system 408 can receive speaker location data and can calculate a phase angle difference for the input audio data that corresponds to the location of the speaker. In this exemplary embodiment,surround processing system 408 can generate audio data for any suitable number of speakers in any suitable locations, by adjusting the phase angle of the parametric audio data to reflect the speaker location relative to other speakers. - In operation,
system 400 allows audio data to be processed to generate parametric audio data, which can then be processed based on predetermined speaker locations to generate N-dimensional audio data.System 400 eliminates the phase data of the input audio data, which is not needed when the input audio data is processed to be output from speakers in non-stereophonic speaker locations. -
FIG. 5 is a diagram of analgorithm 500 for parametric stereo processing, in accordance with an exemplary embodiment of the present disclosure.Algorithm 500 can be implemented in hardware or a suitable combination of hardware and software. Althoughalgorithm 500 is shown as a flow chart, other suitable programming paradigms such as state diagrams and object-oriented programming can also or alternatively be used. -
Algorithm 500 begins at 502 where audio data is received, such as analog or digital audio data in the time domain. The algorithm then proceeds to 504. - At 504, the audio data is converted from the time domain to the frequency domain, such as by performing a fast Fourier transform on the audio data or in other suitable manners. The algorithm then proceeds to 506.
- At 506, it is determined whether the magnitude of a left channel frequency component of the frequency domain audio data is greater than the magnitude of the associated right channel frequency component. In one exemplary embodiment, 506 can be performed on a frequency component basis for each of the frequency components of the audio data, or in other suitable manners. If it is determined that the magnitude of a left channel frequency component of the frequency domain audio data is not greater than the magnitude of the associated right channel frequency component, the algorithm proceeds to 510, otherwise the algorithm proceeds to 508, where the phase of the right channel frequency component is replaced with the phase of the left channel frequency component. The algorithm then proceeds to 510.
- At 510, the audio data is processed for an N-channel surround playback environment. In one exemplary embodiment, the locations of each of a plurality of speakers can be input into a system which can then determine a preferred phase relationship of the left and right channel audio data for that speaker. The phase and magnitude of the audio data can then be generated as a function of the speaker location, or other suitable processes can also or alternatively be used.
- It should be emphasized that the above-described embodiments are merely examples of possible implementations. Many variations and modifications may be made to the above- described embodiments without departing from the principles of the present disclosure. All such modifications and variations are intended to be included herein within the scope of this disclosure and protected by the following claims.
Claims (18)
1-19. (canceled)
20. A system for processing audio data comprising:
a clipping system configured to receive audio input data and to generate a clipped audio input signal;
a high pass filter coupled to the clipping system and configured to receive the clipped audio input signal and to generate high-pass filtered clipped audio data;
a scaler multiplier coupled to the high pass filter and configured to multiply the high-pass filtered clipped audio data by a predetermined value; and
a table coupled to the scaler multiplier and configured to receive a difference signal and to select the predetermined value as a function of the difference signal for use in generating an audio output signal.
21. The system of claim 20 further comprising one or more gain adjustment devices, each gain adjustment device having an associated audio input frequency band, wherein a gain adjustment device control signal is configured to decrease a gain setting of an associated gain adjustment device for a predetermined period of time as a function of a transient in the associated audio input frequency band.
22. The system of claim 20 further comprising a plurality of control signal processing systems, each control signal processing system configured to receive audio input data for an associated audio input frequency band and to generate a gain adjustment device control signal for use in generating the audio output signal.
23. The system of claim 20 further comprising a crossover filter coupled to a subtractor, the crossover filter configured to generate a low frequency output and a high frequency output and the subtractor configured to receive the low frequency output and the high frequency output and to generate the difference signal.
24. The system of claim 23 further comprising an RMS converter coupled between the crossover filter and the subtractor.
25. The system of claim 23 further comprising a linear to log converter coupled between the crossover filter and the subtractor.
26. The system of claim 20 further comprising an adder coupled to the table and one of the gain adjustment devices.
27. A method for processing audio data comprising:
generating a clipped audio input signal from audio input data with a clipping system;
generating high-pass filtered clipped audio data from the clipped audio input signal using a high pass filter coupled to the clipping system;
multiplying the high-pass filtered clipped audio data by a predetermined value using a scaler multiplier coupled to the high pass filter; and
receiving a difference signal at a table coupled to the scaler multiplier and selecting the predetermined value as a function of the difference signal to generate an output audio signal.
28. The method of claim 27 further comprising:
generating a low frequency output and a high frequency output with a crossover filter coupled to the subtractor; and
receiving the low frequency output and the high frequency output at the subtractor and generating the difference signal.
29. The method of claim 28 further comprising coupling an RMS converter between the crossover filter and the subtractor.
30. The method of claim 28 further comprising coupling a linear to log converter between the crossover filter and the subtractor.
31. The method of claim 27 further comprising coupling an adder to the table and one of the gain adjustment devices.
32. A non-transitory computer-readable medium encoded with computer-executable instructions that when executed by one or more processors cause the processor to:
generate a clipped audio input signal from audio input data with a clipping system;
generate high-pass filtered clipped audio data from the clipped audio input signal using a high pass filter coupled to the clipping system;
multiply the high-pass filtered clipped audio data by a predetermined value using a scaler multiplier coupled to the high pass filter; and
receive a difference signal at a table coupled to the scaler multiplier and select the predetermined value as a function of the difference signal to generate an output audio signal.
32. The non-transitory computer-readable medium of claim 31 wherein the instructions cause the processor to:
generate a low frequency output and a high frequency output with a crossover filter coupled to a subtractor; and
receive the low frequency output and the high frequency output at the subtractor and generate the difference signal.
33. The non-transitory computer-readable medium of claim 31 wherein the instructions cause the processor to couple an RMS converter between a crossover filter and a subtractor.
34. The non-transitory computer-readable medium of claim 31 wherein the instructions cause the processor to couple a linear to log converter between a crossover filter and a subtractor.
35. The non-transitory computer-readable medium of claim 31 wherein the instructions cause the processor to couple an adder to the table and a gain adjustment device.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/634,586 US20170294890A1 (en) | 2014-12-16 | 2017-06-27 | System and method for dynamic equalization of audio data |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462092603P | 2014-12-16 | 2014-12-16 | |
US201562133167P | 2015-03-13 | 2015-03-13 | |
US201562156065P | 2015-05-01 | 2015-05-01 | |
US201562156061P | 2015-05-01 | 2015-05-01 | |
US14/863,357 US9691408B2 (en) | 2014-12-16 | 2015-09-23 | System and method for dynamic equalization of audio data |
US15/634,586 US20170294890A1 (en) | 2014-12-16 | 2017-06-27 | System and method for dynamic equalization of audio data |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/863,357 Continuation US9691408B2 (en) | 2014-12-16 | 2015-09-23 | System and method for dynamic equalization of audio data |
Publications (1)
Publication Number | Publication Date |
---|---|
US20170294890A1 true US20170294890A1 (en) | 2017-10-12 |
Family
ID=56111772
Family Applications (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/863,365 Active US9875756B2 (en) | 2014-12-16 | 2015-09-23 | System and method for artifact masking |
US14/863,350 Active US9852744B2 (en) | 2014-12-16 | 2015-09-23 | System and method for dynamic recovery of audio data |
US14/863,357 Active US9691408B2 (en) | 2014-12-16 | 2015-09-23 | System and method for dynamic equalization of audio data |
US14/863,368 Abandoned US20160171987A1 (en) | 2014-12-16 | 2015-09-23 | System and method for compressed audio enhancement |
US14/863,374 Abandoned US20160173808A1 (en) | 2014-12-16 | 2015-09-23 | System and method for level control at a receiver |
US14/863,376 Active US9830927B2 (en) | 2014-12-16 | 2015-09-23 | System and method for decorrelating audio data |
US15/634,586 Abandoned US20170294890A1 (en) | 2014-12-16 | 2017-06-27 | System and method for dynamic equalization of audio data |
Family Applications Before (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/863,365 Active US9875756B2 (en) | 2014-12-16 | 2015-09-23 | System and method for artifact masking |
US14/863,350 Active US9852744B2 (en) | 2014-12-16 | 2015-09-23 | System and method for dynamic recovery of audio data |
US14/863,357 Active US9691408B2 (en) | 2014-12-16 | 2015-09-23 | System and method for dynamic equalization of audio data |
US14/863,368 Abandoned US20160171987A1 (en) | 2014-12-16 | 2015-09-23 | System and method for compressed audio enhancement |
US14/863,374 Abandoned US20160173808A1 (en) | 2014-12-16 | 2015-09-23 | System and method for level control at a receiver |
US14/863,376 Active US9830927B2 (en) | 2014-12-16 | 2015-09-23 | System and method for decorrelating audio data |
Country Status (2)
Country | Link |
---|---|
US (7) | US9875756B2 (en) |
WO (1) | WO2016100422A1 (en) |
Families Citing this family (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9875756B2 (en) * | 2014-12-16 | 2018-01-23 | Psyx Research, Inc. | System and method for artifact masking |
EP3171362B1 (en) * | 2015-11-19 | 2019-08-28 | Harman Becker Automotive Systems GmbH | Bass enhancement and separation of an audio signal into a harmonic and transient signal component |
EP3185589B1 (en) * | 2015-12-22 | 2024-02-07 | Oticon A/s | A hearing device comprising a microphone control system |
US10374566B2 (en) * | 2016-07-29 | 2019-08-06 | Maxim Integrated Products, Inc. | Perceptual power reduction system and method |
CN106060746A (en) * | 2016-08-18 | 2016-10-26 | 佛山博智医疗科技有限公司 | Portable player with multichannel hearing compensation function |
JP2019062514A (en) | 2016-12-26 | 2019-04-18 | キヤノン株式会社 | Audio processing apparatus and method of controlling same |
US10121489B1 (en) * | 2017-07-21 | 2018-11-06 | Htc Corporation | Method, device, and non-transitory computer readable storage medium for processing audio signal |
US10530318B2 (en) * | 2017-11-30 | 2020-01-07 | Apple Inc. | Audio system having variable reset volume |
CN108389590B (en) * | 2018-02-06 | 2021-08-03 | 广东电网有限责任公司惠州供电局 | Time-frequency joint voice top cutting detection method |
US10728667B2 (en) * | 2018-10-19 | 2020-07-28 | Ernest Michael Poholek | Active channel crossover system |
US11019301B2 (en) | 2019-06-25 | 2021-05-25 | The Nielsen Company (Us), Llc | Methods and apparatus to perform an automated gain control protocol with an amplifier based on historical data corresponding to contextual data |
US11133787B2 (en) | 2019-06-25 | 2021-09-28 | The Nielsen Company (Us), Llc | Methods and apparatus to determine automated gain control parameters for an automated gain control protocol |
US11340704B2 (en) | 2019-08-21 | 2022-05-24 | Subpac, Inc. | Tactile audio enhancement |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4052571A (en) * | 1975-11-07 | 1977-10-04 | National Research Development Corporation | Hearing aid with amplitude compression achieved by clipping a modulated signal |
US20040170283A1 (en) * | 2002-03-12 | 2004-09-02 | Yasuhiro Terada | Howling control device and howling control method |
US20100023321A1 (en) * | 2008-07-25 | 2010-01-28 | Yamaha Corporation | Voice processing apparatus and method |
US20130010972A1 (en) * | 2011-07-04 | 2013-01-10 | Gn Resound A/S | Binaural compressor preserving directional cues |
Family Cites Families (83)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4243840A (en) * | 1978-12-22 | 1981-01-06 | Teledyne Industries, Inc. | Loudspeaker system |
US4922537A (en) * | 1987-06-02 | 1990-05-01 | Frederiksen & Shu Laboratories, Inc. | Method and apparatus employing audio frequency offset extraction and floating-point conversion for digitally encoding and decoding high-fidelity audio signals |
US5953049A (en) * | 1996-08-02 | 1999-09-14 | Lucent Technologies Inc. | Adaptive audio delay control for multimedia conferencing |
US5737434A (en) * | 1996-08-26 | 1998-04-07 | Orban, Inc. | Multi-band audio compressor with look-ahead clipper |
US6449368B1 (en) * | 1997-03-14 | 2002-09-10 | Dolby Laboratories Licensing Corporation | Multidirectional audio decoding |
KR19980076752A (en) * | 1997-04-14 | 1998-11-16 | 윤종용 | Broadcast signal receiving method and receiving device for automatically switching screen and sound |
US6760448B1 (en) * | 1999-02-05 | 2004-07-06 | Dolby Laboratories Licensing Corporation | Compatible matrix-encoded surround-sound channels in a discrete digital sound format |
US6961432B1 (en) * | 1999-04-29 | 2005-11-01 | Agere Systems Inc. | Multidescriptive coding technique for multistream communication of signals |
US20020009000A1 (en) * | 2000-01-18 | 2002-01-24 | Qdesign Usa, Inc. | Adding imperceptible noise to audio and other types of signals to cause significant degradation when compressed and decompressed |
US6741874B1 (en) * | 2000-04-18 | 2004-05-25 | Motorola, Inc. | Method and apparatus for reducing echo feedback in a communication system |
US7046812B1 (en) * | 2000-05-23 | 2006-05-16 | Lucent Technologies Inc. | Acoustic beam forming with robust signal estimation |
DE60114638T2 (en) * | 2000-08-16 | 2006-07-20 | Dolby Laboratories Licensing Corp., San Francisco | MODULATION OF ONE OR MORE PARAMETERS IN A PERCEPTIONAL AUDIO OR VIDEO CODING SYSTEM IN RESPONSE TO ADDITIONAL INFORMATION |
US20020122468A1 (en) * | 2001-01-12 | 2002-09-05 | Terion, Inc. | Quasi orthogonal hybrid walsh-PN codes for CDMA application in HF modems |
US6915264B2 (en) * | 2001-02-22 | 2005-07-05 | Lucent Technologies Inc. | Cochlear filter bank structure for determining masked thresholds for use in perceptual audio coding |
US7447631B2 (en) | 2002-06-17 | 2008-11-04 | Dolby Laboratories Licensing Corporation | Audio coding system using spectral hole filling |
US20040098149A1 (en) * | 2002-11-15 | 2004-05-20 | Chuan Liu | Digital audio sampling scheme |
US7395210B2 (en) * | 2002-11-21 | 2008-07-01 | Microsoft Corporation | Progressive to lossless embedded audio coder (PLEAC) with multiple factorization reversible transform |
US7191025B2 (en) * | 2002-12-20 | 2007-03-13 | Texas Instruments Incorporated | Variable digital high and low pass filters |
ATE359687T1 (en) * | 2003-04-17 | 2007-05-15 | Koninkl Philips Electronics Nv | AUDIO SIGNAL GENERATION |
US7158632B2 (en) * | 2003-08-20 | 2007-01-02 | Intel Corporation | Adaptive scaling and echo reduction |
JP2005136647A (en) * | 2003-10-30 | 2005-05-26 | New Japan Radio Co Ltd | Bass booster circuit |
DE10355146A1 (en) * | 2003-11-26 | 2005-07-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for generating a bass channel |
KR100528876B1 (en) * | 2003-12-15 | 2005-11-16 | 삼성전자주식회사 | Apparatus for sharpening image adaptively to multi-video formats and method thereof |
US7929708B2 (en) * | 2004-01-12 | 2011-04-19 | Dts, Inc. | Audio spatial environment engine |
US7394903B2 (en) * | 2004-01-20 | 2008-07-01 | Fraunhofer-Gesellschaft Zur Forderung Der Angewandten Forschung E.V. | Apparatus and method for constructing a multi-channel output signal or for generating a downmix signal |
JP2006019770A (en) * | 2004-05-31 | 2006-01-19 | Toshiba Corp | Broadcast receiving apparatus and broadcast receiving method, and sound reproducing apparatus and sound reproducing method |
US9281794B1 (en) * | 2004-08-10 | 2016-03-08 | Bongiovi Acoustics Llc. | System and method for digital signal processing |
KR101177677B1 (en) * | 2004-10-28 | 2012-08-27 | 디티에스 워싱턴, 엘엘씨 | Audio spatial environment engine |
KR100707186B1 (en) | 2005-03-24 | 2007-04-13 | 삼성전자주식회사 | Audio coding and decoding apparatus and method, and recoding medium thereof |
US8964997B2 (en) * | 2005-05-18 | 2015-02-24 | Bose Corporation | Adapted audio masking |
US20060262938A1 (en) * | 2005-05-18 | 2006-11-23 | Gauger Daniel M Jr | Adapted audio response |
US7953605B2 (en) * | 2005-10-07 | 2011-05-31 | Deepen Sinha | Method and apparatus for audio encoding and decoding using wideband psychoacoustic modeling and bandwidth extension |
US20070223740A1 (en) * | 2006-02-14 | 2007-09-27 | Reams Robert W | Audio spatial environment engine using a single fine structure |
ATE448638T1 (en) * | 2006-04-13 | 2009-11-15 | Fraunhofer Ges Forschung | AUDIO SIGNAL DECORRELATOR |
AU2007243586B2 (en) * | 2006-04-27 | 2010-12-23 | Dolby Laboratories Licensing Corporation | Audio gain control using specific-loudness-based auditory event detection |
US8712061B2 (en) * | 2006-05-17 | 2014-04-29 | Creative Technology Ltd | Phase-amplitude 3-D stereo encoder and decoder |
US20080049950A1 (en) * | 2006-08-22 | 2008-02-28 | Poletti Mark A | Nonlinear Processor for Audio Signals |
JP4972742B2 (en) * | 2006-10-17 | 2012-07-11 | 国立大学法人九州工業大学 | High-frequency signal interpolation method and high-frequency signal interpolation device |
KR101329308B1 (en) * | 2006-11-22 | 2013-11-13 | 삼성전자주식회사 | Method for enhancing Bass of Audio signal and apparatus therefore, Method for calculating fundamental frequency of audio signal and apparatus therefor |
US8345887B1 (en) * | 2007-02-23 | 2013-01-01 | Sony Computer Entertainment America Inc. | Computationally efficient synthetic reverberation |
JP5213339B2 (en) * | 2007-03-12 | 2013-06-19 | アルパイン株式会社 | Audio equipment |
WO2008126496A1 (en) * | 2007-03-20 | 2008-10-23 | Nec Corporation | Acoustic processing system and method for electronic device and mobile telephone terminal |
US8611557B2 (en) * | 2007-08-17 | 2013-12-17 | J. Craig Oxford | Method and apparatus for audio processing |
US20090052676A1 (en) * | 2007-08-20 | 2009-02-26 | Reams Robert W | Phase decorrelation for audio processing |
US20090198500A1 (en) * | 2007-08-24 | 2009-08-06 | Qualcomm Incorporated | Temporal masking in audio coding based on spectral dynamics in frequency sub-bands |
US8103005B2 (en) * | 2008-02-04 | 2012-01-24 | Creative Technology Ltd | Primary-ambient decomposition of stereo audio signals using a complex similarity index |
US8645144B2 (en) * | 2008-04-24 | 2014-02-04 | Broadcom Corporation | Audio signal shaping for playback by audio devices |
JP5202090B2 (en) * | 2008-05-07 | 2013-06-05 | アルパイン株式会社 | Surround generator |
US20090319901A1 (en) * | 2008-06-20 | 2009-12-24 | Brian David Johnson | Aggregration of multiple media types of user consumption habits and device preferences |
EP4372744A1 (en) | 2008-07-11 | 2024-05-22 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoder, audio decoder, methods for encoding and decoding an audio signal, audio stream and computer program |
TWI413109B (en) * | 2008-10-01 | 2013-10-21 | Dolby Lab Licensing Corp | Decorrelator for upmixing systems |
US8280068B2 (en) * | 2008-10-03 | 2012-10-02 | Adaptive Sound Technologies, Inc. | Ambient audio transformation using transformation audio |
PL4231291T3 (en) * | 2008-12-15 | 2024-04-15 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio bandwidth extension decoder, corresponding method and computer program |
EP2436176A4 (en) * | 2009-05-27 | 2012-11-28 | Nokia Corp | Spatial audio mixing arrangement |
US9083288B2 (en) * | 2009-06-11 | 2015-07-14 | Invensense, Inc. | High level capable audio amplification circuit |
US9071214B2 (en) * | 2009-06-11 | 2015-06-30 | Invensense, Inc. | Audio signal controller |
US20110300874A1 (en) * | 2010-06-04 | 2011-12-08 | Apple Inc. | System and method for removing tdma audio noise |
US20110317841A1 (en) * | 2010-06-25 | 2011-12-29 | Lloyd Trammell | Method and device for optimizing audio quality |
US20120076324A1 (en) * | 2010-09-23 | 2012-03-29 | AudioControl, Inc. | System and methods for applying bass compensation in an automobile |
EP2521377A1 (en) * | 2011-05-06 | 2012-11-07 | Jacoti BVBA | Personal communication device with hearing support and method for providing the same |
WO2012094335A1 (en) * | 2011-01-04 | 2012-07-12 | Srs Labs, Inc. | Immersive audio rendering system |
US9130683B2 (en) * | 2011-03-07 | 2015-09-08 | Texas Instruments Incorporated | Silence based attenuation for enhanced idle-channel FM or other receiver co-existence with a coexisting radio and circuits, processes, and systems |
US8805312B2 (en) * | 2011-04-06 | 2014-08-12 | Texas Instruments Incorporated | Methods, circuits, systems and apparatus providing audio sensitivity enhancement in a wireless receiver, power management and other performances |
US8972251B2 (en) * | 2011-06-07 | 2015-03-03 | Qualcomm Incorporated | Generating a masking signal on an electronic device |
EP2560161A1 (en) * | 2011-08-17 | 2013-02-20 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Optimal mixing matrices and usage of decorrelators in spatial audio processing |
DE102011082310A1 (en) * | 2011-09-07 | 2013-03-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus, method and electroacoustic system for reverberation time extension |
KR101803293B1 (en) * | 2011-09-09 | 2017-12-01 | 삼성전자주식회사 | Signal processing apparatus and method for providing 3d sound effect |
DE112011105624B4 (en) * | 2011-09-15 | 2023-03-02 | Mitsubishi Electric Corp. | Dynamic range control device |
US20130077795A1 (en) * | 2011-09-28 | 2013-03-28 | Texas Instruments Incorporated | Over-Excursion Protection for Loudspeakers |
CN103918030B (en) * | 2011-09-29 | 2016-08-17 | 杜比国际公司 | High quality detection in the FM stereo radio signal of telecommunication |
JP5944403B2 (en) * | 2011-10-21 | 2016-07-05 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | Acoustic rendering apparatus and acoustic rendering method |
US20130129117A1 (en) * | 2011-11-21 | 2013-05-23 | Henrik Thomsen | Audio amplification circuit |
US8953901B2 (en) * | 2012-01-12 | 2015-02-10 | Sony Corporation | Method and system for applying filter to image |
US9173020B2 (en) * | 2012-03-27 | 2015-10-27 | Htc Corporation | Control method of sound producing, sound producing apparatus, and portable apparatus |
US20130262687A1 (en) * | 2012-03-29 | 2013-10-03 | Ryan L. Avery | Connecting a mobile device as a remote control |
US9264838B2 (en) * | 2012-12-27 | 2016-02-16 | Dts, Inc. | System and method for variable decorrelation of audio signals |
US9571054B2 (en) * | 2013-02-28 | 2017-02-14 | Rovi Guides, Inc. | Systems and methods for dynamically adjusting volume based on media content |
US9060223B2 (en) * | 2013-03-07 | 2015-06-16 | Aphex, Llc | Method and circuitry for processing audio signals |
US9099980B2 (en) * | 2013-05-06 | 2015-08-04 | Pratt & Whitney Canada Corp | Dynamically detecting resonating frequencies of resonating structures |
CN110619882B (en) * | 2013-07-29 | 2023-04-04 | 杜比实验室特许公司 | System and method for reducing temporal artifacts of transient signals in decorrelator circuits |
JP6212645B2 (en) * | 2013-09-12 | 2017-10-11 | ドルビー・インターナショナル・アーベー | Audio decoding system and audio encoding system |
WO2015165076A1 (en) * | 2014-04-30 | 2015-11-05 | Motorola Solutions, Inc. | Method and apparatus for discriminating between voice signals |
US9875756B2 (en) * | 2014-12-16 | 2018-01-23 | Psyx Research, Inc. | System and method for artifact masking |
-
2015
- 2015-09-23 US US14/863,365 patent/US9875756B2/en active Active
- 2015-09-23 US US14/863,350 patent/US9852744B2/en active Active
- 2015-09-23 US US14/863,357 patent/US9691408B2/en active Active
- 2015-09-23 US US14/863,368 patent/US20160171987A1/en not_active Abandoned
- 2015-09-23 US US14/863,374 patent/US20160173808A1/en not_active Abandoned
- 2015-09-23 US US14/863,376 patent/US9830927B2/en active Active
- 2015-12-16 WO PCT/US2015/065936 patent/WO2016100422A1/en active Application Filing
-
2017
- 2017-06-27 US US15/634,586 patent/US20170294890A1/en not_active Abandoned
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4052571A (en) * | 1975-11-07 | 1977-10-04 | National Research Development Corporation | Hearing aid with amplitude compression achieved by clipping a modulated signal |
US20040170283A1 (en) * | 2002-03-12 | 2004-09-02 | Yasuhiro Terada | Howling control device and howling control method |
US20100023321A1 (en) * | 2008-07-25 | 2010-01-28 | Yamaha Corporation | Voice processing apparatus and method |
US20130010972A1 (en) * | 2011-07-04 | 2013-01-10 | Gn Resound A/S | Binaural compressor preserving directional cues |
Also Published As
Publication number | Publication date |
---|---|
US20160171985A1 (en) | 2016-06-16 |
US20160171987A1 (en) | 2016-06-16 |
US20160171968A1 (en) | 2016-06-16 |
US20160173808A1 (en) | 2016-06-16 |
US20160173995A1 (en) | 2016-06-16 |
US9875756B2 (en) | 2018-01-23 |
US9691408B2 (en) | 2017-06-27 |
US9852744B2 (en) | 2017-12-26 |
WO2016100422A1 (en) | 2016-06-23 |
US9830927B2 (en) | 2017-11-28 |
US20160173979A1 (en) | 2016-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9691408B2 (en) | System and method for dynamic equalization of audio data | |
JP6177798B2 (en) | Bus enhancement system | |
US10750278B2 (en) | Adaptive bass processing system | |
US9985597B2 (en) | Digital compressor for compressing an audio signal | |
US20160344356A1 (en) | Audio Compression System for Compressing an Audio Signal | |
US9530429B2 (en) | Reverberation suppression apparatus used for auditory device | |
US20210067122A1 (en) | Systems and methods of volume limiting | |
US10547927B1 (en) | Systems and methods for processing an audio signal for replay on stereo and multi-channel audio devices | |
US20160322949A1 (en) | Frequency-Domain DRC | |
EP3811514B1 (en) | Audio enhancement in response to compression feedback | |
US9967663B2 (en) | Loudspeaker protection against excessive excursion | |
US10152961B2 (en) | Signal processing device and signal processing method | |
EP3599775B1 (en) | Systems and methods for processing an audio signal for replay on stereo and multi-channel audio devices | |
EP2963816B1 (en) | Adaptive detector and auto mode for dynamics processor | |
KR20230147638A (en) | Virtualizer for binaural audio | |
WO2016149085A2 (en) | System and method for dynamic recovery of audio data and compressed audio enhancement | |
US20140119546A1 (en) | Apparatus and method for keeping output loudness and quality of sound among different equalizer modes | |
US11757420B2 (en) | Method for dynamically adjusting adjustable gain value to equalize input signal to generate equalizer output signal and associated leveling equalizer | |
US9473869B2 (en) | Audio signal processing device | |
US20240170001A1 (en) | Improving perceptual quality of dereverberation | |
CN116964665A (en) | Improving perceived quality of dereverberation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |