CN106792253B - sound effect processing method and system - Google Patents
sound effect processing method and system Download PDFInfo
- Publication number
- CN106792253B CN106792253B CN201611092855.7A CN201611092855A CN106792253B CN 106792253 B CN106792253 B CN 106792253B CN 201611092855 A CN201611092855 A CN 201611092855A CN 106792253 B CN106792253 B CN 106792253B
- Authority
- CN
- China
- Prior art keywords
- sound
- signal
- feature
- reference voice
- preset
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000000694 effects Effects 0.000 title claims abstract description 62
- 238000003672 processing method Methods 0.000 title abstract 2
- 230000005236 sound signal Effects 0.000 claims abstract description 50
- 230000007613 environmental effect Effects 0.000 claims abstract description 45
- 238000000034 method Methods 0.000 claims abstract description 21
- 238000000605 extraction Methods 0.000 claims abstract description 18
- 238000010183 spectrum analysis Methods 0.000 claims description 12
- 230000011218 segmentation Effects 0.000 claims description 7
- 238000006243 chemical reaction Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 4
- 241000208340 Araliaceae Species 0.000 claims 2
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 claims 2
- 235000003140 Panax quinquefolius Nutrition 0.000 claims 2
- 235000008434 ginseng Nutrition 0.000 claims 2
- 238000004458 analytical method Methods 0.000 description 6
- 238000010561 standard procedure Methods 0.000 description 5
- 238000007405 data analysis Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000013139 quantization Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/485—End-user interface for client configuration
- H04N21/4852—End-user interface for client configuration for modifying audio parameters, e.g. switching between mono and stereo
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/439—Processing of audio elementary streams
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
The invention relates to a sound effect processing method and a system, wherein the method comprises the following steps: collecting an environmental sound signal within a preset time range; carrying out feature extraction on the environmental sound signal to obtain environmental sound features; selecting a reference sound feature with the maximum similarity to the environmental sound feature from a plurality of preset reference sound features; searching a preset sound effect mode corresponding to the selected reference sound characteristic to obtain a matched sound effect mode; and performing sound effect processing on the sound signal to be played according to the matched sound effect mode. Therefore, the most suitable matching sound effect mode can be automatically selected according to the environmental sound characteristics of the environmental sound signal, and the sound effect processing effect is good; meanwhile, user operation is not needed, and the use convenience of the user is improved.
Description
Technical field
The present invention relates to signal processing technology fields, more particularly to a kind of sound effect treatment method and system.
Background technique
When audio and video display device carries out audio-video or program broadcasting, it can according to need and different audio modes is set.
By taking television set as an example, the method for traditional replacement audio mode is that menu interface of the user using remote controler in display selects
It selects, the audio mode that the processing system of television set is selected according to user exports after handling voice signal.
However, usually there is noise in external environment locating for television set, user is not aware that corresponding to current environment,
Which kind of most suitable audio mode is, is easy to appear the audio effect processing result and the incongruent situation of current environment of television set, sound
Imitate poor processing effect.
Summary of the invention
Based on this, it is necessary in view of the above-mentioned problems, providing the sound effect treatment method and system of a kind of high treating effect.
A kind of sound effect treatment method, comprising:
Acquire the environmental sound signal in preset time range;
Feature extraction is carried out to the environmental sound signal, obtains ambient sound feature;
It is chosen and the maximum reference voice of ambient sound characteristic similarity from preset multiple reference voice features
Feature;
Default audio mode corresponding with the reference voice feature chosen is searched, matching audio mode is obtained;
Audio effect processing is carried out to voice signal to be played according to the matching audio mode.
A kind of sound effect processing system, comprising:
Environmental sound signal acquisition module, for acquiring the environmental sound signal in preset time range;
Ambient sound feature obtains module, for carrying out feature extraction to the environmental sound signal, obtains ambient sound
Feature;
Reference voice characteristic selecting module, for being chosen and the ambient sound from preset multiple reference voice features
The maximum reference voice feature of characteristic similarity;
Match audio mode searching module, the corresponding default audio mode of reference voice feature for searching with choosing,
Obtain matching audio mode;
Audio effect processing module, for carrying out audio effect processing to voice signal to be played according to the matching audio mode.
Above-mentioned sound effect treatment method and system, by the environmental sound signal in acquisition preset time range, to ambient sound
Sound signal carries out feature extraction and obtains ambient sound feature;Then selection and ambient sound from preset multiple reference voice features
The maximum reference voice feature of sound characteristic similarity is searched default audio mode corresponding with the reference voice feature chosen and is obtained
Audio mode is matched, and audio effect processing is carried out to voice signal to be played according to matching audio mode.In this way, can be according to ambient sound
The ambient sound feature of sound signal, chooses most suitable matching audio mode automatically, and audio effect processing effect is good;Meanwhile without using
Family operation improves the convenience that user uses.
Detailed description of the invention
Fig. 1 is the flow chart of sound effect treatment method in an embodiment;
Fig. 2 is to carry out feature extraction to environmental sound signal in an embodiment, obtains the detailed process of ambient sound feature
Figure;
Fig. 3 is to choose from preset multiple reference voice features in an embodiment and ambient sound characteristic similarity maximum
Reference voice feature specific flow chart;
Fig. 4 is the module map of sound effect processing system in an embodiment.
Specific embodiment
With reference to Fig. 1, sound effect treatment method in an embodiment includes the following steps.
S110: the environmental sound signal in acquisition preset time range.
Preset time range refers to period default setting or the pre-set time range of duration.Voice signal specifically can be with
It is obtained by the sound that microphone acquires ambient enviroment.
In one embodiment, it is as initial time, with preset value at the time of preset time range is to receive play instruction
The time range of duration.Wherein, play instruction refers to the instruction for being used to indicate and opening that audio/video plays or TV programme play, example
The instruction of processing system is waken up when such as television boot-strap.Preset value can be specifically arranged according to actual needs.In the present embodiment, in advance
If value is 5 seconds;When corresponding to the television boot-strap moment at the time of receiving play instruction, preset time range is after television boot-strap
First 5 seconds.
Usually after receiving play instruction, processing system just plays audio/video after needing the shorter response time.Pass through choosing
The time range that taking is initial time at the time of receiving play instruction, preset value is duration carries out the acquisition of voice signal, obtains
Environmental sound signal be starting play audio/video before voice signal, avoid the sound of actual play to ambient sound
The influence of sound signal acquisition, can be improved the accuracy of sound signal collecting.It is appreciated that in other embodiments, preset time
Range can also with other times range, such as using current time as initial time, using preset value as the time range of duration, currently
Moment can set in real time, realize and acquire environmental sound signal simultaneously during playing audio/video.
S130: feature extraction is carried out to environmental sound signal, obtains ambient sound feature.
Carrying out the ambient sound feature that feature extraction obtains to environmental sound signal can be numerical value or image.
In one embodiment, ambient sound feature is numerical value.With reference to Fig. 2, step S130 includes step S131 to step
S134。
S131: environmental sound signal is converted into digital signal.
The ambient sound of acquisition is analog signal, can convert analog signals into digital signal by analog-to-digital conversion.
S132: to digital signal carry out spectrum analysis obtain include multiple Frequency points frequency information.
Spectrum analysis is carried out to digital signal, specifically can be and analyzed using Fourier transformation, obtain digital signal
The Frequency point for inside including.
S133: the average value of the Frequency point in each predeterminated frequency section is calculated separately according to frequency information, as each pre-
If the characteristic value of frequency band.
Predeterminated frequency section has multiple, can be arranged according to the audible audio frequency range of practical human ear.By frequency information
In be not belonging to the Frequency point of any one predeterminated frequency section and give up, classify according to the size of Frequency point to each Frequency point, it is right
The Frequency point belonged in a predeterminated frequency section calculates average value, the characteristic value of available corresponding predeterminated frequency section.Specifically
Ground, if Frequency point is successive value in the same predeterminated frequency section, the calculating of average value be can be by predeterminated frequency section
Frequency point after divided by frequency spectrum length;If Frequency point in the same predeterminated frequency section is discrete value, average value can be with
Be obtained divided by Frequency point number after the sum of each Frequency point by directly calculating, for example, 30hz (hertz) in frequency information,
50hz, 100hz, 110hz belong to the same predeterminated frequency section, then the average value for calculating 30,50,100 and 110 presets frequency as this
The characteristic value of rate section.
In one embodiment, predeterminated frequency section include 20hz-200hz, 200hz-700hz, 700hz-2000hz,
2000hz-7000hz and 7000hz-15000hz.In this way, the audio frequency range that human ear under normal conditions is heard is drawn
Point, feature extraction is targetedly carried out, data-handling efficiency is improved.It is appreciated that in other embodiments, predeterminated frequency section
It may be set to be other numerical value.
S134: calculating separately the characteristic value of each predeterminated frequency section and the product of corresponding predetermined coefficient, and calculate each product it
With obtain ambient sound feature.
Each predeterminated frequency section is corresponding with a predetermined coefficient in advance, and passes through step S133, each predeterminated frequency
Section is corresponding with a characteristic value.By by the characteristic value of each predeterminated frequency section predetermined coefficient phase corresponding with the predeterminated frequency section
Multiply, each predeterminated frequency section is corresponding to obtain a product, and calculating the sum of products then can be obtained ambient sound feature.In the present embodiment,
Each predeterminated frequency section 20hz-200hz, 200hz-700hz, 700hz-2000hz, 2000hz-7000hz and 7000hz-
The corresponding predetermined coefficient of 15000hz are as follows: -100, -10,0,10 and 100.
By carrying out frequency analysis to environmental sound signal, the frequency information obtained according to frequency analysis calculate
The numerical value arrived is indicated in the form of quantization as ambient sound feature, is convenient for Data Analysis Services.It is appreciated that
In other embodiments, sound characteristic can also be extracted using other methods, for example, to the number after environmental sound signal analog-to-digital conversion
Word signal carries out spectrum analysis, and the spectrogram that spectrum analysis is obtained is directly as environmental sound signal.
S150: it is chosen and the maximum reference voice of ambient sound characteristic similarity from preset multiple reference voice features
Feature.
Reference voice feature, which refers to, to be preset for referring to the sound characteristic compared.If ambient sound feature is numerical value,
Reference voice feature is similarly numerical value, and reference voice feature and the similarity of ambient sound feature are obtained by calculating the difference of the two
It arrives, difference is smaller, then similarity is more maximum.If ambient sound feature is image, reference voice feature is similarly image, refers to
Sound characteristic obtains compared with the similarity of ambient sound feature is by image, and image comparing difference is smaller, then similarity is bigger.
Wherein, reference voice feature can be obtained by preparatory collection analysis.In one embodiment, with reference to Fig. 3, step
It further include step S101 to step S103 before S110.
S101: respectively to the voice signal in multiple model scenes acquisition preset duration, multiple reference voice signals are obtained.
Model scene refers to real life scene.The quantity of model scene is equal with the quantity of reference voice feature, mould
The voice signal of type scene can specifically be acquired by microphone.In the present embodiment, model scene includes parlor in the daytime, night visitor
Five kinds of the Room, hotel, sales field and dining room, respectively correspond The clamors of the people bubble up, it is quiet, quiet, mix and spacious five kinds of situations.It can manage
Solution, in other embodiments, model scene can also be other scenes.
Preset duration is greater than the corresponding duration of preset time range.Model scene voice signal is carried out for a long time by setting
Acquisition, obtained voice signal can accurately indicate the corresponding ambient acoustical circumstances of model scene, so that obtained reference
Voice signal is more acurrate.In the present embodiment, preset duration is 10000 seconds.
S102: according to the ratio between preset duration duration corresponding with preset time range, each reference voice signal is divided into more
A signal segment.
The ratio between preset duration duration corresponding with preset time range may be the integer value greater than 1, it is also possible to be greater than 1
Non integer value.Reference voice signal is the voice signal continuous collection at each moment in preset duration, and each reference voice is believed
Number it is divided into multiple signal segments, it is a length of time interval when specifically can be corresponding with preset time range, suitable according to time order and function
Ordered pair reference voice signal carries out interception segmentation, and the last voice signal less than in a time interval is as a signal segment.
In this way, the corresponding reference voice signal of each model scene can be divided into multiple signal segments.
S103: extracting the sound characteristic of each signal segment, obtains corresponding reference voice according to the sound characteristic of each signal segment
The sound characteristic of signal obtains reference voice feature.
The specific method for extracting the sound characteristic of each signal segment, the specific side with the sound characteristic of extraction environment voice signal
Method is identical, and this will not be repeated here.If extracting obtained sound characteristic is numerical value, according to the acquisition pair of the sound characteristic of each signal segment
The sound characteristic for the reference voice signal answered specifically can be using the average value of the numerical value of each signal segment as reference voice spy
Sign;If extracting obtained sound characteristic is image, corresponding reference voice signal is obtained according to the sound characteristic of each signal segment
Sound characteristic, specifically can be and image processing and analyzing carried out to the image of each signal segment, obtain representing entire reference voice letter
Number image as reference voice feature.
Acquire the reference voice signal of multiple model scenes in advance to the mode of step S103 by using step S101, and
It carries out feature extraction and obtains the reference voice feature of each model scene, the representative strong and accuracy of obtained reference voice feature
It is high.
S170: searching default audio mode corresponding with the reference voice feature chosen, and obtains matching audio mode.
Each reference voice feature is corresponding with a kind of default audio mode in advance, specifically can be by that will refer to sound in advance
Sound feature storage corresponding with default audio mode, so that corresponding default audio mould can be found according to reference voice feature
Formula.
In the present embodiment, presetting audio mode includes five kinds of news, night, movie theatre, standard and music modes, is respectively corresponded
In the daytime the reference voice feature in parlor, night parlor, hotel, sales field and five kinds of dining room model scene.In the setting of audio mode,
Be typically used for three kinds of standard techniques: total-sonic, total volume and total surround, total-sonic have
On/off two states, total volume have tri- kinds of states of normal/night/off, and total surround has on/off
Two states.The state that each default audio mode corresponds to accepted standard technology is as shown in table 1 below.
Table 1
total-sonic | total volume | total surround | |
News | on | normal | off |
Night | on | night | off |
Movie theatre | on | off | on |
Standard | off | off | off |
Music | on | off | off |
S190: audio effect processing is carried out to voice signal to be played according to matching audio mode.
After obtaining matching audio mode, audio effect processing is carried out according to matching audio mode, specifically using matching audio mould
The corresponding standard technique of formula carries out audio effect processing to voice signal to be played automatically, so that the voice signal to be played of output adapts to
In the environment being presently in.For example, if matching audio mode be music, be arranged total-sonic, total volume and
The state of tri- kinds of standard techniques of total surround is respectively on, off, off.Voice signal to be played can be television set
Sound of television signal.
Above-mentioned sound effect treatment method, by the environmental sound signal in acquisition preset time range, to environmental sound signal
It carries out feature extraction and obtains ambient sound feature;Then it is chosen and ambient sound feature from preset multiple reference voice features
The maximum reference voice feature of similarity searches default audio mode corresponding with the reference voice feature chosen and obtains matching sound
Effect mode, and audio effect processing is carried out to voice signal to be played according to matching audio mode.In this way, can be according to environmental sound signal
Ambient sound feature, choose most suitable matching audio mode automatically, audio effect processing effect is good;Meanwhile it being not necessarily to user's operation,
Improve the convenience that user uses.
Above-mentioned sound effect treatment method can be applied to the processing system of television set, allow television set automatic according to environment
Selection matching audio mode carries out audio effect processing.Above-mentioned sound effect treatment method also can be applied to other audio and video display devices,
Such as mobile phone, plate etc., so that audio and video display device when opening player, can automatically select matching audio according to environment
Mode carries out audio effect processing.
With reference to Fig. 4, sound effect processing system in an embodiment, including environmental sound signal acquisition module 110, ambient sound
Feature obtains module 130, reference voice characteristic selecting module 150, matching audio mode searching module 170 and audio effect processing module
190。
Environmental sound signal acquisition module 110 is used to acquire the environmental sound signal in preset time range.
Preset time range refers to period default setting or the pre-set time range of duration.Voice signal specifically can be with
It is obtained by the sound that microphone acquires ambient enviroment.
In one embodiment, it is as initial time, with preset value at the time of preset time range is to receive play instruction
The time range of duration.Wherein, play instruction refers to the instruction for being used to indicate and opening that audio/video plays or TV programme play, example
The instruction of processing system is waken up when such as television boot-strap.Preset value can be specifically arranged according to actual needs.In the present embodiment, in advance
If value is 5 seconds;When corresponding to the television boot-strap moment at the time of receiving play instruction, preset time range is after television boot-strap
First 5 seconds.
Usually after receiving play instruction, processing system just plays audio/video after needing the shorter response time.Pass through choosing
The time range that taking is initial time at the time of receiving play instruction, preset value is duration carries out the acquisition of voice signal, obtains
Environmental sound signal be starting play audio/video before voice signal, avoid the sound of actual play to ambient sound
The influence of sound signal acquisition, can be improved the accuracy of sound signal collecting.
Ambient sound feature obtains module 130 and is used to carry out feature extraction to environmental sound signal, obtains ambient sound spy
Sign.
Carrying out the ambient sound feature that feature extraction obtains to environmental sound signal can be numerical value or image.
In one embodiment, ambient sound feature is numerical value.It includes analog-to-digital conversion list that ambient sound feature, which obtains module 130,
First (not shown), spectral analysis unit (not shown), characteristic value computing unit (not shown) and ambient sound feature calculation unit
(not shown).
AD conversion unit is used to environmental sound signal being converted to digital signal.
Spectral analysis unit be used for digital signal carry out spectrum analysis obtain include multiple Frequency points frequency information.It is right
Digital signal carries out spectrum analysis, specifically can be and is analyzed using Fourier transformation, obtains the frequency for including in digital signal
Rate point.
Characteristic value computing unit is used to calculate separately the flat of the Frequency point in each predeterminated frequency section according to frequency information
Mean value, the characteristic value as each predeterminated frequency section.
If Frequency point is successive value in the same predeterminated frequency section, the calculating of average value be can be by predeterminated frequency
Divided by frequency spectrum length after frequency point in section;If the Frequency point in the same predeterminated frequency section is discrete value, average value
It can be and obtained divided by Frequency point number after the sum of each Frequency point by directly calculating.
In one embodiment, predeterminated frequency section include 20hz-200hz, 200hz-700hz, 700hz-2000hz,
2000hz-7000hz and 7000hz-15000hz.In this way, the audio frequency range that human ear under normal conditions is heard is drawn
Point, feature extraction is targetedly carried out, data-handling efficiency is improved.
Ambient sound feature calculation unit is used to calculate separately the characteristic value of each predeterminated frequency section and corresponding predetermined coefficient
Product, and calculate each sum of products and obtain ambient sound feature.In the present embodiment, each predeterminated frequency section 20hz-200hz,
The corresponding predetermined coefficient of 200hz-700hz, 700hz-2000hz, 2000hz-7000hz and 7000hz-15000hz are as follows:-
100, -10,0,10 and 100.
By using AD conversion unit, spectral analysis unit, characteristic value computing unit and ambient sound feature calculation list
Member carries out frequency analysis, the numerical value that the frequency information obtained according to frequency analysis is calculated to environmental sound signal
It as ambient sound feature, is indicated in the form of quantization, is convenient for Data Analysis Services.
Reference voice characteristic selecting module 150 is used to choose from preset multiple reference voice features special with ambient sound
Levy the maximum reference voice feature of similarity.
If ambient sound feature is numerical value, reference voice feature is similarly numerical value, reference voice feature and ambient sound
The similarity of feature is obtained by calculating the difference of the two, and difference is smaller, then similarity is more maximum.If ambient sound feature is figure
Picture, then reference voice feature is similarly image, and reference voice feature is compared with the similarity of ambient sound feature is by image
It arrives, image comparing difference is smaller, then similarity is bigger.
Reference voice feature can be obtained by preparatory collection analysis.In one embodiment, above-mentioned sound effect processing system is also
It is obtained including reference voice signal acquisition module (not shown), reference voice signal segmentation module (not shown) and reference voice feature
Modulus block (not shown).
Reference voice signal acquisition module for the voice signal in multiple model scenes acquisition preset duration, obtaining respectively
To multiple reference voice signals.Wherein, the quantity of model scene is equal with the quantity of reference voice feature, and preset duration is greater than pre-
If the corresponding duration of time range.In the present embodiment, model scene includes parlor, night parlor, hotel, sales field and dining room in the daytime
Five kinds, respectively correspond The clamors of the people bubble up, it is quiet, quiet, mix and spacious five kinds of situations.It is appreciated that in other embodiments, mould
Type scene can also be other scenes.In the present embodiment, preset duration is 10000 seconds.
Reference voice signal segmentation module is used for according to the ratio between preset duration duration corresponding with preset time range, will be each
Reference voice signal is divided into multiple signal segments.
Reference voice signal is the voice signal continuous collection at each moment in preset duration, by each reference voice signal point
For multiple signal segments, a length of time interval when specifically can be corresponding with preset time range, according to chronological order pair
Reference voice signal carries out interception segmentation, and the last voice signal less than in a time interval is as a signal segment.In this way,
The corresponding reference voice signal of each model scene can be divided into multiple signal segments.
Reference voice feature obtains the sound characteristic that module is used to extract each signal segment, according to the sound characteristic of each signal segment
The sound characteristic for obtaining corresponding reference voice signal obtains reference voice feature.
Mould is obtained by using reference voice signal acquisition module, reference voice signal segmentation module and reference voice feature
Block acquires the reference voice signal of multiple model scenes in advance, and carries out feature extraction and obtain the reference voice of each model scene
Feature, the representative strong and accuracy of obtained reference voice feature are high.
Matching audio mode searching module 170 is for searching default audio mould corresponding with the reference voice feature chosen
Formula obtains matching audio mode.
In the present embodiment, presetting audio mode includes five kinds of news, night, movie theatre, standard and music modes, is respectively corresponded
In the daytime the reference voice feature in parlor, night parlor, hotel, sales field and five kinds of dining room model scene.In the setting of audio mode,
Be typically used for three kinds of standard techniques: total-sonic, total volume and total surround, total-sonic have
On/off two states, total volume have tri- kinds of states of normal/night/off, and total surround has on/off
Two states.
Audio effect processing module 190 is used to carry out audio effect processing to voice signal to be played according to matching audio mode.
After obtaining matching audio mode, audio effect processing is carried out according to matching audio mode, specifically using matching audio mould
The corresponding standard technique of formula carries out audio effect processing to voice signal to be played automatically, so that the voice signal to be played of output adapts to
In the environment being presently in.Wherein, voice signal to be played can be the sound of television signal of television set.
Above-mentioned sound effect processing system acquires the environment in preset time range by environmental sound signal acquisition module 110
Voice signal, ambient sound feature obtain module 130 and obtain ambient sound feature to environmental sound signal progress feature extraction;So
Reference voice characteristic selecting module 150 is chosen from preset multiple reference voice features with ambient sound characteristic similarity most afterwards
Big reference voice feature, matching audio mode searching module 170 search default sound corresponding with the reference voice feature chosen
Effect mode obtains matching audio mode, and audio effect processing module 190 carries out sound to voice signal to be played according to matching audio mode
Effect processing.In this way, most suitable matching audio mode, sound can be chosen automatically according to the ambient sound feature of environmental sound signal
Imitate high treating effect;Meanwhile it being not necessarily to user's operation, improve the convenience that user uses.
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality
It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not deposited
In contradiction, all should be considered as described in this specification.
The embodiments described above only express several embodiments of the present invention, and the description thereof is more specific and detailed, but simultaneously
It cannot therefore be construed as limiting the scope of the patent.It should be pointed out that coming for those of ordinary skill in the art
It says, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to protection of the invention
Range.Therefore, the scope of protection of the patent of the invention shall be subject to the appended claims.
Claims (8)
1. a kind of sound effect treatment method characterized by comprising
Acquire the environmental sound signal in preset time range;
Feature extraction is carried out to the environmental sound signal, obtains ambient sound feature;
It is chosen and the maximum reference voice feature of the ambient sound characteristic similarity from preset multiple reference voice features;
Default audio mode corresponding with the reference voice feature chosen is searched, matching audio mode is obtained;
Audio effect processing is carried out to voice signal to be played according to the matching audio mode;
It before environmental sound signal in the acquisition preset time range, further include multiple model scenes being acquired respectively pre-
If the voice signal in duration, multiple reference voice signals, the quantity of the model scene and the reference voice feature are obtained
Quantity it is equal, the preset duration be greater than the corresponding duration of the preset time range;According to the preset duration with it is described
Each reference voice signal is divided into multiple signal segments by the ratio between corresponding duration of preset time range;Extract the sound of each signal segment
It is special to obtain the reference voice according to the sound characteristic that the sound characteristic of each signal segment obtains corresponding reference voice signal for feature
Sign.
2. sound effect treatment method according to claim 1, which is characterized in that the preset time range is to receive and play
It is initial time at the time of instruction, using preset value as the time range of duration.
3. sound effect treatment method according to claim 1, which is characterized in that described to carry out spy to the environmental sound signal
Sign is extracted, and ambient sound feature is obtained, comprising:
The environmental sound signal is converted into digital signal;
To the digital signal carry out spectrum analysis obtain include multiple Frequency points frequency information;
The average value that the Frequency point in each predeterminated frequency section is calculated separately according to the frequency information, as each predeterminated frequency
The characteristic value of section;
Calculate separately the characteristic value of each predeterminated frequency section and the product of corresponding predetermined coefficient, and calculate each sum of products obtain it is described
Ambient sound feature.
4. sound effect treatment method according to claim 3, which is characterized in that the predeterminated frequency section includes 20hz-
200hz, 200hz-700hz, 700hz-2000hz, 2000hz-7000hz and 7000hz-15000hz.
5. a kind of sound effect processing system characterized by comprising
Environmental sound signal acquisition module, for acquiring the environmental sound signal in preset time range;
Ambient sound feature obtains module, for carrying out feature extraction to the environmental sound signal, obtains ambient sound feature;
Reference voice characteristic selecting module, for being chosen and the ambient sound feature from preset multiple reference voice features
The maximum reference voice feature of similarity;
Match audio mode searching module, the corresponding default audio mode of reference voice feature for searching with choosing obtains
Match audio mode;
Audio effect processing module, for carrying out audio effect processing to voice signal to be played according to the matching audio mode;
It further include that reference voice signal acquisition module, reference voice signal segmentation module and reference voice feature obtain module;Ginseng
Sound signal collecting module is examined, for the voice signal in multiple model scenes acquisition preset duration, obtaining multiple ginsengs respectively
Voice signal is examined, the quantity of the model scene is equal with the quantity of the reference voice feature, and the preset duration is greater than institute
State the corresponding duration of preset time range;Reference voice signal segmentation module, for being preset according to the preset duration with described
Each reference voice signal is divided into multiple signal segments by the ratio between corresponding duration of time range;Reference voice feature obtains module, uses
In the sound characteristic for extracting each signal segment, the sound for obtaining corresponding reference voice signal according to the sound characteristic of each signal segment is special
Obtain the reference voice feature.
6. sound effect processing system according to claim 5, which is characterized in that the preset time range is to receive and play
It is initial time at the time of instruction, using preset value as the time range of duration.
7. sound effect processing system according to claim 5, which is characterized in that the ambient sound feature obtains module packet
It includes:
AD conversion unit, for the environmental sound signal to be converted to digital signal;
Spectral analysis unit, for the digital signal carry out spectrum analysis obtain include multiple Frequency points frequency information;
Characteristic value computing unit, for calculating separately the flat of the Frequency point in each predeterminated frequency section according to the frequency information
Mean value, the characteristic value as each predeterminated frequency section;
Ambient sound feature calculation unit, for calculating separately the characteristic value of each predeterminated frequency section and multiplying for corresponding predetermined coefficient
Product, and calculate each sum of products and obtain the ambient sound feature.
8. sound effect processing system according to claim 7, which is characterized in that the predeterminated frequency section includes 20hz-
200hz, 200hz-700hz, 700hz-2000hz, 2000hz-7000hz and 7000hz-15000hz.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611092855.7A CN106792253B (en) | 2016-11-30 | 2016-11-30 | sound effect processing method and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611092855.7A CN106792253B (en) | 2016-11-30 | 2016-11-30 | sound effect processing method and system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106792253A CN106792253A (en) | 2017-05-31 |
CN106792253B true CN106792253B (en) | 2019-07-09 |
Family
ID=58915750
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201611092855.7A Active CN106792253B (en) | 2016-11-30 | 2016-11-30 | sound effect processing method and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106792253B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107910018A (en) * | 2017-10-30 | 2018-04-13 | 广州视源电子科技股份有限公司 | Sound effect processing method and system, computer storage medium and equipment |
CN109002275B (en) * | 2018-07-03 | 2021-12-07 | 百度在线网络技术(北京)有限公司 | AR background audio processing method and device, AR equipment and readable storage medium |
CN109121068A (en) * | 2018-07-04 | 2019-01-01 | 广州视源电子科技股份有限公司 | Sound effect control method and device and electronic equipment |
WO2021237650A1 (en) * | 2020-05-29 | 2021-12-02 | Nokia Technologies Oy | Noise control |
CN116453492A (en) * | 2023-06-16 | 2023-07-18 | 成都小唱科技有限公司 | Method and device for switching jukebox airport scenes, computer equipment and storage medium |
CN117593949B (en) * | 2024-01-19 | 2024-03-29 | 成都金都超星天文设备有限公司 | Control method, equipment and medium for astronomical phenomena demonstration of astronomical phenomena operation |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103051794A (en) * | 2012-12-18 | 2013-04-17 | 广东欧珀移动通信有限公司 | Method and device for dynamically setting sound effect of mobile terminal |
CN103219011A (en) * | 2012-01-18 | 2013-07-24 | 联想移动通信科技有限公司 | Noise reduction method, noise reduction device and communication terminal |
CN104811864A (en) * | 2015-04-20 | 2015-07-29 | 深圳市冠旭电子有限公司 | Method and system for self-adaptive adjustment of audio effect |
WO2015142659A1 (en) * | 2014-03-17 | 2015-09-24 | Adaptive Sound Technologies, Inc. | Systems and methods for automatic signal attenuation |
CN104966522A (en) * | 2015-06-30 | 2015-10-07 | 广州酷狗计算机科技有限公司 | Sound effect regulation method, cloud server, stereo device and system |
CN106060643A (en) * | 2016-06-28 | 2016-10-26 | 乐视控股(北京)有限公司 | Method and device for playing multimedia file and earphones |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW201414274A (en) * | 2012-09-21 | 2014-04-01 | Hon Hai Prec Ind Co Ltd | Sound effect managing system and sound effect managing method |
-
2016
- 2016-11-30 CN CN201611092855.7A patent/CN106792253B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103219011A (en) * | 2012-01-18 | 2013-07-24 | 联想移动通信科技有限公司 | Noise reduction method, noise reduction device and communication terminal |
CN103051794A (en) * | 2012-12-18 | 2013-04-17 | 广东欧珀移动通信有限公司 | Method and device for dynamically setting sound effect of mobile terminal |
WO2015142659A1 (en) * | 2014-03-17 | 2015-09-24 | Adaptive Sound Technologies, Inc. | Systems and methods for automatic signal attenuation |
CN104811864A (en) * | 2015-04-20 | 2015-07-29 | 深圳市冠旭电子有限公司 | Method and system for self-adaptive adjustment of audio effect |
CN104966522A (en) * | 2015-06-30 | 2015-10-07 | 广州酷狗计算机科技有限公司 | Sound effect regulation method, cloud server, stereo device and system |
CN106060643A (en) * | 2016-06-28 | 2016-10-26 | 乐视控股(北京)有限公司 | Method and device for playing multimedia file and earphones |
Also Published As
Publication number | Publication date |
---|---|
CN106792253A (en) | 2017-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106792253B (en) | sound effect processing method and system | |
JP6855527B2 (en) | Methods and devices for outputting information | |
US7593618B2 (en) | Image processing for analyzing video content | |
CN107708100B (en) | Advertisement playing method based on user position information | |
CN105074697B (en) | For inferring the accumulation of the real-time crowdsourcing data of the metadata about entity | |
US11445242B2 (en) | Media content identification on mobile devices | |
US8447761B2 (en) | Lifestyle collecting apparatus, user interface device, and lifestyle collecting method | |
US10166472B2 (en) | Methods and systems for determining a reaction time for a response and synchronizing user interface(s) with content being rendered | |
CN103688531A (en) | Control device, control method and program | |
CN1282470A (en) | Voice recognition unit for audience measurement system | |
CN110519539A (en) | For refunding the method, system and medium of media content based on audio event detected | |
CN110047497B (en) | Background audio signal filtering method and device and storage medium | |
CN103873919B (en) | A kind of information processing method and electronic equipment | |
CN103945140B (en) | The generation method and system of video caption | |
CN108429955A (en) | Release ambient sound enters the intelligent apparatus and method of earphone | |
CN103347070A (en) | Method, terminal, server and system for voice data pushing | |
CN104900237B (en) | A kind of methods, devices and systems for audio-frequency information progress noise reduction process | |
CN103945074B (en) | A kind of CRBT method for customizing and system | |
CN112738705A (en) | Sound effect adjusting method and device, electronic equipment and storage medium | |
CN109040778B (en) | Video cover determining method, user equipment, storage medium and device | |
CN111552836A (en) | Lyric display method, device and storage medium | |
CN111477244A (en) | User-defined sports event description enhancing method | |
CN110708600A (en) | Method and apparatus for identifying valid viewers of a television | |
CN105551504A (en) | Method and device for triggering function application of intelligent mobile terminal based on crying sound | |
CN110460800B (en) | Remote rehearsal system and rehearsal method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |