CN105321526A - Audio processing method and electronic device - Google Patents

Audio processing method and electronic device Download PDF

Info

Publication number
CN105321526A
CN105321526A CN201510612358.4A CN201510612358A CN105321526A CN 105321526 A CN105321526 A CN 105321526A CN 201510612358 A CN201510612358 A CN 201510612358A CN 105321526 A CN105321526 A CN 105321526A
Authority
CN
China
Prior art keywords
voice data
audio
frequency
data
processing method
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510612358.4A
Other languages
Chinese (zh)
Other versions
CN105321526B (en
Inventor
王少敏
陈文辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Lenovo Beijing Ltd
Original Assignee
Lenovo Beijing Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lenovo Beijing Ltd filed Critical Lenovo Beijing Ltd
Priority to CN201510612358.4A priority Critical patent/CN105321526B/en
Publication of CN105321526A publication Critical patent/CN105321526A/en
Application granted granted Critical
Publication of CN105321526B publication Critical patent/CN105321526B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/18Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/003Changing voice quality, e.g. pitch or formants
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Telephone Function (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

The invention discloses an audio processing method and an electronic device using the same. The audio processing method comprises: executing first processing on audio data to be processed to obtain first audio data in a first band and second audio data out of the first band; executing second processing on the first audio data to obtain third audio data; and synthesizing the third audio data and the second audio data to generate post-processing audio data.

Description

Audio-frequency processing method and electronic equipment
Technical field
The present invention relates to field of audio processing, more specifically, the present invention relates to the electronic equipment of a kind of audio-frequency processing method and this audio-frequency processing method of use.
Background technology
In the audio frequency process of the such as change of voice process of current use, normally carry out pitch conversion for whole voice data.Such audio frequency process lacks the passage change of voice accurately, does not add differentiation process for the human voice signal comprised in voice data and background sound signal.
Therefore, desirable to provide a kind of audio-frequency processing method and the electronic equipment using this audio-frequency processing method, it can for the audio frequency process needed for the target audio signal comprised in voice data (such as human voice signal) execution, and keep other non-targeted sound signals constant, thus achieve the accurate audio frequency process for target audio signal.
Summary of the invention
Because above-mentioned situation, the invention provides the electronic equipment of a kind of audio-frequency processing method and this audio-frequency processing method of use.
According to an embodiment of the present disclosure, provide a kind of audio-frequency processing method, comprising: treat processing audio data and perform the first process, acquisition is in the first voice data in the first frequency band and is in described first out-of-band second audio data; Second process is performed to described first voice data, obtains the 3rd voice data; And synthesize described 3rd voice data and described second audio data, voice data after generating process.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said first frequency band is the frequency band in specific frequency range, and described particular frequency range corresponds to the frequency range of people's sounding.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said processing audio data for the treatment of performs the first process and describedly performs the second process to described first voice data, obtains described first voice data of frequency shift and the described second audio data that remains unchanged.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said process pending voice data execution first comprises: perform the first filtering process to pending voice data, to obtain described first voice data in described first frequency band, and the second filtering process is performed to pending voice data, to obtain described first out-of-band described second audio data.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said process described first voice data execution second comprises: the frequency changing described first voice data.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, the frequency of described first voice data of wherein said change comprises: perform the first conversion, to obtain the first audio frequency data corresponding to described first voice data to described first voice data; Change the frequency values of described first audio frequency data, to obtain the 3rd audio frequency data; And the second conversion is performed to described 3rd audio frequency number, to obtain described 3rd voice data, wherein said first conversion is the inverse transformation of described first conversion.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said process described first voice data execution second also comprises: to described first voice data corresponding to predetermined amount of time, perform described second process.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said process described first voice data execution second also comprises: based on the feature of described first voice data, identify in described first voice data and have the first sub-audio data of fisrt feature and have the second sub-audio data of second feature; And the first son process is performed to described first sub-audio data, and the second son process is performed to described second sub-audio data.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, wherein said feature comprises vocal print feature, tamber characteristic and/or tonality feature.
In addition, according to the audio-frequency processing method of an embodiment of the present disclosure, described 3rd voice data of wherein said synthesis and described second audio data, after generating process, voice data comprises: extract the timestamp mark in described 3rd voice data; Mark based on described timestamp, determine that described 3rd voice data is relative to the start time point of described second audio data and termination time point; And based on described start time point and termination time point, aim at and combine described 3rd voice data and described second audio data, with voice data after generating process.
According to another embodiment of the present invention, provide a kind of electronic equipment, comprising: filter unit, perform the first process for treating processing audio data, acquisition is in the first voice data in the first frequency band and is in described first out-of-band second audio data; Modified tone unit, for performing the second process to described first voice data, obtains the 3rd voice data; And synthesis unit, for the synthesis of described 3rd voice data and described second audio data, voice data after generating process.
In addition, electronic equipment according to another embodiment of the present invention, wherein said first frequency band is the frequency band in specific frequency range, and described particular frequency range corresponds to the frequency range of people's sounding.
In addition, electronic equipment according to another embodiment of the present invention, described first voice data that wherein said synthesis unit frequency synthesis changes and the described second audio data remained unchanged.
In addition, electronic equipment according to another embodiment of the present invention, wherein said filter unit comprises the first filtering subelement, for performing the first filtering process to pending voice data, to obtain described first voice data in described first frequency band, and the second filtering process is performed to pending voice data, to obtain described first out-of-band described second audio data.
In addition, electronic equipment according to another embodiment of the present invention, wherein said modified tone unit changes the frequency of described first voice data.
In addition, electronic equipment according to another embodiment of the present invention, wherein said modified tone unit performs the first conversion, to obtain the first audio frequency data corresponding to described first voice data to described first voice data; Change the frequency values of described first audio frequency data, to obtain the 3rd audio frequency data; And the second conversion is performed to described 3rd audio frequency number, to obtain described 3rd voice data,
In addition, electronic equipment according to another embodiment of the present invention, wherein said modified tone unit, to described first voice data corresponding to predetermined amount of time, performs described second process.
In addition, electronic equipment according to another embodiment of the present invention, wherein said modified tone unit, based on the feature of described first voice data, identifies in described first voice data and has the first sub-audio data of fisrt feature and have the second sub-audio data of second feature; And the first son process is performed to described first sub-audio data, and the second son process is performed to described second sub-audio data.
In addition, electronic equipment according to another embodiment of the present invention, wherein said feature comprises vocal print feature, tamber characteristic and/or tonality feature.
In addition, electronic equipment according to another embodiment of the present invention, wherein said synthesis unit extracts the timestamp mark in described 3rd voice data; Mark based on described timestamp, determine that described 3rd voice data is relative to the start time point of described second audio data and termination time point; And based on described start time point and termination time point, aim at and combine described 3rd voice data and described second audio data, with voice data after generating process.
It being understood that general description above and detailed description below both exemplary, and intention is to provide further illustrating of claimed technology.
Accompanying drawing explanation
Be described in more detail the embodiment of the present invention in conjunction with the drawings, above-mentioned and other object of the present invention, Characteristics and advantages will become more obvious.Accompanying drawing is used to provide the further understanding to the embodiment of the present invention, and forms a part for instructions, is used from explanation the present invention, is not construed as limiting the invention with the embodiment of the present invention one.In the accompanying drawings, identical reference number represents same parts or step usually.
Fig. 1 is the process flow diagram of general introduction according to the audio-frequency processing method of the embodiment of the present invention;
Fig. 2 is the structured flowchart of diagram according to the electronic equipment of the embodiment of the present invention;
Fig. 3 is the process flow diagram of diagram according to the first example of the audio-frequency processing method of the embodiment of the present invention;
Fig. 4 is the block diagram of diagram according to the audio processing process of the embodiment of the present invention;
Fig. 5 is the process flow diagram of diagram according to the second example of the audio-frequency processing method of the embodiment of the present invention;
Fig. 6 is the schematic diagram of diagram according to the audio frequency process of the embodiment of the present invention; And
Fig. 7 is the process flow diagram of diagram according to the 3rd example of the audio-frequency processing method of the embodiment of the present invention.
Embodiment
In order to make the object, technical solutions and advantages of the present invention more obvious, describe in detail below with reference to accompanying drawings according to example embodiment of the present invention.Obviously, described embodiment is only a part of embodiment of the present invention, instead of whole embodiment of the present invention, should be understood that the present invention not by the restriction of example embodiment described herein.Based on the embodiment of the present invention described in the disclosure, other embodiments all that those skilled in the art obtain when not paying creative work all should fall within protection scope of the present invention.
Below, the preferred embodiments of the present invention are described in detail with reference to accompanying drawing.
Fig. 1 is the process flow diagram of general introduction according to the audio-frequency processing method of the embodiment of the present invention.Audio-frequency processing method according to the embodiment of the present invention comprises the following steps.
In step S101, treat processing audio data and perform the first process, acquisition is in the first voice data in the first frequency band and is in the first out-of-band second audio data.As follows by what describe in detail, pending voice data is by the voice data gathered according to the electronic equipment of the embodiment of the present invention.In one embodiment of the invention, in described pending voice data, the sound signal being in different frequency bands is respectively comprised.Usually, sound is produced by object vibration, and the essential characteristic key element of sound comprises: tone, loudness of a sound and tone color.Particularly, different vibration frequencies is different tones.In the voice data that collection sound obtains, voice and background have different characteristic, are in different frequency bands.In a preferred embodiment of the invention, described first frequency band is the frequency band in specific frequency range, and described particular frequency range corresponds to the frequency range of people's sounding, such as 64Hz-523Hz.Below, describe in further detail with reference to the accompanying drawings and how to be processed by execution first, acquisition is in the first voice data in the first frequency band and is in the first out-of-band second audio data.After this, process enters step S102.
In step s 102, the second process is performed to the first voice data, obtains the 3rd voice data.As follows by what describe in detail, the second process is performed to the first voice data, obtains described first voice data of frequency shift.That is, described second process is the process for changing the first voice data.Described in step S101, the first voice data is the voice data being in (such as, the frequency range corresponding to people's sounding) in the first frequency band, so processes the frequency of voice in change first voice data by described second.
Such as, usually adopt 12 equal temperances, be divided into sound level CDEFGAB by sound.Certain sound level and the distance on the net between several 8 sound levels are called " pure octave ".A pure octave is divided into 12 impartial semitones by 12 equal temperances, the physical vibration frequency phase-difference 2 between adjacent semitone 1/12doubly.That is, the frequency between each semitone becomes geometric ratio relation.
In one embodiment of the invention, the signal frequency of the first voice data is such as f, and the signal frequency so processing by described second the 3rd voice data obtained is f '
f′=f×2 d/12,d=±1、±2、±3…(1)
As d >, represent the frequency that the second process raises the first voice data, by the first voice data rising tune; As d <, represent the frequency that the second process reduces the first voice data, by the first voice data falling tone.
After described first voice data (that is, the 3rd voice data) obtaining frequency shift and the described second audio data remained unchanged, process enters step S103.
In step s 103, synthesis the 3rd voice data and second audio data, voice data after generating process.Particularly, described first voice data of frequency shift and the constant described second audio data of holding frequency are synthesized, obtains the voice data after process.Below, describe in further detail with reference to the accompanying drawings according to the voice data synthesis in the audio-frequency processing method of the embodiment of the present invention.
As mentioned above, treat compared with processing audio data with what start according to the audio-frequency processing method of the embodiment of the present invention just to input, in voice data after treatment, corresponding in specific frequency range (such as, 64Hz-523Hz corresponding to voice) the first voice data frequency of occurrences change (such as, rising tune or falling tone, namely female voice is converted to boy student, or old man's sound is converted to children's sound), second audio data in unspecific frequency range then holding frequency is constant, thus only accurately change of voice modified tone process has been carried out to voice in voice data after the process of synthesis, and keep background sound constant.
Fig. 2 is the structured flowchart of diagram according to the electronic equipment of the embodiment of the present invention.Electronic equipment 10 is as shown in Figure 2 for performing the audio-frequency processing method according to the embodiment of the present invention described with reference to Fig. 1.
Described electronic equipment 10 is such as preferably the electronic equipment with audio frequency processing power, include but not limited to portable electric appts (such as, smart phone, personal digital assistant, panel computer), personal computer, household audio and video system, commercial Karaoke entertainment systems etc.
As shown in Figure 2, processing module 100 and audio frequency input/output module 200 is comprised according to the electronic equipment 10 of the embodiment of the present invention.It is easily understood that in order to simplified characterization only illustrates the assembly be closely related with the present invention in Fig. 1, the electronic equipment 10 according to the embodiment of the present invention can also comprise other assembly certainly, such as display module, memory module etc.
Described processing module 100 is for performing the audio frequency process according to the embodiment of the present invention.In one embodiment of the invention, described processing module 100 can be configured by the CPU (central processing unit) of electronic equipment 10 (CPU).Alternately, described processing module 100 can be configured by the special audio processing unit (APU) of electronic equipment 10.
Voice data after the process performed via described processing module 100 after audio frequency process for obtaining pending voice data, and exports by described audio frequency input/output module 200.In one embodiment of the invention, described audio frequency input/output module 200 can utilize the audio collection unit such as such as microphone to gather pending voice data.Alternately, described audio frequency input/output module 200 can be retrieved from the memory module (not shown) described electronic equipment 10 and extract pending voice data, or receives pending voice data via wired or radio communication channel from another electronic equipment.After performing audio frequency process via described processing module 100, described audio frequency input/output module 200 can via voice data after the audio output unit output processing such as such as loudspeaker.Alternately, voice data after process can be stored in the memory module in described electronic equipment 10 by described audio frequency input/output module 200, or via wired or radio communication channel, voice data after process is sent to another electronic equipment.
More specifically, as shown in Figure 2, described processing module 100 comprises filter unit 110, modified tone unit 120 and synthesis unit 130.Described filter unit 110 performs the first process for treating processing audio data, and acquisition is in the first voice data in the first frequency band and is in described first out-of-band second audio data.Described modified tone unit 120, for performing the second process to described first voice data, obtains the 3rd voice data.Described synthesis unit 130 for the synthesis of described 3rd voice data and described second audio data, voice data after generating process.Below, the audio-frequency processing method according to the embodiment of the present invention performed by the described processing module 100 comprising filter unit 110, modified tone unit 120 and synthesis unit 130 is described in further detail with reference to the accompanying drawings.
With reference to Fig. 3 and Fig. 4, the first example according to the audio-frequency processing method of the embodiment of the present invention is described.Fig. 3 is the process flow diagram of diagram according to the first example of the audio-frequency processing method of the embodiment of the present invention.Fig. 4 is the block diagram of diagram according to the audio processing process of the embodiment of the present invention.
As shown in Figure 3, comprise the following steps according to the first example of the audio-frequency processing method of the embodiment of the present invention.
In step S301, the first filtering process is performed to pending voice data, to obtain the first voice data in the first frequency band.
After step S301, or the while of with step S301, in step s 302, the second filtering process is performed to pending voice data, to obtain the first out-of-band second audio data.
As shown in Figure 4, in order to obtain the first voice data and second audio data, pending voice data A 0the first filtering subelement 111 and the second filtering subelement respectively in input filter unit 110.Particularly, described first filtering subelement 111 performs the first filtering process (such as, the first bandpass filtering treatment), obtains and exports the first voice data A of (such as, corresponding to the 64Hz-523Hz of voice) in the first frequency band 1.Described second filtering subelement 112 performs the second filtering process (such as, the second bandpass filtering treatment), obtains the first out-of-band second audio data A 2.
Referring back to Fig. 3, after step S302, process enters step S303.In step S303, the first conversion is performed, to obtain the first audio frequency data corresponding to the first voice data to the first voice data.
As shown in Figure 4, from the first voice data A that described first filtering subelement 111 exports 1enter the first varitron unit 121 in modified tone unit 120.Described first varitron unit 121 performs the first conversion to the first voice data.In one embodiment of the invention, described first is transformed to fast Fourier transform, for by the first voice data A 1be converted to the first audio frequency data A of frequency domain 1f.
Referring back to Fig. 3, after step S303, process enters step S304.In step s 304, change the frequency values of the first audio frequency data, to obtain the 3rd audio frequency data.
As shown in Figure 4, from the first audio frequency data A that described first varitron unit 121 exports 1fenter the frequency inverted subelement 122 in modified tone unit 120.Described frequency inverted subelement 122 changes the first audio frequency data A 1ffrequency, to obtain the 3rd audio frequency data A 3f.Such as, if improve the first audio frequency data A 1ffrequency, namely perform rising tune process, then may obtain and male voice is become female voice, old man's sound be become the effect of children's sound; If reduce the first audio frequency data A 1ffrequency, namely perform falling tone process, then may obtain and female voice is become male voice, children's sound be become the effect of old man's sound.
Particularly, for rising tune process, by the first audio frequency data A 1fspectral line stretch to high frequency.
Such as, at the first audio frequency data A 1fspectrum signal in, by performing selection at a certain distance, with realize reduce sampling thus rising tune.Similarly, for falling tone process, by the first audio frequency data A 1fspectral line shrink to low frequency.Such as, at the first audio frequency data A 1fspectrum signal in, by the data point that interpolation between adjacent two data point is new, with realize increase sampling thus falling tone.
Such as, suppose that the modified tone factor is:
&alpha; = f , f = M L - - - ( 2 )
Wherein M and L is positive integer, and the modified tone of any rational number overtones band, can be expressed as:
X 1 ( n ) = X ( &lsqb; n L &rsqb; ) + mod ( n , L ) L &times; { X ( &lsqb; n L &rsqb; + 1 ) - X ( &lsqb; n L &rsqb; ) } , n = 0 , 1 , 2 , ... , ( N - 1 ) &times; L + 1 - - - ( 3 )
Wherein, N is frame length, and [] represents rounding operation, and mod is modulo operation.Realize rising tune as M>L, during M<L, realize falling tone.
Main attention, after modified tone, in order to keep the duration of general audio data constant, needs the data after to modified tone to carry out suitable process.Such as, when rising tune, because rising tune may make raw data shorten, therefore first the data after modified tone aftertreatment can be divided into frame one by one, after each frame, compensate the data of this frame last part.In one embodiment of the invention, the concrete length of the data of compensation can be frame length ( ) doubly.
Referring back to Fig. 3, after step S304, process enters step S305.In step S305, the second conversion is performed, to obtain the 3rd voice data to the 3rd audio frequency number.
As shown in Figure 4, from the 3rd audio frequency data A that frequency inverted subelement 122 exports 3fenter the second varitron unit 123 in modified tone unit 120.Described second varitron unit 123 is to the 3rd audio frequency data A 3fperform the second conversion.In one embodiment of the invention, described second is transformed to inverse fast Fourier transform, for by the 3rd audio frequency data A 3fbe converted to the 3rd voice data A of time domain 3.
Referring back to Fig. 3, after step S305, process enters step S306.In step S306, synthesis the 3rd voice data and second audio data, voice data after generating process.
As shown in Figure 4, from the 3rd voice data A that the second varitron unit 123 modified tone unit 120 exports 3with the second audio data A exported from described second filtering subelement 112 2enter synthesis unit 130.Described synthesis unit 130 is synthesized with voice data A after generating process p.
In the first example of the audio-frequency processing method according to the embodiment of the present invention described with reference to Fig. 3 and Fig. 4, corresponding in specific frequency range (such as, 64Hz-523Hz corresponding to voice) the first voice data perform and change the modified tone process of frequency, keep the overall length of the voice data after process constant to avoid distortion simultaneously.Thus only accurately change of voice modified tone process has been carried out to voice in voice data after the process of synthesis, and keep background sound constant.
With reference to Fig. 5 and Fig. 6, the second example according to the audio-frequency processing method of the embodiment of the present invention is described.Fig. 5 is the process flow diagram of diagram according to the second example of the audio-frequency processing method of the embodiment of the present invention.Fig. 6 is the schematic diagram of diagram according to the audio frequency process of the embodiment of the present invention.
As shown in Figure 5, comprise the following steps according to the second example of the audio-frequency processing method of the embodiment of the present invention.
In step S501, treat processing audio data and perform the first process, acquisition is in the first voice data in the first frequency band and is in the first out-of-band second audio data.Process in step S501 is equivalent to the step S301 and S302 that describe with reference to Fig. 3 and the process performed in the first filtering subelement 111 and the second filtering subelement 112 described with reference to Fig. 4.After this, process enters step S502.
In step S502, to the first voice data corresponding to predetermined amount of time, perform the second process, obtain the 3rd voice data.Be different from the first example of the audio-frequency processing method according to the embodiment of the present invention described with reference to Fig. 3, in the second example of the audio-frequency processing method according to the embodiment of the present invention, only for the first voice data A corresponding to predetermined amount of time 1perform the second process.In one embodiment of the invention, described predetermined amount of time can be specified by user, thus only performs modified tone to the voice of this predetermined amount of time, and for the first voice data A 1in the voice of other periods do not perform modified tone.In the case, at predetermined amount of time, single voice is modified tone (such as, male voice becomes female voice), thus the sound that do not modify tone outside the modified tone sound of predetermined amount of time and predetermined amount of time combines the effect obtaining the antiphonal singing of men and women's sound.After this, process enters step S503.
The building-up process that step S503 to S505 describes corresponding to the step S306 with reference to Fig. 3.At this, schematic diagram by three step S503 to S505 and Fig. 6 is described in further detail this building-up process.
In step S503, extract the timestamp mark in the 3rd voice data.
As shown in Figure 6, described synthesis unit 130 extracts the 3rd voice data A exported via described modified tone unit 120 3timestamp mark T 1and T 2, and via the second audio data A that described filter unit 110 exports 2timestamp mark T 0and T t,
After this, process enters step S504.In step S504, based on timestamp mark, determine that the 3rd voice data is relative to the start time point of second audio data and termination time point.
As shown in Figure 6, described synthesis unit 130 is based on the 3rd voice data A extracted in step S503 3timestamp mark T 1and T 2and second audio data A 2timestamp mark T 0and T t, determine the 3rd voice data A 3relative to second audio data A 2start time point and termination time point.
After this, process enters step S505.In step S505, based on start time point and termination time point, aim at and combine the 3rd voice data and second audio data, with voice data after generating process.
As shown in Figure 6, described synthesis unit 130, based on the start time point obtained in step S504 and termination time point, is aimed at and combines the 3rd voice data A 3with second audio data A 2, with voice data A after generating process p.
In the second example of the audio-frequency processing method according to the embodiment of the present invention described with reference to Fig. 5 and Fig. 6, the modified tone process changing frequency is performed to the first voice data corresponding to predetermined amount of time, achieves the combination of the voice simultaneously comprising modified tone and the voice do not modified tone.In addition, based on the timestamp mark in each voice data in the process of synthesis, achieve the aligning of each voice data accurately, avoid the unnecessary noise because misalignment causes.
With reference to Fig. 7, the 3rd example according to the audio-frequency processing method of the embodiment of the present invention is described.As shown in Figure 7, comprise the following steps according to the 3rd example of the audio-frequency processing method of the embodiment of the present invention.
In step s 701, treat processing audio data and perform the first process, acquisition is in the first voice data in the first frequency band and is in the first out-of-band second audio data.The step S501 that process in step S701 is equivalent to step S301 and the S302 reference Fig. 5 description described with reference to Fig. 3 and the process performed in the first filtering subelement 111 and the second filtering subelement 112 described with reference to Fig. 4.After this, process enters step S702.
Step S702 and step S703 is performed by described modified tone unit 120.Be different from step S303 to the S305 described with reference to Fig. 3 and the step S502 described with reference to Fig. 5, in the 3rd example of the audio-frequency processing method according to the embodiment of the present invention, in step S702, based on the feature of the first voice data, identify in the first voice data first sub-audio data with fisrt feature and second sub-audio data with second feature.In one embodiment of the invention, described feature comprises vocal print feature, tamber characteristic and/or tonality feature.
Such as, described modified tone unit 120, based on the vocal print feature of the first voice data, identifies the different user in the first voice data.Further, when recognizing the first sub-audio data meeting fisrt feature (that is, having specific first vocal print) and/or the second sub-audio data meeting second feature (that is, having specific second vocal print), process enters step S703.
In step S703, described modified tone unit 120 performs the first son process to the first sub-audio data, and performs the second son process to the second sub-audio data, obtains the 3rd voice data.That is, for the first sub-audio data and second sub-audio data with different vocal print feature, described modified tone unit 120 performs the first different son process and the second son process respectively.That is, the specific user achieved for having specific vocal print feature performs the process that specifically modifies tone.
Similarly, described modified tone unit 120, based on the tonality feature of the first voice data, identifies the voice data being in different tone (that is, frequency band) in the first voice data.When recognizing the first sub-audio data meeting fisrt feature (that is, having specific first frequency band) and/or the second sub-audio data meeting second feature (that is, having specific second frequency band), process enters step S703.
In step S703, described modified tone unit 120 performs the first son process to the first sub-audio data, and performs the second son process to the second sub-audio data, obtains the 3rd voice data.That is, for the first sub-audio data and second sub-audio data with different tonality feature, described modified tone unit 120 performs the first different son process and the second son process respectively.Such as, the first sub-audio data for specific first frequency band performs rising tune process, and performs falling tone process for the second sub-audio data of specific second frequency band.
After this, process enters step S704.The synthesis process of step S704 to S706 is identical with step S503 to the S505 described with reference to Fig. 5 respectively, will omit its repeated description at this.
In the 3rd example of the audio-frequency processing method according to the embodiment of the present invention described with reference to Fig. 7, based on the feature of the first voice data, different sub-process is performed for the sub-audio data meeting different characteristic, thus the specific features achieved based on the sound signal gathered, perform the process of corresponding audio frequency.
Describe according to the audio-frequency processing method of the embodiment of the present invention and the electronic equipment using this audio-frequency processing method referring to figs. 1 through Fig. 7 above, it can for the audio frequency process needed for the target audio signal comprised in voice data (such as human voice signal) execution, and keep other non-targeted sound signals constant, thus achieve the accurate audio frequency process for target audio signal.
It should be noted that, in this manual, term " comprises ", " comprising " or its any other variant are intended to contain comprising of nonexcludability, thus make to comprise the process of a series of key element, method, article or equipment and not only comprise those key elements, but also comprise other key elements clearly do not listed, or also comprise by the intrinsic key element of this process, method, article or equipment.When not more restrictions, the key element limited by statement " comprising ... ", and be not precluded within process, method, article or the equipment comprising described key element and also there is other identical element.
Finally, also it should be noted that, above-mentioned a series of process not only comprises with the order described here temporally process that performs of sequence, and comprises process that is parallel or that perform respectively instead of in chronological order.
Through the above description of the embodiments, those skilled in the art can be well understood to the mode that the present invention can add required hardware platform by software and realize, and can certainly all be implemented by hardware.Based on such understanding, what technical scheme of the present invention contributed to background technology can embody with the form of software product in whole or in part, this computer software product can be stored in storage medium, as ROM/RAM, magnetic disc, CD etc., comprising some instructions in order to make a computer equipment (can be personal computer, server, or the network equipment etc.) perform the method described in some part of each embodiment of the present invention or embodiment.
Above to invention has been detailed introduction, applying specific case herein and setting forth principle of the present invention and embodiment, the explanation of above embodiment just understands method of the present invention and core concept thereof for helping; Meanwhile, for one of ordinary skill in the art, according to thought of the present invention, all will change in specific embodiments and applications, in sum, this description should not be construed as limitation of the present invention.

Claims (20)

1. an audio-frequency processing method, comprising:
Treat processing audio data and perform the first process, acquisition is in the first voice data in the first frequency band and is in described first out-of-band second audio data;
Second process is performed to described first voice data, obtains the 3rd voice data; And
Synthesize described 3rd voice data and described second audio data, voice data after generating process.
2. audio-frequency processing method as claimed in claim 1, wherein said first frequency band is the frequency band in specific frequency range, and described particular frequency range corresponds to the frequency range of people's sounding.
3. audio-frequency processing method as claimed in claim 1 or 2, wherein said processing audio data for the treatment of performs the first process and describedly performs the second process to described first voice data, obtains described first voice data of frequency shift and the described second audio data that remains unchanged.
4. audio-frequency processing method as claimed in claim 1 or 2, wherein said process pending voice data execution first comprises:
First filtering process is performed to pending voice data, to obtain described first voice data in described first frequency band, and the second filtering process is performed to pending voice data, to obtain described first out-of-band described second audio data.
5. audio-frequency processing method as claimed in claim 1 or 2, wherein said process described first voice data execution second comprises:
Change the frequency of described first voice data.
6. audio-frequency processing method as claimed in claim 5, the frequency of described first voice data of wherein said change comprises:
First conversion is performed, to obtain the first audio frequency data corresponding to described first voice data to described first voice data;
Change the frequency values of described first audio frequency data, to obtain the 3rd audio frequency data; And
Second conversion is performed to described 3rd audio frequency number, to obtain described 3rd voice data,
Wherein said first conversion is the inverse transformation of described first conversion.
7. audio-frequency processing method as claimed in claim 1 or 2, wherein said process described first voice data execution second also comprises:
To described first voice data corresponding to predetermined amount of time, perform described second process.
8. audio-frequency processing method as claimed in claim 7, wherein said process described first voice data execution second also comprises:
Based on the feature of described first voice data, identify in described first voice data and there is the first sub-audio data of fisrt feature and there is the second sub-audio data of second feature; And
First son process is performed to described first sub-audio data, and the second son process is performed to described second sub-audio data.
9. audio-frequency processing method as claimed in claim 8, wherein said feature comprises vocal print feature, tamber characteristic and/or tonality feature.
10. audio-frequency processing method as claimed in claim 1 or 2, described 3rd voice data of wherein said synthesis and described second audio data, after generating process, voice data comprises:
Extract the timestamp mark in described 3rd voice data;
Mark based on described timestamp, determine that described 3rd voice data is relative to the start time point of described second audio data and termination time point; And
Based on described start time point and termination time point, aim at and combine described 3rd voice data and described second audio data, with voice data after generating process.
11. 1 kinds of electronic equipments, comprising:
Filter unit, perform the first process for treating processing audio data, acquisition is in the first voice data in the first frequency band and is in described first out-of-band second audio data;
Modified tone unit, for performing the second process to described first voice data, obtains the 3rd voice data; And
Synthesis unit, for the synthesis of described 3rd voice data and described second audio data, voice data after generating process.
12. electronic equipments as claimed in claim 11, wherein said first frequency band is the frequency band in specific frequency range, and described particular frequency range corresponds to the frequency range of people's sounding.
13. electronic equipments as described in claim 11 or 12, described first voice data that wherein said synthesis unit frequency synthesis changes and the described second audio data remained unchanged.
14. electronic equipments as described in claim 11 or 12, wherein said filter unit comprises the first filtering subelement, for performing the first filtering process to pending voice data, to obtain described first voice data in described first frequency band, and the second filtering process is performed to pending voice data, to obtain described first out-of-band described second audio data.
15. electronic equipments as described in claim 11 or 12, wherein said modified tone unit changes the frequency of described first voice data.
16. electronic equipments as claimed in claim 15, wherein said modified tone unit performs the first conversion, to obtain the first audio frequency data corresponding to described first voice data to described first voice data;
Change the frequency values of described first audio frequency data, to obtain the 3rd audio frequency data; And
Second conversion is performed, to obtain described 3rd voice data to described 3rd audio frequency number.
17. electronic equipments as described in claim 11 or 12, wherein said modified tone unit, to described first voice data corresponding to predetermined amount of time, performs described second process.
18. electronic equipments as claimed in claim 17, wherein said modified tone unit, based on the feature of described first voice data, identifies in described first voice data and has the first sub-audio data of fisrt feature and have the second sub-audio data of second feature; And
First son process is performed to described first sub-audio data, and the second son process is performed to described second sub-audio data.
19. electronic equipments as claimed in claim 18, wherein said feature comprises vocal print feature, tamber characteristic and/or tonality feature.
20. electronic equipments as described in claim 11 or 12, wherein said synthesis unit extracts the timestamp mark in described 3rd voice data;
Mark based on described timestamp, determine that described 3rd voice data is relative to the start time point of described second audio data and termination time point; And
Based on described start time point and termination time point, aim at and combine described 3rd voice data and described second audio data, with voice data after generating process.
CN201510612358.4A 2015-09-23 2015-09-23 Audio processing method and electronic equipment Active CN105321526B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510612358.4A CN105321526B (en) 2015-09-23 2015-09-23 Audio processing method and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510612358.4A CN105321526B (en) 2015-09-23 2015-09-23 Audio processing method and electronic equipment

Publications (2)

Publication Number Publication Date
CN105321526A true CN105321526A (en) 2016-02-10
CN105321526B CN105321526B (en) 2020-07-24

Family

ID=55248774

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510612358.4A Active CN105321526B (en) 2015-09-23 2015-09-23 Audio processing method and electronic equipment

Country Status (1)

Country Link
CN (1) CN105321526B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105869621A (en) * 2016-05-20 2016-08-17 广州华多网络科技有限公司 Audio synthesizing device and audio synthesizing method applied to same
CN106128474A (en) * 2016-07-04 2016-11-16 广东小天才科技有限公司 Audio processing method and device
CN107707974A (en) * 2017-09-18 2018-02-16 广东九联科技股份有限公司 A kind of realization method and system of special efficacy voice function
CN108965757A (en) * 2018-08-02 2018-12-07 广州酷狗计算机科技有限公司 video recording method, device, terminal and storage medium
CN111210833A (en) * 2019-12-30 2020-05-29 联想(北京)有限公司 Audio processing method, electronic device, and medium
CN113409801A (en) * 2021-08-05 2021-09-17 云从科技集团股份有限公司 Noise processing method, system, medium, and apparatus for real-time audio stream playback

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001306100A (en) * 2000-04-25 2001-11-02 Matsushita Electric Works Ltd Voice conversion system
CN1967657A (en) * 2005-11-18 2007-05-23 成都索贝数码科技股份有限公司 Automatic tracking and tonal modification system of speaker in program execution and method thereof
CN101354889A (en) * 2008-09-18 2009-01-28 北京中星微电子有限公司 Method and apparatus for tonal modification of voice
CN102592607A (en) * 2012-03-30 2012-07-18 北京交通大学 Voice converting system and method using blind voice separation
CN102982809A (en) * 2012-12-11 2013-03-20 中国科学技术大学 Conversion method for sound of speaker
CN103280215A (en) * 2013-05-28 2013-09-04 北京百度网讯科技有限公司 Audio frequency feature library establishing method and device
CN103310796A (en) * 2013-06-28 2013-09-18 姜鸿彦 Voice signal extraction method
CN104078051A (en) * 2013-03-29 2014-10-01 中兴通讯股份有限公司 Voice extracting method and system and voice audio playing method and device
CN104704558A (en) * 2012-09-14 2015-06-10 杜比实验室特许公司 Multi-channel audio content analysis based upmix detection
CN104916288A (en) * 2014-03-14 2015-09-16 深圳Tcl新技术有限公司 Human voice highlighting processing method and device in audio

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001306100A (en) * 2000-04-25 2001-11-02 Matsushita Electric Works Ltd Voice conversion system
CN1967657A (en) * 2005-11-18 2007-05-23 成都索贝数码科技股份有限公司 Automatic tracking and tonal modification system of speaker in program execution and method thereof
CN101354889A (en) * 2008-09-18 2009-01-28 北京中星微电子有限公司 Method and apparatus for tonal modification of voice
CN102592607A (en) * 2012-03-30 2012-07-18 北京交通大学 Voice converting system and method using blind voice separation
CN104704558A (en) * 2012-09-14 2015-06-10 杜比实验室特许公司 Multi-channel audio content analysis based upmix detection
CN102982809A (en) * 2012-12-11 2013-03-20 中国科学技术大学 Conversion method for sound of speaker
CN104078051A (en) * 2013-03-29 2014-10-01 中兴通讯股份有限公司 Voice extracting method and system and voice audio playing method and device
CN103280215A (en) * 2013-05-28 2013-09-04 北京百度网讯科技有限公司 Audio frequency feature library establishing method and device
CN103310796A (en) * 2013-06-28 2013-09-18 姜鸿彦 Voice signal extraction method
CN104916288A (en) * 2014-03-14 2015-09-16 深圳Tcl新技术有限公司 Human voice highlighting processing method and device in audio

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
曹强: "《数字音频规范与程序设计》", 30 June 2012, 中国水利水电出版社 *
蔡杰: "一种改进的时域音频变调方法及其软件实现", 《电声基础》 *
赵力: "《语音信号处理》", 31 May 2009, 机械工业出版社 *
陈启美 等: "《现代数据通信教程》", 28 February 2006, 南京大学出版社 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105869621A (en) * 2016-05-20 2016-08-17 广州华多网络科技有限公司 Audio synthesizing device and audio synthesizing method applied to same
CN105869621B (en) * 2016-05-20 2019-10-25 广州华多网络科技有限公司 Audio synthesizer and its audio synthetic method
CN106128474A (en) * 2016-07-04 2016-11-16 广东小天才科技有限公司 Audio processing method and device
CN107707974A (en) * 2017-09-18 2018-02-16 广东九联科技股份有限公司 A kind of realization method and system of special efficacy voice function
CN108965757A (en) * 2018-08-02 2018-12-07 广州酷狗计算机科技有限公司 video recording method, device, terminal and storage medium
CN111210833A (en) * 2019-12-30 2020-05-29 联想(北京)有限公司 Audio processing method, electronic device, and medium
CN113409801A (en) * 2021-08-05 2021-09-17 云从科技集团股份有限公司 Noise processing method, system, medium, and apparatus for real-time audio stream playback
CN113409801B (en) * 2021-08-05 2024-03-19 云从科技集团股份有限公司 Noise processing method, system, medium and device for real-time audio stream playing

Also Published As

Publication number Publication date
CN105321526B (en) 2020-07-24

Similar Documents

Publication Publication Date Title
CN105321526A (en) Audio processing method and electronic device
CN1893412B (en) Frequency extension of harmonic signals
CN102264022B (en) Stability improvements in hearing aids
Goodwin Residual modeling in music analysis-synthesis
RU2487426C2 (en) Apparatus and method for converting audio signal into parametric representation, apparatus and method for modifying parametric representation, apparatus and method for synthensising parametrick representation of audio signal
CN100474939C (en) Sounder signal processor and method
US20200211577A1 (en) Method and device for audio signal processing, and storage medium
CN112506341B (en) Vibration effect generation method and device, terminal equipment and storage medium
CN103109317A (en) Masking sound outputting device, and masking sound outputting means
CN108922558B (en) Voice processing method, voice processing device and mobile terminal
CN107248415A (en) Alarm clock ringtone generation method and device and user terminal
CN106775568A (en) Sound effect adjusting method and device and mobile terminal
CN102484759A (en) Processing audio signals
CN111667803B (en) Audio processing method and related products
CN109416911A (en) Speech synthesizing device and speech synthesizing method
CN113035167A (en) Audio frequency tuning method and storage medium for active noise reduction
Gupta Fourier transform and its application in cell phones
CN105575414A (en) Generating method and device of lyric file
CN109841232B (en) Method and device for extracting note position in music signal and storage medium
CN112837670A (en) Voice synthesis method and device and electronic equipment
CN110097618A (en) A kind of control method, device, vehicle and the storage medium of music animation
JP6078358B2 (en) Noise reduction device, broadcast reception device, and noise reduction method
CN110992969B (en) Filter bank configuration method and device of electronic cochlea
CN109995941B (en) Data adjusting method, device and storage medium
CN112997511B (en) Generating harmonics in an audio system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant