CN117275527A - Sliding sound audio simulation method and device, storage medium and electronic equipment - Google Patents

Sliding sound audio simulation method and device, storage medium and electronic equipment Download PDF

Info

Publication number
CN117275527A
CN117275527A CN202310922550.8A CN202310922550A CN117275527A CN 117275527 A CN117275527 A CN 117275527A CN 202310922550 A CN202310922550 A CN 202310922550A CN 117275527 A CN117275527 A CN 117275527A
Authority
CN
China
Prior art keywords
target
audio
frequency
sliding
intensity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310922550.8A
Other languages
Chinese (zh)
Inventor
张子健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202310922550.8A priority Critical patent/CN117275527A/en
Publication of CN117275527A publication Critical patent/CN117275527A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/75Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 for modelling vocal tract parameters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H7/00Instruments in which the tones are synthesised from a data store, e.g. computer organs
    • G10H7/08Instruments in which the tones are synthesised from a data store, e.g. computer organs by calculating functions or polynomial approximations to evaluate amplitudes at successive sample points of a tone waveform
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters

Abstract

The application discloses a sliding sound audio simulation method and device, a storage medium and electronic equipment. Wherein the method comprises the following steps: acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio; acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio; determining a target intensity profile based on the first intensity profile and the second intensity profile; and responding to the sliding sound operation triggered by the target single sound identification, and utilizing the target strength change curve to change the energy or the amplitude of the target single sound audio matched with the target single sound identification so as to simulate the target sliding sound audio. The method and the device solve the technical problem that the simulation efficiency of the sliding sound audio frequency is low.

Description

Sliding sound audio simulation method and device, storage medium and electronic equipment
Technical Field
The present invention relates to the field of computers, and in particular, to a method and apparatus for simulating a sliding sound audio, a storage medium, and an electronic device.
Background
In the analog scene of the sliding audio, as many audio samples as possible are required to be recorded, however, because the combination of music is endless, the problem that the analog efficiency of the sliding audio is low because the required audio sample amount is massive is solved by recording alone. Therefore, there is a problem in that the analog efficiency of the slide sound audio is low.
In view of the above problems, no effective solution has been proposed at present.
Disclosure of Invention
The embodiment of the application provides a method and a device for simulating sliding sound audio, a storage medium and electronic equipment, and aims to at least solve the technical problem that the sliding sound audio is low in simulation efficiency.
According to an aspect of the embodiments of the present application, there is provided a method for simulating a sliding sound audio, including: acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio; acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies; determining a target intensity variation curve based on the first intensity variation curve and the second intensity variation curve, wherein the target intensity variation curve is used for representing energy or amplitude variation of the single-tone sample audio frequency generated by sliding sounds at different frequencies; and responding to the sliding sound operation triggered by the target single sound identification, and utilizing the target strength change curve to change the energy or the amplitude of the target single sound frequency matched with the target single sound identification so as to simulate the target sliding sound frequency.
According to another aspect of the embodiments of the present application, there is also provided an apparatus for simulating a sliding sound audio, including: the first acquisition unit is used for acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio; a second obtaining unit, configured to obtain a first intensity variation curve corresponding to the sliding-tone sample audio and a second intensity variation curve corresponding to the single-tone sample audio, where the first intensity variation curve is used to represent energy or amplitude of the sliding-tone sample audio at different frequencies, and the second intensity variation curve is used to represent energy or amplitude of the single-tone sample audio at different frequencies; a determining unit configured to determine a target intensity profile based on the first intensity profile and the second intensity profile, where the target intensity profile is used to represent energy or amplitude changes of the single-tone sample audio generated by a sliding sound at different frequencies; and the simulation unit is used for responding to the sliding sound operation triggered by the target single sound identification, and performing energy or amplitude change on the target single sound audio matched with the target single sound identification by utilizing the target strength change curve to simulate the target sliding sound audio.
As an alternative, the analog unit includes: the first acquisition module is used for acquiring each target frequency associated with the target single-tone audio; the first determining module is used for determining intensity change information matched with each target frequency from the target intensity change curve, wherein the intensity change information is used for energy or amplitude change of the target single-tone audio frequency, which is generated by sliding sound, at each target frequency; and the first simulation module is used for simulating the target sliding sound audio by using the intensity change information and gain/attenuation of the energy or the amplitude of the target single sound audio at each target frequency, wherein the energy or the amplitude of the target sliding sound audio at each target frequency and the energy or the amplitude of the target single sound audio at each target frequency accord with the intensity change information.
As an alternative, the second obtaining unit includes: a second obtaining module, configured to obtain the first intensity variation curve for representing energy or amplitude of the sliding sound sample audio at N key frequencies, and the second intensity variation curve for representing energy or amplitude of the single sound sample audio at the N key frequencies, where the N key frequencies include the target frequency.
As an optional solution, the second obtaining module includes: the first transformation submodule is used for carrying out short-time Fourier transformation on the sliding sound sample audio to obtain a first intensity data matrix of the total frequency of the sliding sound sample audio at each time point; determining the first intensity change curves corresponding to the N key frequencies from the first intensity data matrix; and a second transformation submodule, configured to perform the short-time fourier transform on the single-tone sample audio to obtain a second intensity data matrix of the full-scale frequency of the single-tone sample audio at each time point; and determining the second intensity change curves corresponding to the N key frequencies from the second intensity data matrix.
As an alternative, the determining unit includes: and the subtraction module is used for obtaining the target intensity change curve by subtracting the values of the first intensity change curve and the second intensity change curve point by point.
As an alternative, the apparatus further includes: a third obtaining module, configured to obtain an initial sliding sound sample audio before obtaining the target intensity variation curve by subtracting the values of the first intensity variation curve and the second intensity variation curve point by point, where an initial pitch of the initial sliding sound sample audio is the same as an initial pitch of the single sound sample audio; and the processing module is used for processing the initial sliding sound sample audio into the sliding sound sample audio with the fundamental frequency variation less than or equal to a preset threshold before the target intensity variation curve is obtained by subtracting the values of the first intensity variation curve and the second intensity variation curve point by point.
As an alternative, the processing module includes: the acquisition sub-module is used for acquiring a fundamental frequency change curve corresponding to the initial sliding sound sample audio, wherein the fundamental frequency change curve is used for representing the change condition of the fundamental frequency of the initial sliding sound sample audio along with time; the dividing sub-module is used for dividing the sliding sound sample audio into a plurality of audio paragraphs according to the time axis resolution of the fundamental frequency change curve; the conversion sub-module is used for reversely converting the fundamental frequency change curve into an inversely proportional sample density change curve; and the sampling sub-module is used for carrying out inverse resampling on the plurality of audio paragraphs according to the inversely proportional sample density change curve, and supplementing new audio paragraphs through the inverse resampling to obtain the sliding sound sample audio.
As an alternative, the apparatus further includes: and the scaling sub-module is used for scaling a second time axis of the first intensity change curve according to a first time axis of the second intensity change curve before the target intensity change curve is obtained by subtracting the values of the first intensity change curve and the second intensity change curve point by point, wherein the first time axis corresponds to the scaled second time axis.
As an alternative, the apparatus further includes: the sampling unit is used for carrying out energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, and respectively carrying out response sampling on an entity musical instrument to be simulated by using a sweep frequency test signal and a white noise test signal before simulating the target single-tone audio frequency to obtain a sweep frequency response signal and an impact response signal, wherein the sweep frequency test signal is a signal with continuously changed signal frequency in a certain frequency range, the white noise test signal is a random signal with uniform frequency spectrum distribution, the sweep frequency response signal is a response signal after the entity musical instrument receives the sweep frequency test signal, and the impact response signal is a response signal after the entity musical instrument receives the white noise test signal; a difference unit, configured to perform energy or amplitude change on the target tone audio frequency matched with the target tone identifier by using the target intensity change curve, and obtain a frequency intensity difference between the frequency sweep response signal and the impact response signal before simulating the target tone audio frequency, where the frequency intensity difference is used to represent a frequency spectrum characteristic of resonance of the physical musical instrument; the device further comprises: the combination unit is used for carrying out energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, acquiring the current pitch of the target single-tone audio frequency in the process of simulating the target single-tone audio frequency, and determining convolution operation parameters by combining the frequency intensity difference; and the first integration unit is used for carrying out energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, and simulating the target single-tone audio frequency in the process of simulating the target single-tone audio frequency.
As an alternative, the apparatus further includes: a third obtaining unit, configured to obtain an operation acceleration corresponding to a target operation triggered by the target tone identifier in a process of performing energy or amplitude change on the target tone audio frequency matched with the target tone identifier by using the target intensity change curve and simulating a target tone audio frequency, where the target operation includes the tone sliding operation; a fourth obtaining unit, configured to obtain noise audio frequency matched with the operation acceleration in a process of performing energy or amplitude change on the target tone audio frequency matched with the target tone identifier by using the target intensity change curve and simulating a target sliding tone audio frequency; and the second integration unit is used for carrying out energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, and carrying out integration processing on the noise audio frequency and the alternative single-tone audio frequency under the condition that the energy or amplitude change is carried out on the target single-tone audio frequency by utilizing the target intensity change curve to obtain the alternative single-tone audio frequency in the process of simulating the target single-tone audio frequency, so as to simulate the target single-tone audio frequency.
According to yet another aspect of embodiments of the present application, there is provided a computer program product or computer program comprising computer instructions stored in a computer readable storage medium. The processor of the computer device reads the computer instructions from the computer-readable storage medium, and the processor executes the computer instructions so that the computer device performs the simulation method of the sliding sound audio as above.
According to still another aspect of the embodiments of the present application, there is further provided an electronic device including a memory, a processor, and a computer program stored on the memory and executable on the processor, wherein the processor executes the above-mentioned method for simulating the sliding sound audio through the computer program.
In the embodiment of the application, a sliding sound sample audio and a single sound sample audio are obtained, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio; acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies; determining a target intensity variation curve based on the first intensity variation curve and the second intensity variation curve, wherein the target intensity variation curve is used for representing energy or amplitude variation of the single-tone sample audio frequency generated by sliding sounds at different frequencies; and responding to the sliding sound operation triggered by the target single sound identification, and utilizing the target strength change curve to change the energy or the amplitude of the target single sound frequency matched with the target single sound identification so as to simulate the target sliding sound frequency. The intensity change curve expresses the energy or amplitude change of the single-tone audio frequency on different frequencies caused by the sliding sound, and the sliding sound audio frequency is simulated by utilizing the energy or amplitude change caused by the sliding sound, so that the aim of reducing the use quantity of the sliding sound sample audio frequency and the single-tone sample audio frequency in the sliding sound audio frequency simulation process is fulfilled, the technical effect of improving the simulation efficiency of the sliding sound audio frequency is realized, and the technical problem of lower simulation efficiency of the sliding sound audio frequency is solved.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this application, illustrate embodiments of the application and together with the description serve to explain the application and do not constitute an undue limitation to the application. In the drawings:
FIG. 1 is a schematic illustration of an application environment of an alternative method of simulating slide sound audio according to an embodiment of the present application;
FIG. 2 is a schematic illustration of a flow of an alternative method of simulating a slide sound audio according to an embodiment of the present application;
FIG. 3 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 4 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 5 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 6 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 7 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 8 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 9 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 10 is a schematic diagram of an alternative method of simulating a sliding sound audio according to an embodiment of the present application;
FIG. 11 is a schematic diagram of an alternative slide tone audio simulation device according to an embodiment of the present application;
fig. 12 is a schematic structural view of an alternative electronic device according to an embodiment of the present application.
Detailed Description
In order to make the present application solution better understood by those skilled in the art, the following description will be made in detail and with reference to the accompanying drawings in the embodiments of the present application, it is apparent that the described embodiments are only some embodiments of the present application, not all embodiments. All other embodiments, which can be made by one of ordinary skill in the art based on the embodiments herein without making any inventive effort, shall fall within the scope of the present application.
It should be noted that the terms "first," "second," and the like in the description and claims of the present application and the above figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that embodiments of the present application described herein may be implemented in sequences other than those illustrated or otherwise described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
According to an aspect of the embodiments of the present application, there is provided a method for simulating a sliding sound audio, optionally, as an optional implementation manner, the method for simulating a sliding sound audio may be applied, but is not limited to, in an environment as shown in fig. 1. Which may include, but is not limited to, a user device 102 and a server 112, which may include, but is not limited to, a display 104, a processor 106, and a memory 108, the server 112 including a database 114 and a processing engine 116.
The specific process comprises the following steps:
step S102, the user equipment 102 obtains a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as the initial pitch of the single sound sample audio;
step S104-S106, the sliding sound sample audio and the single sound sample audio are sent to the server 112 through the network 110;
step S108-S110, the server 112 obtains a first intensity variation curve corresponding to the sliding sound sample audio and a second intensity variation curve corresponding to the single sound sample audio through the processing engine 116, and further determines a target intensity variation curve based on the first intensity variation curve and the second intensity variation curve;
steps S112-S114, transmitting the target intensity profile variation to the user equipment 102 via the network 110;
In steps S116-S118, the user equipment 102 obtains a sliding operation triggered by the target tone identifier, and further responds to the sliding operation triggered by the target tone identifier through the processor 106, and uses the target intensity variation curve to perform energy or amplitude variation on the target tone audio matched by the target tone identifier, simulate out the target sliding audio, display audio information corresponding to the target sliding audio on the display 104, and store the target sliding audio in the memory 108.
In addition to the example shown in fig. 1, the above steps may be performed by the user device or the server independently, or by the user device and the server cooperatively, such as by the user device 102 performing the steps of S108-S110 described above, to relieve the processing pressure of the server 112. The user device 102 includes, but is not limited to, a handheld device (e.g., a mobile phone), a notebook computer, a tablet computer, a desktop computer, a vehicle-mounted device, a smart television, etc., and the present application is not limited to a specific implementation of the user device 102. The server 112 may be a single server or a server cluster composed of a plurality of servers, or may be a cloud server.
Alternatively, as an alternative implementation manner, as shown in fig. 2, the method for simulating the sliding voice frequency may be performed by an electronic device, such as the user device or the server shown in fig. 1, and the specific steps include:
S202, acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitches of the sliding sound sample audio and the single sound sample audio are the same;
s204, acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies;
s206, determining a target intensity change curve based on the first intensity change curve and the second intensity change curve, wherein the target intensity change curve is used for representing energy or amplitude change of the single-tone sample audio frequency, which is generated by sliding sounds at different frequencies;
and S208, responding to the sliding sound operation triggered by the target single-tone identification, and utilizing a target strength change curve to change the energy or the amplitude of the target single-tone audio matched with the target single-tone identification so as to simulate the target sliding sound audio.
Alternatively, in the present embodiment, the above-mentioned method for simulating the sliding sound audio may be, but not limited to, applied to an application scenario of the guqin simulation software. The playing technique of the ancient musical instrument can be generally divided into fixed tones and hand-moving tones, and is characterized by the hand-moving tones or the speaking sliding tones. The fixed tone principle is simpler, but the smooth tone is very complex in the playing process, so that mature seven-stringed plucked instrument simulation software is lacking, and a user can play the seven-stringed plucked instrument in real time and truly through a virtual interface. The embodiment provides a software virtual playing real-time algorithm for converting the single tone of the ancient piano into the sliding tone with any length and any amplitude, so that a user can more simply and freely touch and play the ancient piano. In addition, the present embodiment can be applied not only to the sliding sound simulation of Yu Guqin but also to the sliding sound simulation of other similar string-type musical instruments, such as urheen, violin, cello, and the like.
Alternatively, in the present embodiment, the sliding sound may refer to, but is not limited to, sliding through a series of notes or ranges in a continuous and smooth manner during a musical performance, and is generally used in a performance of a keyboard instrument, a string instrument, a wind instrument, or the like, and may be added to a musical composition to increase expressivity and musical effect. The slide sample audio may then be, but is not limited to, an audio sample acquired during a series of notes or fields slid over in a continuous and smooth manner during a musical performance.
Alternatively, in the present embodiment, a single tone may refer to, but is not limited to, only one sound part or only one sound in music being played independently. Further, the slide sample audio may be, but is not limited to, an audio sample collected in a single sound during an independent performance during a musical performance.
Alternatively, in this embodiment, the starting pitch may refer to, but is not limited to, the pitch location of a certain note or sequence in the music, which indicates the position of the note or sequence on the music chart, to determine from which pitch the performance should be started when the performance is performed. The starting pitch is typically represented by music notation, e.g., using a standard staff, and the specific pitch can be determined by the clef and the location of the notes. In different modes and musical compositions, the initial pitch is changed to determine the tone and the range of the voice domain of the music;
Further, pitch may refer to, but is not limited to, the attribute of the level of a tone that is subjectively perceived by a user, in relation to the frequency of the sound wave. The higher the pitch, the higher the corresponding frequency; the lower the pitch, the lower the corresponding frequency. In music, standard pitch A4 (440 Hz) is commonly used as a reference, and the pitches of other notes are relatively measured and represented at half-tone or full-tone intervals. The fundamental frequency may refer, but is not limited to, the most dominant frequency component in a sound, the pitch of which determines the pitch of the sound (pitch). In physics and acoustics, each sound can be considered to consist of a series of harmonics. The fundamental frequency is the lowest frequency component of these harmonics and is the frequency most easily perceived by the human ear. The fundamental frequency may, but is not limited to, determine the pitch of the sound, with higher fundamental frequencies corresponding to higher pitches and lower fundamental frequencies corresponding to lower pitches.
Alternatively, in this embodiment, the frequency may refer to, but not limited to, the number of times of occurrence of periodic vibration in the sound or signal, but may refer to, but not limited to, a specific frequency by means of a frequency point, and for the energy or amplitude magnitude on the frequency, it may also be understood, but not limited to, the energy or amplitude magnitude on the frequency point.
Alternatively, in the present embodiment, the magnitudes of the energy or amplitude at the different frequencies are typically determined by the spectral characteristics of the signal, where the spectrum is a representation describing the energy distribution of the signal at the different frequencies. Spectrograms are commonly used to visualize such energy distributions. In the spectrogram, the horizontal axis represents frequency, and the vertical axis represents energy or amplitude. For a particular sound or signal, it is possible to have higher energy or amplitude at some frequencies and lower energy or amplitude at other frequencies. Depending on the frequency content of the signal and the intensity occupied by a particular sound or signal.
Taking music as an example, sounds played by different music instruments tend to have different energy distributions at different frequencies. For example, bass violins typically produce higher energy sounds in the low frequency range, while violins produce higher energy sounds in the mid-high frequency range. Thus, the user may perceive a difference in tone color audibly.
Optionally, in this embodiment, the energy or amplitude at different frequencies may be understood as, but not limited to, harmonic intensity, and the energy or amplitude at a fundamental frequency may be understood as, but not limited to, fundamental frequency intensity, where the fundamental frequency generally varies with the pitch, and thus the sliding sample audio whose pitch varies with time changes with respect to the fundamental frequency, and the single sample audio whose pitch varies with time changes with respect to the fundamental frequency, so that the sliding sample audio corresponds to a plurality of fundamental frequencies, and the different fundamental frequencies correspond to respective harmonics, and the single sample audio corresponds to only one fundamental frequency, so that the difference in harmonic intensity between the sliding sample audio and the single sample audio (the energy or amplitude variation generated by the sliding at different frequencies) needs to be processed into the same dimension, such as unified into the same unique fundamental frequency, otherwise, the difference in harmonic intensity cannot be obtained smoothly.
Optionally, in this embodiment, as shown in fig. 3, a plurality of tone marks, such as tone mark 1, tone mark 2, tone mark 3, tone mark 4, tone mark 5, tone mark 6, are displayed on the analog interface 304 of the virtual musical instrument 302, a tone operation triggered on the tone mark 6 (target tone mark) is further obtained, and during the execution of the tone operation, a tone sliding operation triggered on the tone mark 6 (target tone mark) is obtained, and the target tone sliding audio simulated for the virtual musical instrument 302 is further output.
Optionally, in this embodiment, the simulated target sliding sound includes, in addition to energy or amplitude variation generated by sliding sound, sound factors such as pitch, noise, resonance, etc. may be integrated into the target sliding sound, for example, resampling (resampling) is performed on a single sound, and by inserting an intermediate value, the density of a sampling point in a current Buffer (Buffer) is changed in real time to implement lossless transposition, so as to complete simulation on the pitch, where the reference may refer, but not limited to, changing the sampling rate of data, that is, changing the time or space interval between data points to adapt to different requirements or signal processing operations, and the process of the reference may refer, but not limited to, techniques such as interpolation, filtering, truncation, etc. to maintain signal quality and prevent the introduction of aliasing errors (aliasing).
It should be noted that, the target sliding voice audio frequency simulated in this embodiment does not directly output the target sliding voice audio frequency with the target single voice identifier matched in advance, but changes the energy or the amplitude of the target single voice audio frequency in real time according to the real-time audio frequency change of the target single voice audio frequency with the target single voice identifier matched, and further, no matter what kind of change occurs in the target single voice audio frequency, the real-time energy or the amplitude change can be performed according to the real-time audio frequency.
By way of further illustration, as shown in FIG. 4, optionally, by performance operation on the physical instrument 402, a slide sample audio 404-1 and a single sample audio 404-2 are collected, wherein the slide sample audio 404-1 and the single sample audio 404-2 have the same starting pitch; acquiring a first intensity variation curve 406-1 corresponding to the sliding sound sample audio 404-1 and a second intensity variation curve 406-2 corresponding to the single sound sample audio 404-2; determining a target intensity profile 408 based on the first intensity profile 406-1 and the second intensity profile 406-2; in response to a sliding operation triggered by the target tone identifier 412 displayed on the corresponding analog interface 410 of the physical instrument 402, the target tone audio 416 is simulated by using the target intensity variation curve 408 to perform an energy or amplitude variation on the target tone audio 414 matched by the target tone identifier 412.
According to the embodiment provided by the application, the sliding sound sample audio and the single sound sample audio are obtained, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio; acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies; determining a target intensity change curve based on the first intensity change curve and the second intensity change curve, wherein the target intensity change curve is used for representing energy or amplitude change of the single-tone sample audio frequency, which is generated by sliding sounds at different frequencies; and responding to the sliding sound operation triggered by the target single sound identification, and utilizing the target strength change curve to change the energy or the amplitude of the target single sound audio matched with the target single sound identification so as to simulate the target sliding sound audio. The intensity change curve expresses the energy or amplitude change of the single-tone audio frequency on different frequencies caused by the sliding sound, and the sliding sound audio frequency is simulated by utilizing the energy or amplitude change caused by the sliding sound, so that the aim of reducing the number of the sliding sound sample audio frequency and the single-tone sample audio frequency in the sliding sound audio frequency simulation process is fulfilled, and the technical effect of improving the simulation efficiency of the sliding sound audio frequency is realized.
As an alternative, using the target intensity variation curve to perform energy or amplitude variation on the target tone audio frequency matched with the target tone identifier, simulating the target sliding tone audio frequency, including:
s1-1, acquiring each target frequency associated with target single-tone audio;
s1-2, determining intensity change information matched with each target frequency from a target intensity change curve, wherein the intensity change information is used for energy or amplitude change of target single-tone audio frequency on each target frequency due to sliding sound;
s1-3, using the intensity change information, gain/attenuation target single-tone audio frequency energy or amplitude on each target frequency, simulating target sliding-tone audio frequency, wherein the target sliding-tone audio frequency energy or amplitude on each target frequency and the target single-tone audio frequency energy or amplitude on each target frequency accord with the intensity change information.
It should be noted that, due to the diversity of combinations among the sliding operation force, angle, position and the single-tone identifier, the sliding change of the single-tone audio is also various, if the sliding audio corresponding to each single-tone audio is pre-generated and matched with the single-tone identifier, a great deal of cost and time are required to be consumed no matter what the matching workload is or what the sampled data volume is, and the simulation efficiency of the sliding audio is further reduced.
In this embodiment, because the target intensity change curve is used to represent the energy or amplitude change of the single-tone sample audio frequency caused by the sliding tone on different frequencies, so as to obtain the sliding tone operation triggered by the target single-tone identifier, the real-time audio frequency corresponding to the target single-tone audio frequency matched by the target single-tone identifier can be, but is not limited to, then the energy or amplitude change of the real-time audio frequency caused by the sliding tone is determined from the target intensity change curve, and then the energy or amplitude change (attenuation or amplification) is performed on the target single-tone audio frequency by using the energy or amplitude change, so that the low-cost and high-efficiency sliding tone audio simulation is realized.
According to the embodiment provided by the application, each target frequency associated with the target single-tone audio is obtained; determining intensity change information matched with each target frequency from a target intensity change curve, wherein the intensity change information is used for energy or amplitude change of target single-tone audio frequency on each target frequency due to sliding sound; the intensity change information is utilized to gain/attenuate the energy or the amplitude of the target single-tone audio frequency on each target frequency, and the target single-tone audio frequency is simulated, wherein the energy or the amplitude of the target single-tone audio frequency on each target frequency and the energy or the amplitude of the target single-tone audio frequency on each target frequency accord with the intensity change information, so that the purpose of reducing the use quantity of the single-tone sample audio frequency and the single-tone sample audio frequency in the process of the sliding-tone audio frequency simulation is achieved, and the technical effect of improving the simulation efficiency of the sliding-tone audio frequency is achieved.
As an alternative, acquiring a first intensity variation curve corresponding to the sliding sound sample audio and a second intensity variation curve corresponding to the single sound sample audio includes:
a first intensity profile representing the energy or amplitude magnitude of the sliding sample audio over N key frequencies is obtained, and a second intensity profile representing the energy or amplitude magnitude of the single sample audio over N key frequencies is obtained, wherein the N key frequencies comprise the target frequency.
Alternatively, in the present embodiment, the key frequencies may be, but are not limited to, frequencies to which human ears are more sensitive to musical tones, such as frequencies in a low frequency region (20 hz to 200 hz), frequencies in a medium frequency region (200 hz to 2,000 hz), frequencies in a high frequency region (2,000Hz to 20,000Hz), and the like.
It should be noted that, not all frequencies are obtained by harmonic intensity variation, but intensity variation of key frequency points of the sliding sound in the hearing perception range of the human ear is simulated in a limited manner, so that the calculated amount in the simulation process of the sliding sound frequency is reduced, and the simulation efficiency of the sliding sound frequency is improved.
According to the embodiment, the first intensity change curve for representing the energy or amplitude of the sliding sound sample audio frequency on the N key frequencies and the second intensity change curve for representing the energy or amplitude of the single sound sample audio frequency on the N key frequencies are obtained, wherein the N key frequencies comprise target frequencies, and therefore the purpose of reducing the calculated amount in the simulation process of the sliding sound audio frequency is achieved, and the technical effect of improving the simulation efficiency of the sliding sound audio frequency is achieved.
As an alternative, acquiring a first intensity profile representing the energy or amplitude magnitude of the sliding sample audio at N key frequencies and a second intensity profile representing the energy or amplitude magnitude of the single sample audio at N key frequencies, comprising:
s2-1, carrying out short-time Fourier transform on the sliding sound sample audio to obtain a first intensity data matrix of the total frequency of the sliding sound sample audio at each time point; determining first intensity change curves corresponding to N key frequencies from a first intensity data matrix; and, a step of, in the first embodiment,
s2-2, carrying out short-time Fourier transform on the single-tone sample audio to obtain a second intensity data matrix of the full frequency of the single-tone sample audio at each time point; and determining second intensity change curves corresponding to the N key frequencies from the second intensity data matrix.
Optionally, in this embodiment, a Short-time fourier transform (Short-Time Fourier Transform, known as STFT) is used to convert the signal between the time and frequency domains and to provide spectral information of the signal over different time periods.
Alternatively, in this embodiment, the intensity data matrix may refer, but is not limited to, a spectrum matrix obtained by performing time-frequency analysis on the audio signal, where each element represents signal energy or intensity at a specific time and frequency location. For an audio signal, the time-frequency analysis may be performed using a short-time fourier transform. The short-time fourier transform will divide the audio signal into a plurality of short-time windows and calculate the frequency spectrum within each window. By applying a fourier transform to each window, amplitude or intensity information for the different frequency components within the window can be obtained. The spectra within each window are combined to form a matrix, commonly referred to as a spectrogram or intensity map. The rows of the matrix represent time, the columns represent frequency, and each element in the matrix represents signal strength at a corresponding time and frequency location.
It should be noted that, in this embodiment, the short-time fourier transform is used to provide information of the audio signal in two dimensions of time and frequency, so as to improve the comprehensiveness of the information contained in the intensity change curve, thereby improving the simulation accuracy of the sliding sound audio.
By the embodiment provided by the application, short-time Fourier transform is carried out on the sliding sound sample audio to obtain a first intensity data matrix of the total frequency of the sliding sound sample audio at each time point; determining first intensity change curves corresponding to N key frequencies from a first intensity data matrix; performing short-time Fourier transform on the single-tone sample audio to obtain a second intensity data matrix of the full frequency of the single-tone sample audio at each time point; and determining second intensity change curves corresponding to the N key frequencies from the second intensity data matrix, so that the aim of improving the comprehensiveness of information contained in the intensity change curves is fulfilled, and the technical effect of improving the simulation accuracy of the sliding sound audio frequency is realized.
As an alternative, determining the target intensity profile based on the first intensity profile and the second intensity profile includes:
and obtaining a target intensity change curve by subtracting the values of the first intensity change curve and the second intensity change curve point by point.
In order to improve the accuracy of the target intensity change curve, a calculation mode more conforming to the curve characteristic is used to perform point-by-point subtraction on the values of the first intensity change curve and the second intensity change curve, where point-by-point can be, but is not limited to, understood as subtraction on the harmonic intensity values corresponding to frequency points (frequencies).
According to the embodiment provided by the application, the target intensity change curve is obtained by subtracting the values of the first intensity change curve and the second intensity change curve point by point, so that the purpose of using a calculation mode which is more in line with curve characteristics is achieved, and the technical effect of improving the accuracy of the target intensity change curve is achieved.
As an alternative, before obtaining the target intensity profile by subtracting the values of the first intensity profile and the second intensity profile point by point, the method further comprises:
s3-1, acquiring initial sliding sound sample audio, wherein the initial sliding sound sample audio and the initial pitch of single sound sample audio are the same;
s3-2, processing the initial sliding sound sample audio into sliding sound sample audio with the fundamental frequency variation less than or equal to a preset threshold value.
In the process of subtracting the values of the first intensity change curve and the second intensity change curve point by point to obtain the target intensity change curve, the first intensity change curve and the second intensity change curve are ensured to be the same-dimension curve, and the values of the first intensity change curve and the second intensity change curve can be further ensured to be subtracted point by point. Further, in this embodiment, the initial sliding sound sample audio is processed into sliding sound sample audio with the fundamental frequency variation less than or equal to the preset threshold value, so as to ensure that the fundamental frequencies of the first intensity variation curve and the second intensity variation curve are the same or similar.
According to the embodiment provided by the application, initial sliding sound sample audio is obtained, wherein the initial sliding sound sample audio and the initial pitch of single sound sample audio are the same; the initial sliding sound sample audio is processed into the sliding sound sample audio with the fundamental frequency variation smaller than or equal to the preset threshold value, so that the aim of ensuring that the fundamental frequencies of the first intensity variation curve and the second intensity variation curve are the same or similar is fulfilled, and the technical effect of ensuring that the values of the first intensity variation curve and the second intensity variation curve can be subtracted point by point is realized.
As an alternative, processing the slide sample audio into slide sample audio with a fundamental frequency variation less than or equal to a preset threshold value includes:
s4-1, acquiring a fundamental frequency change curve corresponding to the initial sliding sound sample audio, wherein the fundamental frequency change curve is used for representing the change condition of the fundamental frequency of the initial sliding sound sample audio along with time;
s4-2, dividing the sliding sound sample audio into a plurality of audio paragraphs according to the time axis resolution of the fundamental frequency change curve;
s4-3, reversely converting the fundamental frequency change curve into an inversely proportional sample density change curve;
s4-4, reversely resampling the plurality of audio paragraphs according to the inversely proportional sample density change curve, and supplementing new audio paragraphs through reversely resampling to obtain the sliding sound sample audio.
Alternatively, in the present embodiment, the sliding sound sample is first divided into individual small segments at the time axis resolution of the fundamental frequency variation curve. And inversely converting the fundamental frequency change curve into a change curve of the sampling point density. And finally, according to an inversely proportional sampling point density change curve, reversely resampling each audio small section, thereby realizing the function of straightening the pitch of the sliding tone. The small segment of the original sliding sound higher than the target pitch gets longer in the period of the wave, i.e. the frequency gets shorter and the pitch gets lower by inserting an intermediate value in its sampling point. On the contrary, the small section lower than the target sound in the original sliding sound realizes the pitch rising by deleting the sampling point. Further, according to the fundamental frequency variation curve, the originally slid pitch is straightened to a constant pitch dynamically and time-period by time-period.
As an alternative, before obtaining the target intensity profile by subtracting the values of the first intensity profile and the second intensity profile point by point, the method further comprises:
and scaling the second time axis of the first intensity change curve according to the first time axis of the second intensity change curve, wherein the first time axis corresponds to the scaled second time axis.
It should be noted that, before the target intensity profile is obtained by subtracting the values of the first intensity profile and the second intensity profile point by point, the second time axis of the first intensity profile needs to be scaled, because the first intensity profile is obtained by inverse resampling, and therefore the time axis of the first intensity profile is different from the second intensity profile, so that in order to control the variable, the most accurate tone change due to the sliding tone is compared, and the time axis is scaled and restored.
According to the embodiment provided by the application, the second time axis of the first intensity change curve is scaled according to the first time axis of the second intensity change curve, wherein the first time axis corresponds to the second time axis after scaling, and further the time axis scaling reduction is achieved, so that the aim of comparing the most accurate tone change generated by sliding sound is achieved, and the technical effect of improving the accuracy of the target intensity change curve is achieved.
As an alternative, before performing energy or amplitude variation on the target tone audio frequency matched with the target tone identifier by using the target intensity variation curve, the method further includes:
S5-1, respectively using a sweep frequency test signal and a white noise test signal to perform response sampling on an entity musical instrument to be simulated to obtain a sweep frequency response signal and an impulse response signal, wherein the sweep frequency test signal is a signal with continuously changed signal frequency in a certain frequency range, the white noise test signal is a random signal with uniform frequency spectrum distribution, the sweep frequency response signal is a response signal of the entity musical instrument after receiving the sweep frequency test signal, and the impulse response signal is a response signal of the entity musical instrument after receiving the white noise test signal;
s5-2, obtaining a frequency intensity difference between the sweep frequency response signal and the impact response signal, wherein the frequency intensity difference is used for representing the frequency spectrum characteristics of the resonance of the physical musical instrument;
in the process of using the target intensity change curve to change the energy or the amplitude of the target single-tone audio frequency matched with the target single-tone identification and simulating the target sliding-tone audio frequency, the method further comprises the following steps:
s6-1, dividing and acquiring the current pitch of target single-tone audio, and determining convolution operation parameters by combining the frequency intensity difference;
s6-2, under the condition that energy or amplitude change is carried out on target single-tone audio by utilizing a target intensity change curve to obtain candidate sliding-tone audio, the candidate sliding-tone audio and an impact response signal are integrated by utilizing convolution operation parameters, and the target sliding-tone audio is simulated.
Alternatively, in this embodiment, the swept test signal is a signal that continuously changes the frequency of the signal over a range of frequencies, and may be, but is not limited to, varied in a linear or logarithmic manner over a range of frequencies, and used to evaluate the response of the system or device at different frequencies. By sampling and analyzing the sweep frequency signal, a frequency response curve of the system or device under test can be obtained.
Alternatively, in the present embodiment, the white noise test signal is a random signal having a uniform spectrum distribution, and includes various frequency components, similar to white noise in nature. White noise test signals are widely used in signal processing and system testing, and because white noise has flat spectral characteristics, the white noise test signals can be used for evaluating the influence of a system or equipment on processing capacity, transmission bandwidth, noise interference and the like of different frequencies.
Alternatively, in the present embodiment, in signal processing, an original signal may be processed into a new signal with characteristics of the system described by IR, using convolution operation and IR, where IR represents the response characteristics of the system to unit pulses, and includes time domain and frequency domain information of the system. By convolving the input signal with the IR, the response of the output signal in the time domain can be obtained, so that the influence of the system on the input signals with different frequencies and time sequence characteristics can be known. The most common application is convolution reverberation, which can process an original dry sound signal into a signal with spatially reflected sound, and which can restore the acoustic properties of a target scene in a very realistic manner.
It should be noted that, to improve the reality of simulating the target sliding sound frequency, the resonance factor is considered in the simulation process of the sliding sound frequency, and the resonance factor is reflected by the frequency intensity difference between the frequency sweep response signal and the impact response signal.
According to the embodiment provided by the application, the sweep frequency test signal and the white noise test signal are respectively used for responding and sampling the entity musical instrument to be simulated, so that a sweep frequency response signal and an impulse response signal are obtained, wherein the sweep frequency test signal is a signal with continuously changed signal frequency in a certain frequency range, the white noise test signal is a random signal with uniform frequency spectrum distribution, the sweep frequency response signal is a response signal of the entity musical instrument after receiving the sweep frequency test signal, and the impulse response signal is a response signal of the entity musical instrument after receiving the white noise test signal; acquiring a frequency intensity difference between the sweep frequency response signal and the impact response signal, wherein the frequency intensity difference is used for representing the frequency spectrum characteristics of the resonance of the physical musical instrument; acquiring the current pitch of target single-tone audio, and determining convolution operation parameters by combining the frequency intensity difference; under the condition that energy or amplitude change is carried out on target single-tone audio by utilizing a target intensity change curve to obtain candidate sliding-tone audio, the candidate sliding-tone audio and an impact response signal are integrated by utilizing convolution operation parameters to simulate target sliding-tone audio, so that the aim of taking the sliding-tone audio into consideration by resonance factors is fulfilled, and the technical effect of improving the reality of the simulated target sliding-tone audio is realized.
As an alternative, in the process of using the target intensity variation curve to perform energy or amplitude variation on the target tone audio frequency matched with the target tone identifier to simulate the target sliding tone audio frequency, the method further comprises:
s7-1, acquiring operation acceleration corresponding to target operation triggered by a target single-tone identifier, wherein the target operation comprises a sliding operation;
s7-2, obtaining noise audio matched with the operation acceleration;
s7-3, under the condition that energy or amplitude change is carried out on the target single-tone audio by utilizing the target intensity change curve to obtain the alternative sliding-tone audio, the noise audio and the alternative sliding-tone audio are integrated, and the target sliding-tone audio is simulated.
It should be noted that, in order to improve the reality of the target sliding voice frequency, the noise factor is considered in the sliding voice frequency simulation process, and because the noise factor has a random and irregular characteristic, the embodiment adds the noise factor in the sliding voice frequency simulation process by acquiring the operation acceleration corresponding to the target operation triggered by the target single-tone identifier and based on the noise frequency matched with the operation acceleration.
According to the embodiment provided by the application, the operation acceleration corresponding to the target operation triggered by the target single-tone identifier is obtained, wherein the target operation comprises a sliding sound operation; acquiring noise audio matched with the operation acceleration; under the condition that energy or amplitude change is carried out on target single-tone audio by utilizing a target intensity change curve to obtain alternative sliding-tone audio, noise audio and alternative sliding-tone audio are integrated, and target sliding-tone audio is simulated, so that the aim of taking noise factors into consideration in a simulation process of the sliding-tone audio is achieved, and the technical effect of improving the reality of simulating the target sliding-tone audio is achieved.
As an alternative, for easy understanding, the above-mentioned method for simulating the sliding sound audio is applied to the simulation scene of the guqin. One of the great characteristics of the ancient musical instrument is that the ancient musical instrument has no 'product' structure. In contrast to the example of a guitar, the construction of the "product" determines that a slip playing is made on the guitar, in fact, that the transition between products is made, i.e. the slip change is not entirely continuous. However, the ancient musical instrument has no "quality" structure, and thus a smooth sound is played on the ancient musical instrument, and the change is smooth and continuous. Also, the pitch variation of the sliding tone on the Guqin is so complex that a wide variety of technical names are available to describe the sliding tone. In principle, the variation of the tones with pitch-shifting on all the palo-keys is subject to wave equation.
It should be noted that, in the technical solution of this embodiment, a large number of recordings and a strong variable-speed tonal modification are not needed, but instead, a small number of sliding sound reference samples for analysis are recorded, and feature analysis of spectral variation is performed, so that by means of analysis data, a single sound sample is used as an original material, and a sliding sound effect is generated in real time and truly according to performance.
Further by way of example, as shown in fig. 5, in the manufacturing process, data is collected by recording, clipping, signal feature analysis, etc., and the sampled data is mapped to musical instrument digital interface (Musical Instrument Digital Interface, MIDI for short) information. In the actual use process, the musical instrument digital interface inputs the calling sampling data, carries out real-time signal processing, outputs the playing audio of the Guqin, and completes the simulation of the Guqin audio.
Alternatively, in the present embodiment, as shown in fig. 6, the sound of the smooth sound of the organ may be divided into two parts, one of the musical tone simulation and the other of the noise simulation. Musical sound refers to sound with pitch that varies with performance, while noise refers to sound without pitch that is produced by friction between fingers, strings, and plates during performance.
Further in this embodiment, each instrument has its unique Harmonic sequence (Harmonic Series), which is a frequency sequence that is an integer multiple of the fundamental frequency (lowest frequency point) of musical tones, and frequency components (Inharmonic) that deviate from the Harmonic sequence, which are very complex and difficult to learn to be regular. For example, the harmonic sequence at the fundamental frequency of 100Hz is: the difference of the intensities of all frequency points in different harmonic sequences of 100Hz,200Hz,300Hz and 400Hz … … is one of important factors for enabling different musical instruments to play homophones with high tones and different sounding timbres, and nharmonic is an important factor for shaping the unique timbre of one musical instrument. Pure Harmonic Series is very difficult to obtain, and in reality, the construction of a piece of musical instrument is very complex, so that frequency components deviating from Harmonic sequences are generated, and the frequency components are one of the sources of tone color differences of similar musical instruments. For example, playing a key of a piano with a certain force, the intensity of the fundamental frequency and the harmonic frequency point that is an integer multiple of the fundamental frequency, can be subjected to nonlinear, non-equal complex attenuation over time.
However, each frequency point has a unique attenuation curve, so that the intensity proportional relationship between the frequency points is continuously changed in the attenuation process, and the tone color change generated in the attenuation process is brought. Obviously, it is already very complex to simulate such a variation, and this is limited to only one strength of one tone. It is difficult to simulate the full key and the piano pressing sounds with various forces. In addition, the piano is just a 'pure tone' generated under the pure theoretical condition of the piano, in real life, the tone of the piano is provided with additional directions such as body resonance, hammer difference and the like, and even if a formula can be fitted and induced to the law, no computer can perform real-time operation through the complex formula.
However, a breakthrough is that although the natural world is infinitely variable, human perception is very limited and even audible as in computers. The present embodiment is based on this logic, i.e. not to summarize rules, but to faithfully analyze, and to simulate the intensity variation of key frequency points of the sliding sounds in the human auditory perception range. The number of frequency points is controlled within a limited range, and the time interval of the change is not necessarily very continuous, so that real-time operation becomes possible.
Specifically, for the simulation of the sliding sound musical sound, the embodiment is split into two parts, namely, the harmonic variation simulation of vibration on strings in the sliding sound process, such as the string vibration simulation shown in fig. 6, and the simulation of resonance of a musical body in the sliding sound process, such as the box simulation shown in fig. 6.
Further in this embodiment, for string vibration simulation, a spectral change analysis of a real sliding sound is performed on a pitch-by-pitch basis, and then a digital filter is used to simulate a spectral change generated by a spectral following sliding sound displacement. Because of the inaccuracy of human ears, the frequency point is not needed to be too much when the filter works, and the change speed of the intensity of the frequency point is not needed to be too fast, namely, the simulation is controlled within the degree that the human ears do not respond but the computer can calculate.
Optionally, as shown in fig. 7, the specific implementation steps are as follows:
s7-1, recording some real sliding sounds, wherein the speed is not limited, but the sliding initial sound and the sliding termination sound need to be in a fixed interval relation, such as sliding by three degrees and pure five degrees. The recording process does not have to cover all intervals of one start of a sound because the sliding sounds above octaves are hardly generated in the performance of the old piano, and the sliding sounds are mostly performed in five scales, and the sliding sound intervals which are very close in distance, such as two degrees and three degrees, are not necessary to be recorded completely, because the difference between them is very small, and most people cannot hear. In short, the more the number of recordings in this step is, the closer to the physical reality, but the reality is that much is not needed, and the human ear can feel the reality. After recording, the embodiment obtains a batch of sliding sound samples with fixed intervals;
S7-2, analyzing the recorded fundamental frequency change curve of the sliding sound sample. The fundamental frequency change curve is the fundamental frequency change curve, and the embodiment is based on short-time Fourier transform and is obtained through analysis. After analysis, the data sheet is stored for later use.
S7-3, sampling the sliding sound with the variable pitch by utilizing the fundamental frequency change curve and the audio resampling technology analyzed in the S7-2, straightening the sliding sound into the sliding sound with the unchanged pitch, and setting the pitch as the initial pitch of the sliding sound. Thus, a batch of audio samples with unchanged pitch but spectral variation in the process of sliding sound are obtained, batch normal processing with a target value of 0dBFS is performed, and the samples are left for standby.
S7-4, recording the single-tone samples of the ancient organ, which are the same in pitch, string plucking position and similar in strength with the initial tone of the sliding tone recorded in S7-1, and carrying out normal processing in batches. They were then subjected to mean spectrum analysis using the Welch method, and the N largest, non-close-coordinate peaks were captured from the results. The N peaks represent key frequency points of the organ at the pitch, that is, characteristic information of the tone. The value of N is finally reflected as the number of frequency points of the real-time filter, so the value of N can be determined according to the hardware calculation force during actual playing, wherein the Welch method is an optimized form of a Periodog algorithm, and has the function similar to STFT, and the difference is that the Welch method averages the data of each time point, so that a frequency point-intensity average value data table is obtained, and the result does not contain time dimension information;
S7-5, carrying out short-time Fourier transform on the single-tone sample obtained in the S7-4 to obtain an intensity data matrix of the full frequency band at each time point. And then, selecting curves with the intensity changing along with time, corresponding to the N frequency point values obtained in the step S7-4, from the matrix.
S7-6, carrying out real-time spectrum analysis on the reserved sample in S7-3 by using the short-time Fourier transform of the same parameters again to obtain a full-band time-point-by-time intensity data matrix. Then, similarly, screening out intensity change curves of N frequency point values in S7-4 from the matrix, wherein the intensity change curves are time-point by time-point;
s7-7, subtracting the corresponding curves obtained in the sixth step from the curves of the N frequency points obtained in the S7-5, thereby obtaining curves which are generated due to the sliding sound and occur in the key frequency points and have intensity changes. However, the curve obtained for S7-6 needs to be scaled before subtraction, because the curve for S7-6 is based on the samples obtained in the third step, but the samples for S7-3 are over-sampled and stretched, so the time axis of all curves for S7-6 will be different from those in S7-5. Therefore, in order to control the variables, the most accurate tone change generated by the sliding sound is compared, and the time axis is scaled and restored;
S7-8, establishing a filter with N frequency points. The bandwidth of each frequency point is 1Hz, and the frequency value of each frequency point is the value of N frequency points obtained in the fourth step. The gain/attenuation values of the frequency points follow the N difference curves obtained in the step S7-7. Then, the filter is applied to a single tone sample, and the initial gain/attenuation value of all frequency points is set to 0. In the virtual performance, if only a single tone is played and the finger does not slide, the filter does not produce any effect. If the single tone is played and the sliding tone is started, the gain/attenuation value of each frequency point is gradually transited from zero to the corresponding curve value according to the interval relation.
Wherein the sample is used to change the sampling rate or sampling interval of the signal. Resampling techniques, such as increasing the sampling rate of the signal, i.e. inserting additional sampling points in the signal, may be used when it is desired to convert the signal from an original sampling rate to a new sampling rate. During the upsampling process, interpolation algorithms (e.g., linear interpolation, raised cosine interpolation, etc.) may be used to estimate the value of the new sample point inserted. Interpolation algorithms infer from known sampling points to obtain continuous signals at higher sampling rates. Thus, the frequency range of the signal can be increased or the time domain precision can be improved; or, the sampling rate of the signal is increased, i.e. additional sampling points are inserted in the signal. During the upsampling process, interpolation algorithms (e.g., linear interpolation, raised cosine interpolation, etc.) may be used to estimate the value of the new sample point inserted. Interpolation algorithms infer from known sampling points to obtain continuous signals at higher sampling rates. This may increase the frequency range of the signal or improve the time domain accuracy.
In this embodiment, for the case simulation, as shown in fig. 8, a sound box 802 is prepared, the guqin 804 is vertically arranged, the diaphragm of the sound box 802 is closely aligned with the "dragon pool" resonance hole on the back of the guqin 804, and a small-volume high-precision capacitor microphone is placed in the "chicken methane" resonance hole on the back of the guqin 804.
The specific flow is shown in fig. 9, two test signals of frequency sweep (20-20000 Hz) and white noise (122 kHz) are respectively played by using a sound box 802, and then the signals of the test signals after passing through the guqin cavity are recorded by recording, so that the frequency sweep response and the impact response of the box are obtained. And analyzing obvious frequency point intensity difference between the new frequency sweep signal and the original frequency sweep signal, and determining the frequency point intensity difference as the frequency spectrum characteristic of the resonance of the Guqin box body. When the sliding sound of the Guqin is played virtually, the current pitch of the sliding sound is obtained, and whether to start and how much proportion of convolution operation is determined according to the spectrum characteristics of the box body resonance obtained in the fourth step, wherein the convolution operation is the convolution operation between a signal obtained by string vibration simulation and a recorded impact response signal, and the box body resonance effect generated by the sliding sound is obtained.
Alternatively, in the present embodiment, since the spectrum of noise is more random and irregular than musical tones, and it is difficult to analyze the spectrum characteristics thereof in detail, the simulation of the sliding noise in the present embodiment uses a method of detecting finger acceleration during virtual performance and taking a sliding noise sample of a corresponding intensity according to the acceleration.
As shown in fig. 10, a batch of sliding noise with different intensity levels divided by speed level is recorded. And then, in the virtual playing process, detecting the acceleration of finger displacement, calling the sliding sound noise corresponding to the gear speed according to the acceleration, limiting playback logic, such as logic of how many milliseconds are not triggered repeatedly, weakening the sound head of the noise of the next sliding sound when the noise of the previous sliding sound is not played, and following the position of musical sound when the sliding sound is played, and the like, so that the noise simulation with more real sliding sound can be realized.
Alternatively, in this embodiment, a time-frequency-intensity data matrix may be derived using short-time Fourier (STFT) sliding tone pitch (fundamental frequency) curve analysis. The STFT is used for carrying out spectrum analysis on the digital audio file in a computer, the digital audio file can be divided into a plurality of parts by setting the size of a window, and then Fourier calculation is carried out respectively to obtain frequency-intensity curves at different time points. In this embodiment, the characteristic of STFT is utilized to analyze the pitch curve of the sliding sound of an seven-stringed plucked instrument with a time axis precision of 10 milliseconds.
It should be noted that the STFT obtains a huge full spectrum intensity-time matrix data, and does not have a "pitch search" function. Thus, the first few maxima at each time point need to be screened out of this huge matrix. The first few maxima, which must contain the fundamental frequency, may contain 2 nd order, 3 rd order, etc. harmonics, and may also contain random interference terms, it is a difficulty to find the fundamental frequency required for this embodiment. Thus, the present embodiment creates a workflow to ensure accuracy of the frequency point search. That is, when a slip is recorded, the starting pitch of the slip and whether it is a rising or falling slip is recorded, and these information are named on its file name. Then, when analyzing the slip curve, the file name is read, the starting pitch thereof is obtained, and the slip tendency thereof is obtained. With them, a frequency-dependent motion range of the sliding sound is created, and the program is defined to find the fundamental frequency in this range. For example, the initial tone is the center a and is the upglide, then at the first moment, the base frequency is found in the range of 10 Hz above and below 440Hz, and at the next moment, only the base frequency point at the first moment is found, and the frequency point less than 880Hz (i.e. the initial tone less than Gao Badu) is found as the base frequency. By analogy, the actually desired sliding tone pitch curve of the embodiment can be accurately found according to the frequency range and the frequency trend of the frame dead.
Alternatively, in the present embodiment, the variation curve of the multiple harmonics is calculated using a harmonic sequence following technique. And then searching the frequency change curve of the harmonic waves from the data obtained by the STFT, so that the real-time intensity change can be obtained.
Alternatively, in the present embodiment, two kinds of test signals are used: the 20-20000Hz sweep frequency signal and the 122kHz sampling rate white noise signal excite and pick up the acoustic characteristics of the tank body resonance. Wherein white noise is eventually processed into IR signals after the entire flow has been passed. The recording and processing method is that the white noise is continuously played at a high sound pressure level by using sound equipment, and then the playing is stopped suddenly, so that the residual sound tail picked up backwards from the stopping moment is generated by Gu Qinqin body resonance. That is, resonance occurs due to white noise with a large sound pressure level, and then the white noise is gradually consumed due to the disappearance of the white noise, so that the process of consumption is that the organ box body faces the unit impact generated after full-band white noise impact, namely the IR required by the embodiment. The sweep frequency signal determines under what conditions and to what extent the IR participates in the processing and output of signals in the flow.
It should be noted that when the present embodiment virtually plays a single sound, it is not desirable for IR to participate therein, because the recording of the single sound itself is accompanied by resonance of the box. The case resonance is required in this embodiment because the single sound slides through the specific resonance frequency point of the case. The sweep frequency is used for detecting the sweep frequency signal with fixed amplitude, and the amplitude of the sweep frequency signal can change after the sweep frequency signal passes through the guqin box body. This change is indicative of which frequency points a slipsound will cause the cabinet to resonate, and the degree of resonance. This data is then obtained and then IR is applied to produce a sliding single sound band that will resonate at a particular frequency with a particular intensity.
Optionally, in this embodiment, the signal is filtered first by a real-time sliding noise filtering technique. The filter calls for an off-the-shelf Peak type filter with a frequency bandwidth of 1Hz. After filtering, the signal is subjected to real-time resetting according to the virtual playing condition and Buffer Size, namely the hearing is subjected to lossless change of the pitch, and the frequency spectrum is dynamically changed. Then, the information control signal for performance is convolved with the IR signal. Finally, the convolved signal is combined with the noise signal.
By the embodiment provided by the application, endless recording sampling is not needed, so that the manufacturing time is greatly shortened, and meanwhile, the storage space is greatly reduced; meanwhile, when virtual playing is carried out, the speed and the interval of the sliding sound are not limited any more; the user can play the game, so that the learning cost of the user is saved; the output audio frequency does not relate to variable speed and tone processing which damages the tone quality of the original sample.
It will be appreciated that in the specific embodiments of the present application, related data such as user information is referred to, and when the above embodiments of the present application are applied to specific products or technologies, user permissions or consents need to be obtained, and the collection, use and processing of related data need to comply with related laws and regulations and standards of related countries and regions.
It should be noted that, for simplicity of description, the foregoing method embodiments are all expressed as a series of action combinations, but it should be understood by those skilled in the art that the present application is not limited by the order of actions described, as some steps may be performed in other order or simultaneously in accordance with the present application. Further, those skilled in the art will also appreciate that the embodiments described in the specification are all preferred embodiments, and that the acts and modules referred to are not necessarily required in the present application.
According to another aspect of the embodiments of the present application, there is also provided a device for simulating a sliding sound audio for implementing the method for simulating a sliding sound audio. As shown in fig. 11, the apparatus includes:
a first obtaining unit 1102, configured to obtain a sliding-tone sample audio and a single-tone sample audio, where starting pitches of the sliding-tone sample audio and the single-tone sample audio are the same;
a second obtaining unit 1104, configured to obtain a first intensity variation curve corresponding to the sliding-tone sample audio and a second intensity variation curve corresponding to the single-tone sample audio, where the first intensity variation curve is used to represent energy or amplitude of the sliding-tone sample audio at different frequencies, and the second intensity variation curve is used to represent energy or amplitude of the single-tone sample audio at different frequencies;
A determining unit 1106 configured to determine a target intensity profile based on the first intensity profile and the second intensity profile, where the target intensity profile is used to represent energy or amplitude changes of the single-tone sample audio generated by the sliding sound at different frequencies;
and a simulation unit 1108, configured to simulate the target sliding sound audio by using the target intensity variation curve to perform energy or amplitude variation on the target sliding sound audio matched with the target voice identifier in response to the sliding sound operation triggered on the target voice identifier.
For specific embodiments, reference may be made to the example shown in the above-mentioned sliding audio simulation apparatus, and this example will not be described herein.
As an alternative, the simulation unit 1108 includes:
the first acquisition module is used for acquiring each target frequency associated with the target single-tone audio;
the first determining module is used for determining intensity change information matched with each target frequency from a target intensity change curve, wherein the intensity change information is used for energy or amplitude change of the target single-tone audio frequency, which is generated by sliding sound, on each target frequency;
the first simulation module is used for simulating the target sliding sound audio by utilizing the intensity change information and gain/attenuation of the energy or the amplitude of the target single sound audio on each target frequency, wherein the energy or the amplitude of the target sliding sound audio on each target frequency and the energy or the amplitude of the target single sound audio on each target frequency accord with the intensity change information.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the second acquisition unit 1104 includes:
a second acquisition module for acquiring a first intensity profile representing the energy or amplitude magnitude of the sliding sample audio at N key frequencies, and a second intensity profile representing the energy or amplitude magnitude of the single sample audio at N key frequencies, wherein the N key frequencies include the target frequency.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the second obtaining module includes:
the first transformation submodule is used for carrying out short-time Fourier transformation on the sliding sound sample audio to obtain a first intensity data matrix of the total frequency of the sliding sound sample audio at each time point; determining first intensity change curves corresponding to N key frequencies from a first intensity data matrix; and, a step of, in the first embodiment,
the second transformation submodule is used for carrying out short-time Fourier transformation on the single-tone sample audio to obtain a second intensity data matrix of the full frequency of the single-tone sample audio at each time point; and determining second intensity change curves corresponding to the N key frequencies from the second intensity data matrix.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the determining unit 1106 includes:
and the subtraction module is used for obtaining a target intensity change curve by subtracting the values of the first intensity change curve and the second intensity change curve point by point.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the apparatus further includes:
the third acquisition module is used for acquiring initial sliding sound sample audio before the target intensity change curve is obtained by subtracting the values of the first intensity change curve and the second intensity change curve point by point, wherein the initial sliding sound sample audio and the single sound sample audio have the same initial pitch;
the processing module is used for processing the initial sliding sound sample audio into the sliding sound sample audio with the fundamental frequency variation less than or equal to a preset threshold before the target intensity variation curve is obtained by subtracting the values of the first intensity variation curve and the second intensity variation curve point by point.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the processing module includes:
the acquisition sub-module is used for acquiring a fundamental frequency change curve corresponding to the initial sliding sound sample audio, wherein the fundamental frequency change curve is used for representing the change condition of the fundamental frequency of the initial sliding sound sample audio along with time;
the dividing sub-module is used for dividing the sliding sound sample audio into a plurality of audio paragraphs according to the time axis resolution of the fundamental frequency change curve;
the conversion sub-module is used for reversely converting the fundamental frequency change curve into an inversely proportional sample density change curve;
and the sampling sub-module is used for carrying out inverse resampling on the plurality of audio paragraphs according to the inversely proportional sample density change curve, and supplementing new audio paragraphs through the inverse resampling to obtain the sliding sound sample audio.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the apparatus further includes:
and the scaling sub-module is used for scaling the second time axis of the first intensity change curve according to the first time axis of the second intensity change curve before the target intensity change curve is obtained by subtracting the values of the first intensity change curve and the second intensity change curve point by point, wherein the first time axis corresponds to the second time axis after the scaling.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the apparatus further includes:
the sampling unit is used for carrying out energy or amplitude change on target single-tone audio frequency matched with the target single-tone identification by utilizing a target intensity change curve, and before simulating target sliding-tone audio frequency, respectively using a sweep frequency test signal and a white noise test signal to respond and sample an entity musical instrument to be simulated to obtain a sweep frequency response signal and an impact response signal, wherein the sweep frequency test signal is a signal with continuously changed signal frequency in a certain frequency range, the white noise test signal is a random signal with uniform frequency spectrum distribution, the sweep frequency response signal is a response signal after the entity musical instrument receives the sweep frequency test signal, and the impact response signal is a response signal after the entity musical instrument receives the white noise test signal;
the difference unit is used for obtaining the frequency intensity difference between the sweep frequency response signal and the impact response signal before the target single tone frequency matched with the target single tone identification is subjected to energy or amplitude change by utilizing the target intensity change curve and the target sliding tone frequency is simulated, wherein the frequency intensity difference is used for representing the frequency spectrum characteristics of the resonance of the entity musical instrument;
The apparatus further comprises:
the combination unit is used for carrying out energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, acquiring the current pitch of the target single-tone audio frequency in the process of simulating the target sliding-tone audio frequency, and determining convolution operation parameters by combining the frequency intensity difference;
the first integration unit is used for performing energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, and simulating the target sliding-tone audio frequency in the process of simulating the target sliding-tone audio frequency.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
As an alternative, the apparatus further includes:
the third acquisition unit is used for acquiring the operation acceleration corresponding to the target operation triggered by the target tone mark in the process of simulating the target tone frequency by utilizing the energy or amplitude change of the target tone frequency matched with the target tone mark by utilizing the target intensity change curve, wherein the target operation comprises a tone slipping operation;
The fourth acquisition unit is used for acquiring noise audio matched with the operation acceleration in the process of simulating the target sliding sound audio by utilizing the energy or amplitude change of the target single sound audio matched with the target single sound identification by utilizing the target intensity change curve;
the second integration unit is used for performing energy or amplitude change on the target single-tone audio frequency matched with the target single-tone identification by utilizing the target intensity change curve, and performing integration processing on the noise audio frequency and the alternative single-tone audio frequency under the condition that the alternative single-tone audio frequency is obtained by performing energy or amplitude change on the target single-tone audio frequency by utilizing the target intensity change curve in the process of simulating the target single-tone audio frequency, so as to simulate the target single-tone audio frequency.
Specific embodiments may refer to examples shown in the above-mentioned method for simulating the sliding sound audio, and this example will not be described herein.
According to yet another aspect of the embodiments of the present application, there is also provided an electronic device for implementing the above-mentioned method of simulating a sliding sound audio, which may be, but is not limited to, the user device 102 or the server 112 shown in fig. 1, the embodiment being illustrated by the electronic device as the user device 102, and further as shown in fig. 12, the electronic device includes a memory 1202 and a processor 1204, the memory 1202 storing a computer program, the processor 1204 being configured to execute the steps of any of the above-mentioned method embodiments by means of the computer program.
Alternatively, in this embodiment, the electronic device may be located in at least one network device of a plurality of network devices of the computer network.
Alternatively, in the present embodiment, the above-described processor may be configured to execute the following steps by a computer program:
s1, acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio;
s2, acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies;
s3, determining a target intensity change curve based on the first intensity change curve and the second intensity change curve, wherein the target intensity change curve is used for representing energy or amplitude change of the single-tone sample audio frequency, which is generated by sliding sounds at different frequencies;
and S4, responding to the sliding sound operation triggered by the target single sound identification, and utilizing a target strength change curve to change the energy or the amplitude of the target single sound audio matched with the target single sound identification so as to simulate the target sliding sound audio.
Alternatively, it will be appreciated by those of ordinary skill in the art that the configuration shown in fig. 12 is merely illustrative, and that fig. 12 is not intended to limit the configuration of the electronic device described above. For example, the electronic device may also include more or fewer components (e.g., network interfaces, etc.) than shown in FIG. 12, or have a different configuration than shown in FIG. 12.
The memory 1202 may be configured to store software programs and modules, such as program instructions/modules corresponding to the method and apparatus for simulating a sliding sound audio in the embodiments of the present application, and the processor 1204 executes the software programs and modules stored in the memory 1202 to perform various functional applications and data processing, that is, to implement the method for simulating a sliding sound audio. Memory 1202 may include high-speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid-state memory. In some examples, the memory 1202 may further include memory located remotely from the processor 1204, which may be connected to the electronic device via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof. The memory 1202 may be used for storing information such as, but not limited to, a sliding sound sample audio, a single sound sample audio, and a target sliding sound audio. As an example, as shown in fig. 12, the memory 1202 may include, but is not limited to, a first acquiring unit 1102, a second acquiring unit 1104, a determining unit 1106, and an simulating unit 1108 in the simulating device including the sliding sound audio. In addition, other module units in the above-mentioned sliding audio simulation device may be also included, but are not limited thereto, and are not described in detail in this example.
Optionally, the transmission device 1206 is configured to receive or transmit data via a network. Specific examples of the network described above may include wired networks and wireless networks. In one example, the transmission means 1206 comprises a network adapter (Network Interface Controller, NIC) that can be connected to other network devices and routers via a network cable to communicate with the internet or a local area network. In one example, the transmission device 1206 is a Radio Frequency (RF) module for communicating wirelessly with the internet.
In addition, the electronic device further includes: a display 1208 for displaying the information such as the sliding sound sample audio, the single sound sample audio, and the target sliding sound audio; and a connection bus 1210 for connecting the respective module parts in the above-described electronic apparatus.
In other embodiments, the user device or the server may be a node in a distributed system, where the distributed system may be a blockchain system, and the blockchain system may be a distributed system formed by connecting the plurality of nodes through a network communication. The nodes may form a peer-to-peer network, and any type of computing device, such as a server, a user device, etc., may become a node in the blockchain system by joining the peer-to-peer network.
According to one aspect of the present application, a computer program product is provided, comprising a computer program/instructions containing program code for performing the method shown in the flow chart. In such embodiments, the computer program may be downloaded and installed from a network via a communication portion, and/or installed from a removable medium. When executed by a central processing unit, performs the various functions provided by the embodiments of the present application.
The foregoing embodiment numbers of the present application are merely for describing, and do not represent advantages or disadvantages of the embodiments.
It should be noted that the computer system of the electronic device is only an example, and should not impose any limitation on the functions and the application scope of the embodiments of the present application.
The computer system includes a central processing unit (Central Processing Unit, CPU) which can execute various appropriate actions and processes according to a program stored in a Read-Only Memory (ROM) or a program loaded from a storage section into a random access Memory (Random Access Memory, RAM). In the random access memory, various programs and data required for the system operation are also stored. The CPU, the ROM and the RAM are connected to each other by bus. An Input/Output interface (i.e., I/O interface) is also connected to the bus.
The following components are connected to the input/output interface: an input section including a keyboard, a mouse, etc.; an output section including a Cathode Ray Tube (CRT), a liquid crystal display (Liquid Crystal Display, LCD), and the like, and a speaker, and the like; a storage section including a hard disk or the like; and a communication section including a network interface card such as a local area network card, a modem, and the like. The communication section performs communication processing via a network such as the internet. The drive is also connected to the input/output interface as needed. Removable media such as magnetic disks, optical disks, magneto-optical disks, semiconductor memories, and the like are mounted on the drive as needed so that a computer program read therefrom is mounted into the storage section as needed.
In particular, according to embodiments of the present application, the processes described in the various method flowcharts may be implemented as computer software programs. For example, embodiments of the present application include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method shown in the flowcharts. In such embodiments, the computer program may be downloaded and installed from a network via a communication portion, and/or installed from a removable medium. The computer program, when executed by a central processing unit, performs the various functions defined in the system of the present application.
According to one aspect of the present application, there is provided a computer-readable storage medium, from which a processor of a computer device reads the computer instructions, the processor executing the computer instructions, causing the computer device to perform the methods provided in the various alternative implementations described above.
Alternatively, in the present embodiment, the above-described computer-readable storage medium may be configured to store a computer program for executing the steps of:
s1, acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio;
s2, acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies;
s3, determining a target intensity change curve based on the first intensity change curve and the second intensity change curve, wherein the target intensity change curve is used for representing energy or amplitude change of the single-tone sample audio frequency, which is generated by sliding sounds at different frequencies;
And S4, responding to the sliding sound operation triggered by the target single sound identification, and utilizing a target strength change curve to change the energy or the amplitude of the target single sound audio matched with the target single sound identification so as to simulate the target sliding sound audio.
Alternatively, in this embodiment, it will be understood by those skilled in the art that all or part of the steps in the methods of the above embodiments may be performed by a program for instructing electronic equipment related hardware, and the program may be stored in a computer readable storage medium, where the storage medium may include: flash disk, read-Only Memory (ROM), random-access Memory (Random Access Memory, RAM), magnetic or optical disk, and the like.
The foregoing embodiment numbers of the present application are merely for describing, and do not represent advantages or disadvantages of the embodiments.
The integrated units in the above embodiments may be stored in the above-described computer-readable storage medium if implemented in the form of software functional units and sold or used as separate products. Based on such understanding, the technical solution of the present application may be embodied in essence or a part contributing to the prior art or all or part of the technical solution in the form of a software product stored in a storage medium, including several instructions to cause one or more computer devices (which may be personal computers, servers or network devices, etc.) to perform all or part of the steps of the methods of the various embodiments of the present application.
In the foregoing embodiments of the present application, the descriptions of the embodiments are emphasized, and for a portion of this disclosure that is not described in detail in this embodiment, reference is made to the related descriptions of other embodiments.
In several embodiments provided in the present application, it should be understood that the disclosed user equipment may be implemented in other manners. The above-described embodiments of the apparatus are merely exemplary, and the division of the units, such as the division of the units, is merely a logical function division, and may be implemented in another manner, for example, multiple units or components may be combined or may be integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be through some interfaces, units or modules, or may be in electrical or other forms.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The foregoing is merely a preferred embodiment of the present application and it should be noted that modifications and adaptations to those skilled in the art may be made without departing from the principles of the present application and are intended to be comprehended within the scope of the present application.

Claims (14)

1. A method for simulating a sliding sound audio, comprising:
acquiring a sliding sound sample audio and a single sound sample audio, wherein the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio;
acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio on different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio on different frequencies;
Determining a target intensity profile based on the first intensity profile and the second intensity profile, wherein the target intensity profile is used for representing energy or amplitude changes of the single-tone sample audio caused by sliding sounds at different frequencies;
and responding to the sliding sound operation triggered by the target single sound identification, and utilizing the target strength change curve to change the energy or the amplitude of the target single sound audio matched with the target single sound identification so as to simulate the target sliding sound audio.
2. The method of claim 1, wherein using the target intensity profile to perform energy or amplitude variation on the target tone audio that matches the target tone identification, simulating target sliding tone audio, comprises:
acquiring each target frequency associated with the target single-tone audio;
determining intensity change information matched with each target frequency from the target intensity change curve, wherein the intensity change information is used for energy or amplitude change of the target single-tone audio frequency, which is generated by sliding sound, on each target frequency;
and using the intensity change information, and simulating the target sliding sound audio by gain/attenuation of the energy or the amplitude of the target single sound audio on each target frequency, wherein the energy or the amplitude of the target sliding sound audio on each target frequency and the energy or the amplitude of the target single sound audio on each target frequency accord with the intensity change information.
3. The method of claim 2, wherein the obtaining a first intensity profile corresponding to the sliding sample audio and a second intensity profile corresponding to the single sample audio comprises:
the method further includes obtaining the first intensity profile representing energy or amplitude magnitudes of the sliding sample audio at N key frequencies, and the second intensity profile representing energy or amplitude magnitudes of the single sample audio at the N key frequencies, wherein the N key frequencies include the target frequency.
4. A method according to claim 3, wherein said obtaining said first intensity profile representing energy or amplitude magnitudes of said slide sample audio at N key frequencies and said second intensity profile representing energy or amplitude magnitudes of said single sample audio at said N key frequencies comprises:
performing short-time Fourier transform on the sliding sound sample audio to obtain a first intensity data matrix of the total frequency of the sliding sound sample audio at each time point; determining the first intensity change curves corresponding to the N key frequencies from the first intensity data matrix; performing short-time Fourier transform on the single-tone sample audio to obtain a second intensity data matrix of the full frequency of the single-tone sample audio at each time point; and determining the second intensity change curves corresponding to the N key frequencies from the second intensity data matrix.
5. The method of claim 1, wherein the determining a target intensity profile based on the first intensity profile and the second intensity profile comprises:
and obtaining the target intensity change curve by subtracting the values of the first intensity change curve and the second intensity change curve point by point.
6. The method of claim 5, wherein prior to said deriving said target intensity profile by subtracting the values of said first intensity profile and said second intensity profile point by point, said method further comprises:
acquiring initial sliding sound sample audio, wherein the initial sliding sound sample audio and the single sound sample audio have the same initial pitch;
and processing the initial sliding sound sample audio into the sliding sound sample audio with the fundamental frequency variation less than or equal to a preset threshold value.
7. The method of claim 6, wherein processing the slide sample audio as slide sample audio with a fundamental frequency variation less than or equal to a preset threshold comprises:
acquiring a fundamental frequency change curve corresponding to the initial sliding sound sample audio, wherein the fundamental frequency change curve is used for representing the change condition of the fundamental frequency of the initial sliding sound sample audio along with time;
Dividing the sliding sound sample audio into a plurality of audio paragraphs according to the time axis resolution of the fundamental frequency change curve;
inversely converting the fundamental frequency change curve into an inversely proportional sample density change curve;
and reversely resampling the plurality of audio paragraphs according to the inversely proportional sample density change curve, and supplementing new audio paragraphs through the reversely resampling to obtain the sliding sound sample audio.
8. The method of claim 7, wherein prior to said deriving said target intensity profile by subtracting the values of said first intensity profile and said second intensity profile point by point, said method further comprises:
and scaling the second time axis of the first intensity change curve according to the first time axis of the second intensity change curve, wherein the first time axis corresponds to the scaled second time axis.
9. The method according to any one of claims 1 to 8, wherein,
before the target tone frequency matched with the target tone identification is subjected to energy or amplitude change by utilizing the target intensity change curve, and the target sliding tone frequency is simulated, the method further comprises the following steps:
The method comprises the steps of responding and sampling an entity musical instrument to be simulated by using a frequency sweep test signal and a white noise test signal respectively to obtain a frequency sweep response signal and an impulse response signal, wherein the frequency sweep test signal is a signal with continuously changed signal frequency in a certain frequency range, the white noise test signal is a random signal with uniform frequency spectrum distribution, the frequency sweep response signal is a response signal after the entity musical instrument receives the frequency sweep test signal, and the impulse response signal is a response signal after the entity musical instrument receives the white noise test signal;
acquiring a frequency intensity difference between the sweep frequency response signal and the impact response signal, wherein the frequency intensity difference is used for representing the frequency spectrum characteristic of the resonance of the physical musical instrument;
in the process of using the target intensity change curve to change the energy or the amplitude of the target single-tone audio frequency matched with the target single-tone identification and simulating the target sliding-tone audio frequency, the method further comprises the following steps:
acquiring the current pitch of the target single-tone audio frequency, and determining convolution operation parameters by combining the frequency intensity differences;
and under the condition that energy or amplitude of the target single-tone audio is changed by utilizing the target intensity change curve to obtain candidate sliding-tone audio, integrating the candidate sliding-tone audio and the impact response signal by utilizing the convolution operation parameters to simulate the target sliding-tone audio.
10. The method according to any one of claims 1 to 8, wherein in said using the target intensity variation curve to perform energy or amplitude variation on the target tone audio matched to the target tone identification, the method further comprises:
acquiring operation acceleration corresponding to target operation triggered by the target single-tone identifier, wherein the target operation comprises the sliding operation;
acquiring noise audio matched with the operation acceleration;
and under the condition that energy or amplitude of the target single-tone audio is changed by utilizing the target intensity change curve to obtain alternative sliding-tone audio, integrating the noise audio and the alternative sliding-tone audio, and simulating the target sliding-tone audio.
11. A device for simulating a sliding sound audio, comprising:
the device comprises a first acquisition unit, a second acquisition unit and a third acquisition unit, wherein the first acquisition unit is used for acquiring a sliding sound sample audio and a single sound sample audio, and the initial pitch of the sliding sound sample audio is the same as that of the single sound sample audio;
the second acquisition unit is used for acquiring a first intensity change curve corresponding to the sliding sound sample audio and a second intensity change curve corresponding to the single sound sample audio, wherein the first intensity change curve is used for representing the energy or amplitude of the sliding sound sample audio at different frequencies, and the second intensity change curve is used for representing the energy or amplitude of the single sound sample audio at different frequencies;
A determining unit configured to determine a target intensity profile based on the first intensity profile and the second intensity profile, wherein the target intensity profile is used to represent energy or amplitude changes of the single-tone sample audio generated by a sliding sound at different frequencies;
and the simulation unit is used for responding to the sliding sound operation triggered by the target single sound identification, and utilizing the target strength change curve to perform energy or amplitude change on the target single sound audio matched with the target single sound identification so as to simulate the target sliding sound audio.
12. A computer-readable storage medium, characterized in that the computer-readable storage medium comprises a stored program, wherein the program, when run by an electronic device, performs the method of any one of claims 1 to 10.
13. A computer program product comprising computer programs/instructions which, when executed by a processor, implement the steps of the method as claimed in any one of claims 1 to 10.
14. An electronic device comprising a memory and a processor, characterized in that the memory has stored therein a computer program, the processor being arranged to execute the method according to any of the claims 1 to 10 by means of the computer program.
CN202310922550.8A 2023-07-25 2023-07-25 Sliding sound audio simulation method and device, storage medium and electronic equipment Pending CN117275527A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310922550.8A CN117275527A (en) 2023-07-25 2023-07-25 Sliding sound audio simulation method and device, storage medium and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310922550.8A CN117275527A (en) 2023-07-25 2023-07-25 Sliding sound audio simulation method and device, storage medium and electronic equipment

Publications (1)

Publication Number Publication Date
CN117275527A true CN117275527A (en) 2023-12-22

Family

ID=89218497

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310922550.8A Pending CN117275527A (en) 2023-07-25 2023-07-25 Sliding sound audio simulation method and device, storage medium and electronic equipment

Country Status (1)

Country Link
CN (1) CN117275527A (en)

Similar Documents

Publication Publication Date Title
CN108417228B (en) Human voice tone similarity measurement method under musical instrument tone migration
US8842847B2 (en) System for simulating sound engineering effects
US8716586B2 (en) Process and device for synthesis of an audio signal according to the playing of an instrumentalist that is carried out on a vibrating body
WO2022012195A1 (en) Audio signal processing method and related apparatus
US9626949B2 (en) System of modeling characteristics of a musical instrument
US20170024495A1 (en) Method of modeling characteristics of a musical instrument
Penttinen et al. Model-based sound synthesis of the guqin
CN106997765A (en) The quantitatively characterizing method of voice tone color
Gabrielli et al. A digital waveguide-based approach for Clavinet modeling and synthesis
Christensen Introduction to audio processing
Jensen The timbre model
JP7359164B2 (en) Sound signal synthesis method and neural network training method
Fritz et al. Perceptual studies of violin body damping and vibrato
Zhang Application of audio visual tuning detection software in piano tuning teaching
CN117275527A (en) Sliding sound audio simulation method and device, storage medium and electronic equipment
CN116959503A (en) Sliding sound audio simulation method and device, storage medium and electronic equipment
Dubnov Polyspectral analysis of musical timbre
CN112289289A (en) Editable universal tone synthesis analysis system and method
Bensa et al. Perceptive and cognitive evaluation of a piano synthesis model
Knees et al. Basic methods of audio signal processing
Masuyama et al. Modal decomposition of musical instrument sounds via optimization-based non-linear filtering
CN113140204B (en) Digital music synthesis method and equipment for pulsar signal control
Penttinen et al. Morphing instrument body models
Cuzzucoli et al. The Sound
Permana et al. Improved Feature Extraction for Sound Recognition Using Combined Constant-Q Transform (CQT) and Mel Spectrogram for CNN Input

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination