CN105513583B - song rhythm display method and system - Google Patents

song rhythm display method and system Download PDF

Info

Publication number
CN105513583B
CN105513583B CN201510827808.1A CN201510827808A CN105513583B CN 105513583 B CN105513583 B CN 105513583B CN 201510827808 A CN201510827808 A CN 201510827808A CN 105513583 B CN105513583 B CN 105513583B
Authority
CN
China
Prior art keywords
beat
audio data
value
acquiring
time interval
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510827808.1A
Other languages
Chinese (zh)
Other versions
CN105513583A (en
Inventor
段岢
林鎏娟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian Star Net eVideo Information Systems Co Ltd
Original Assignee
Fujian Star Net eVideo Information Systems Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian Star Net eVideo Information Systems Co Ltd filed Critical Fujian Star Net eVideo Information Systems Co Ltd
Priority to CN201510827808.1A priority Critical patent/CN105513583B/en
Publication of CN105513583A publication Critical patent/CN105513583A/en
Application granted granted Critical
Publication of CN105513583B publication Critical patent/CN105513583B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/40Rhythm
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/071Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for rhythm pattern analysis or rhythm style recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/076Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of timing, tempo; Beat detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/021Indicator, i.e. non-screen output user interfacing, e.g. visual or tactile instrument status or guidance information using lights, LEDs or seven segments displays
    • G10H2220/081Beat indicator, e.g. marks or flashing LEDs to indicate tempo or beat positions
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2220/00Input/output interfacing specifically adapted for electrophonic musical tools or instruments
    • G10H2220/021Indicator, i.e. non-screen output user interfacing, e.g. visual or tactile instrument status or guidance information using lights, LEDs or seven segments displays
    • G10H2220/086Beats per minute [BPM] indicator, i.e. displaying a tempo value, e.g. in words or as numerical value in beats per minute

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Auxiliary Devices For Music (AREA)

Abstract

the invention provides a method and a system for displaying song rhythm, wherein the method comprises the following steps: acquiring audio data of the song according to a preset acquisition period; acquiring the frequency amplitude of the audio data; acquiring a beat value of audio data; and displaying the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value. According to the method, the beat value of the audio data is obtained, the audio data is displayed together with the frequency amplitude, and the display speed of the audio amplitude is determined according to the beat value; the rhythm, the beat and the frequency of the music can be vividly and visually displayed, and the rhythm of the music can be perfectly reflected by the three components in a complementary manner; meanwhile, better visual enjoyment can be brought to the viewers, and the music atmosphere is activated.

Description

Song rhythm display method and system
Technical Field
the invention relates to the field of audio signal processing, in particular to a method and a system for displaying song rhythm.
background
The music player in the prior art not only supports the playing of music, but also displays the height of the music in the form of dynamic music frequency spectrum during the playing process of the music. In the music frequency spectrum in the prior art, after audio data (PCM data) of music is acquired in real time and the audio data is subjected to fourier transform to transform an audio signal from a time domain to a frequency domain, each discrete frequency corresponds to an amplitude. The music frequency spectrum determines the height of the frequency spectrum animation display by using the amplitude; however, such music spectrum only shows the height and intensity of music (the higher the treble, the right peak is and the higher the bass, the left peak is), and the characteristics of music are many, such as rhythm, beat and timbre of music, which cannot be shown in the general music spectrum. Especially, the rhythm and the beat of the music supplement each other to form a framework of the music together to support the rhythm of the music, so that the rhythm and the beat of the music are very important for the music; however, the conventional music frequency spectrum cannot show the defects of the conventional music frequency spectrum and the conventional music frequency spectrum, so that a dynamic frequency spectrum which more accords with the music characteristic cannot be presented to a user, and the requirement of the user for grasping the music rhythm according to the frequency spectrum in the music appreciation process cannot be met.
The application numbers are: 200910162906.2 discloses a singing and playing computer real-time analysis system, which is used to draw a time-frequency-intensity real-time spectrogram according to short-time fast-reading Fourier transform during the processing of music signals, wherein the spectrogram can reflect the stability of sound, the jitter range, the resonance peak position, the intensity, the adhesion condition of the slip tone in the middle of the fundamental tone, the strong and weak contrast between the fundamental tone and each overtone and the time-dependent progress of frequency, so as to represent the rhythm change of music.
the above patent does not express the change of music rhythm according to the rhythm and beat characteristics of music, or can not express the music rhythm sense in the most intuitive and accurate way; therefore, it is necessary to provide a method and system for acquiring music spectrum based on the rhythm and beat of music.
disclosure of Invention
The technical problem to be solved by the invention is as follows: a method and a system for displaying song rhythm are provided, and audio data can be displayed according to a beat value and a frequency amplitude.
In order to solve the technical problems, the invention adopts the technical scheme that:
a method of displaying song tempos, comprising:
Acquiring audio data of a song according to a preset acquisition period;
Acquiring a beat value of the audio data;
acquiring a beat value of the audio data according to the acquisition period and the time interval;
and displaying the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value.
The invention provides another technical scheme as follows:
a song tempo display system comprising:
the preset module is used for acquiring audio data of the song according to a preset acquisition cycle;
the first acquisition module is used for acquiring the frequency amplitude of the audio data;
The second acquisition module is used for acquiring the beat value of the audio data;
and the display module is used for displaying the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value.
the invention has the beneficial effects that: the invention provides a display method and a system of song rhythm, which are characterized in that the beat value of audio data is obtained, the audio data is displayed by combining with frequency amplitude, and the display speed of the audio amplitude is determined by the beat value; the rhythm, the beat and the frequency of the music can be vividly and visually displayed, and the rhythm of the music can be perfectly reflected by the three components in a complementary manner; meanwhile, better visual enjoyment can be brought to the viewers, and the music atmosphere is activated.
Drawings
FIG. 1 is a flow chart of a song rhythm display method according to the present invention;
fig. 2 is a flowchart illustrating a method for displaying song tempos according to an embodiment of the present invention;
FIG. 3 is a block diagram of the structural components of a song tempo display system of the present invention;
FIG. 4 is a music-frequency interface diagram of an electrocardiogram in accordance with an embodiment of the present invention;
FIG. 5 is a diagram of a phonogram music frequency spectrum interface moving left and right according to an embodiment of the present invention;
FIG. 6 is a music spectrum interface diagram of a 3D animated fret diagram according to an embodiment of the present invention;
FIG. 7 is a spectrogram according to an embodiment of the present invention.
description of reference numerals:
1. presetting a module; 2. a first acquisition module; 3. a second acquisition module; 4. and a display module.
Detailed Description
in order to explain technical contents, achieved objects, and effects of the present invention in detail, the following description is made with reference to the accompanying drawings in combination with the embodiments.
The most key concept of the invention is as follows: the method comprises the steps of obtaining a beat value of audio, displaying audio data according to the beat value and a frequency amplitude, and determining the display speed of the audio amplitude according to the beat value in the displaying process.
the technical terms related to the invention are explained as follows:
referring to fig. 1 and 2, the present invention provides a method for displaying song tempos, including:
acquiring audio data of a song according to a preset acquisition period;
acquiring the frequency amplitude of the audio data;
acquiring more than two beat points in the audio data;
acquiring the time interval between every two adjacent beat points;
acquiring a beat value of the audio data according to the acquisition period and the time interval;
and displaying the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value.
further, the beat value is linearly proportional to the display speed of the audio amplitude.
further, the linear proportional relationship is a direct proportional relationship.
From the above description, if the tempo is faster, the change speed of the amplitude is faster, and vice versa, the display speed of the beat value and the audio amplitude is in a linear and proportional relationship, which conforms to the viewing habit of people and can more intuitively reflect the fluctuation of the tempo of music.
Further, the audio data includes a music score corresponding to the audio, and a tempo value of the audio data is obtained according to the music score.
as can be seen from the above description, each beat point is marked on the music score, and the beat value of the audio data can be obtained directly according to the time interval between adjacent beat points. The music score shown in fig. 7 has four beats, 4 beats per cycle, one beat of quarter note, and the playing time of each bar is t. For example, in the first section, there are four notes, each note is a beat, and the time interval (beat value) between adjacent notes is t/4; and 11 notes exist in the second bar, the first three notes form a beat, the time interval (the beat value) between the first note and the second note of the second bar is t/8, and the rest is done, so that the time intervals (the beat values) of other notes are calculated.
referring to fig. 4, fig. 5 and fig. 6, further, the "displaying the audio data according to the beat value and the frequency amplitude" specifically includes:
displaying the audio data by using an electrocardiogram, wherein the potential change of the electrocardiogram represents the change of the frequency amplitude value, and the speed of the potential change represents the beat value;
or displaying the audio data by using a left-right moving sound column graph, wherein the length of the sound column represents the frequency amplitude, and the moving speed of the sound column represents the beat value;
or displaying the audio data by using a sound column graph with elevation change, wherein the elevation height of the sound column represents the audio amplitude, and the elevation speed of the sound column represents the beat value;
Or displaying the audio data by using the flicker change of the light, wherein the brightness of the light represents the audio amplitude, and the flicker speed of the light represents the beat value.
the above description shows that the invention can show the rhythm change of the audio data in different expression forms, and the invention has various forms, rich and colorful contents and vivid images, and brings a brand new audio-visual experience to people through the combination of the appreciation of frequency spectrum and the auditory enjoyment of audio.
further, the step of obtaining more than two beat points in the audio data; acquiring the time interval between every two adjacent beat points; the step value of the audio data obtained according to the acquisition cycle and the time interval "is specifically:
performing wavelet transform processing on the audio data to obtain a coefficient value of the audio data;
carrying out peak value detection on the coefficient value of the audio data to obtain more than two beat points;
acquiring the time interval of two adjacent beat points in the original audio signal;
Acquiring a time interval with the maximum occurrence probability;
and acquiring a beat value according to the acquisition period and the time interval with the maximum occurrence probability.
As can be seen from the above description, based on the fact that the sampling rate is fixed, for example, 48K audio data is collected per second, the time interval for each piece of audio data to be transmitted to the application layer for processing is almost the same, and therefore, the beat value of each piece of audio data can be approximately recognized as the same due to the short time interval, but even within the same time interval, the number of notes appearing is different, which is the manifestation of the rhythm speed. By performing processing such as wavelet transform and peak detection processing on audio data, the tempo value of the audio data, that is, the rhythm of music can be extracted.
The sampling rate of the original signal is reduced by one time after each wavelet transform is performed on the signal in the audio data, and the interval between the beat values in the invention refers to the beat interval in the original audio signal, so that the transformed beat interval needs to be transformed to the beat interval value in the original signal, and a beat interval histogram is constructed according to the beat interval value, so as to ensure the accuracy of the obtained beat interval.
further, before the step of performing wavelet transform processing on the audio data to obtain detail coefficient values of the audio data, the method further includes:
and normalizing the quantization bit number of the audio data.
According to the description, after the audio data are obtained, normalization preprocessing is firstly carried out, the data length of each sampling point represented by quantization digits can be converted into standard quantization to be represented, so that the influence of certain audio data on the data calculation result is prevented from being too large, the calculation of the audio data can be simplified, and the calculation efficiency is improved.
Further, after the "obtaining the beat value", the method further includes:
constructing a beat interval histogram according to the time interval between every two adjacent beat values;
smoothing the beat interval histogram;
And acquiring the time interval with the highest occurrence probability in the smoothed beat interval histogram to obtain a beat value mean value.
As can be seen from the above, the time interval with the highest frequency of occurrence is extracted from the smoothed beat interval histogram, and the accuracy of the extracted beat value can be improved.
further, the step value average value obtaining by obtaining the time interval with the highest occurrence probability in the smoothed beat interval histogram specifically includes:
Acquiring a time interval with the highest occurrence probability in the beat interval histogram after smoothing processing;
According to the formula: and obtaining a beat value mean value, wherein the beat value is the preset time interval with the highest acquisition cycle/probability.
Further, after the "obtaining the mean value of the beat values", the method further includes:
Acquiring an initial beat point of the audio data;
and acquiring all the beat points of the audio data according to the initial beat point and the beat value mean value.
As can be seen from the above, since the positions of the beat points are determined by the initial beat points and the beat values, the initial beat points are the average of the first possible beat point positions obtained after wavelet transformation, that is, the initial beat value average, we can obtain the positions of all the beat points according to the initial beat value average plus several times of beat values, and assuming that the current beat is Bv, the initial beat value average is P1, the position of the nth beat point is: pn ═ P1+ (n-1) Bv.
referring to fig. 3, another technical solution provided by the present invention is:
A song tempo display system comprising:
The system comprises a preset module 1, a processing module and a control module, wherein the preset module is used for acquiring audio data of a song according to a preset acquisition cycle;
the first obtaining module 2 is configured to obtain a frequency amplitude of the audio data;
the second obtaining module 3 is configured to obtain a beat value of the audio data;
And the display module 4 is configured to display the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value.
from the above description, the beneficial effects of the present invention are: the invention provides a display system of song rhythm, obtain the beat value of the audio data through the second acquisition module 3, and then combine the frequency amplitude that the first acquisition module 2 obtains together, demonstrate the audio data through the display module 4, confirm the display speed of the said audio amplitude from the said beat value; the rhythm, the beat and the frequency of the music can be vividly and visually displayed, and the rhythm of the music can be perfectly reflected by the three components in a complementary manner; meanwhile, better visual enjoyment can be brought to the viewers, and the music atmosphere is activated.
further, the second obtaining module 3 includes:
the first acquisition unit is used for acquiring more than two beat points in the audio data;
a second acquisition unit configured to acquire a time interval between every two adjacent beat points;
and the third acquisition unit is used for acquiring the beat value of the audio data according to the acquisition cycle and the time interval.
further, the first obtaining unit is specifically configured to perform wavelet transform processing on the audio data to obtain a coefficient value of the audio data; performing peak value detection on the coefficient value of the audio data to acquire more than two beat points;
the second obtaining unit is specifically configured to obtain a time interval between two adjacent beat points in the original audio signal; acquiring a time interval with the maximum occurrence probability;
and the third acquisition unit is specifically used for acquiring the beat value according to the acquisition cycle and the time interval with the maximum occurrence probability.
according to the description, the beat value is obtained according to the time interval between the adjacent beat points and the acquisition cycle of the audio data, and a basis is provided for showing the rhythm and the music height of the audio data by combining the beat value and the frequency amplitude.
Referring to fig. 1, fig. 2, fig. 4, fig. 5 and fig. 6, a first embodiment of the present invention is:
there is provided an audio frequency spectrum capable of exhibiting a music tempo, comprising:
S1: acquiring audio data of a song according to a preset acquisition period;
In the playing process of music, recording the music, setting the acquisition period (sampling rate) to be 48K, acquiring PCM audio data, and sending the PCM audio data to an application layer for processing. Because the acquisition period is fixed, the time from acquisition to processing is also fixed, that is, the number of beats acquired is fixed, in the present invention, we extract the number of notes (the number of beats) included in the fixed number of beats, that is, the beat value, which represents the rhythm of music.
S2: acquiring the frequency amplitude of the audio data;
Specifically, it is preferable that the first main board is responsible for performing Fourier transform (Fourier transform) on the audio data to obtain the frequency amplitude of the audio data. The fourier transform is capable of transforming the audio data from the time domain to the frequency domain (the so-called frequency domain, which may be represented by the amplitude corresponding to the frequency on the X-axis and the frequency on the Y-axis); then, segmenting the frequency of the X axis, and dividing the width of each frequency band in a nonlinear way; according to each discrete frequency, a frequency amplitude is corresponding, and in the music frequency spectrum adopting the sound column to display the frequency amplitude, the height of the animation sound column (the height of the sound column rising) is determined by the frequency amplitude, and the sound column falls back to the initial point position after finishing displaying the height. Therefore, the maximum frequency amplitude is obtained in each divided frequency band, and all the frequency amplitudes of the audio data are sent to the second main board; and the height fluctuation of the animation sound column is determined by the second main board according to the frequency amplitude.
S3: extracting music rhythm;
s31: carrying out normalization pretreatment on the acquired audio data;
because PCM audio data is subjected to Fourier transform, when analog audio data is converted into digital audio data, audio signals are discretized in amplitude, the data length of each sampling point is represented by quantization bit numbers (8 bits and 16 bits are commonly used), the quantization bit numbers are different, the data ranges which can be represented by the sampling points are different, for example, the data range which can be represented by the 8-bit sampling point is 0-255, and the data range which can be represented by the 16-bit sampling point is-32768-32767; therefore, for unified processing, after the audio data is acquired, normalization processing needs to be performed on the audio data, and data of sampling points are normalized to [ -1,1 ];
Assuming that the sample point sequence is S [ ], and the normalized sample point sequence is Ns [ ], the formula can be used:
Ns[]=(S[]-S min)/(S max-S min)*(D max-D min)+D min;
wherein Smax is the maximum value in the sequence of sample points; smin is the minimum value in the sampling point sequence, Dmax is the maximum value expected in the sequence after the sampling points are normalized, and Dmin is the minimum value expected in the sequence after the sampling points are normalized; and the data of the sampling points are normalized to prevent certain audio data from influencing the data calculation result too much, simplify the calculation of the audio data and improve the calculation efficiency.
S32: performing wavelet transformation on the preprocessed audio data;
performing discrete wavelet transform on the preprocessed audio data signals, and generating approximate coefficients and detail coefficients after the audio data signals are subjected to wavelet transform for one time; the approximation coefficients preserve the low frequency component of the signal, while the detail coefficients preserve the high frequency component of the signal; and (3) continuously carrying out wavelet decomposition on the approximation coefficient of the low-frequency component to obtain a new approximation coefficient and a detail coefficient, and continuously repeating the decomposition process for n times of wavelet decomposition to obtain n groups of coefficient values. Here, preferably n is 4, resulting in 4 sets of coefficient values.
after the audio data signals are subjected to wavelet transformation, time domain information cannot be lost, and the tone quality effect of the audio data signals can be ensured; meanwhile, the sampling rate of the audio data signal is reduced by one time after each wavelet decomposition, so that the data volume of all processing can be reduced, and the calculation efficiency of the algorithm is improved.
S33: carrying out peak value detection on the audio data after wavelet transformation to obtain a beat point;
since the tempo of music is likely to occur at the peak point of the wavelet decomposition coefficients, it is necessary to apply a peak detection algorithm to the coefficient values in the n sets of coefficient values obtained in S32, respectively, to obtain tempo points;
the method specifically comprises the following steps:
(1) full-wave correction: absolute value conversion is carried out on all data in the audio data signal;
(2) windowing and extreme value solving: for short-time analysis, a window must be added; a window with smooth transition characteristics at both ends should be used in theory, with the purpose of reducing discontinuity of the audio signal at the beginning and end of the audio data frame;
And windowing all corrected audio data, sliding a window while solving a local extremum in the window, continuously solving the extremum, determining a final peak value point by considering that the local extremum is in a window range and the local extremum has the same times, if the communication times are more than 90% of the window width, considering the extremum as a peak value, and simultaneously recording the position of the peak value so as to obtain all peak values of the audio data, wherein the peak value position is assumed to be a beat point position.
S34: acquiring the time interval of the beat points, and constructing a histogram;
after the peak values of n groups of wavelet coefficients of the audio data are detected, the time interval between every two adjacent peak value positions needs to be calculated, namely the time interval IOI between every two adjacent beat points is calculated;
Specifically, after each wavelet transform is performed on the audio data signal, the sampling rate of the original audio data signal is reduced by one time, and the required beat point time interval IOI refers to an interval value in the original audio data signal. Therefore, it is necessary to first transform the transformed time interval values into the time interval values in the original audio data signal.
Supposing that the current interval value is a value obtained by performing nth wavelet transform on the original audio data signal and is recorded as IOI 2; when the time interval value of the original audio data signal is recorded as IOI1, IOI1 is IOI2 × 2 n;
then, according to the probability of the interval occurrence of the beat points, a beat interval histogram is constructed;
S35: smoothing the beat interval histogram, and calculating to obtain a beat value;
the accuracy of the beat value directly obtained according to the constructed beat interval histogram is not high, and in order to improve the accuracy of the beat value, the beat interval histogram needs to be smoothed;
specifically, the weight of the time interval of a beat point is introduced first, and the weight of the time interval of the beat point is estimated by judging the similarity degree of other beat point intervals adjacent to a certain beat point interval. Here, two parameters need to be determined: firstly, comparing with adjacent time intervals, and preferably selecting four beat time intervals around a certain time interval according to historical experience; second, determining whether a time interval is close can define a deviation value, and if the difference between two time intervals is smaller than the deviation value, they can be regarded as the same time interval.
at this time, the following situations may also occur in the histogram: first, the histogram may have more than one extreme (time interval with the same probability of occurrence), how to choose; second, the time interval with the highest probability of occurrence occurs in the low density region; we can simply select the maximum value of this extremum to calculate the beat value, but this maximum value does not represent the main distribution of beat interval values. Therefore, if we do not consider the distribution of the histogram and simply select the maximum value, the calculated beat value will be inaccurate.
to solve the above problem, we can convolve the histogram with a gaussian function to smooth the histogram; the gaussian function may highlight areas of high density; after the smoothing processing, the time interval of the beat point can be judged by obtaining the maximum amplitude;
the time interval of the beat points is obtained as follows:
Obtaining the time interval with the maximum occurrence probability in the smooth histogram, and according to a formula:
The beat value Bv is the time interval with the maximum acquisition period/occurrence probability; and obtaining the mean value of the beat values.
s36: acquiring all beat points in audio data;
the position of the beat point is determined by the initial beat point and the beat value in the audio data, and the initial beat point is the average value of the position of the first possible beat point obtained after wavelet transformation; after the beat value mean value is calculated each time, all beat point positions can be obtained according to the initial beat point and a plurality of times of beat value mean values. The method specifically comprises the following steps:
if the current beat is Bv and the initial beat point is P1, the position of the nth beat point is:
Pn=P1+(n-1)Bv。
s4: displaying the audio data according to the beat value and the frequency amplitude;
And determining the display speed of the audio amplitude according to the beat value, wherein the optimal beat value is in linear and direct proportional relation with the display speed of the audio amplitude.
the audio data may be presented using a wide variety of spectrograms,
as shown in fig. 4, if an electrocardiograph is used to display the audio data, the change of the electrical potential of the electrocardiogram represents the change of the amplitude of the frequency, and the speed of the change of the electrical potential represents the beat value;
as shown in fig. 5, if the audio data is displayed by using a left-right moving fret spectrogram, the length of the fret represents the frequency amplitude, and the moving speed of the fret represents the beat value;
As shown in fig. 6, if the audio data is displayed by using a spectrogram of a sound column with a variable elevation, the elevation height of the sound column represents the audio amplitude, and the elevation speed of the sound column represents the beat value;
The audio data may also be presented using a flash change in light, wherein the intensity of the light represents the audio amplitude and the speed at which the light flashes represents the beat value.
Specifically, we explain in detail a 3D fret spectrogram with elevation change:
In the spectrogram, the rising height of a sound column is determined according to the amplitude of audio, and the frequency spectrum is changed from small to large from left to right; the stronger the music intensity is, the higher the sound column rises, otherwise, the lower the sound column rises; the stronger the music high pitch, the more the right side columns are, and the stronger the music low pitch, the more the left side columns are;
The rising and falling speeds of the sound column are determined according to the beat value, and the higher the rhythm is, the higher the rising and falling speeds are; the slower the tempo, the slower the speed of rise and fall.
in the process of frequency spectrum display, according to the frequency amplitude of the obtained audio data, the sound column with the height of 0 rises to the height corresponding to the frequency amplitude through the upward throwing motion, and then returns to the height of 0 through the downward throwing motion; calculating to obtain a value of V0 according to a motion formula of S ═ V0 × (t + (1/2) × a × (t) × (S is a rising path determined by the maximum frequency amplitude, V0 is an initial velocity, a is an acceleration preset according to an empirical value, t is time, and time t of motion is determined according to a beat value of the obtained audio data), wherein the initial velocity V0 determines the velocity of "upward throwing" and "downward throwing" of the sound column, and specifically, the faster the rhythm is, the shorter the time of "upward throwing motion" and "downward throwing motion" is; conversely, the longer the movement time.
various 3D animations which are edited by self-definition can be preferably added into the music spectrum, and the music atmosphere can be played by the dancing accompanying of characters while the pitch of the sound columns changes along with the music rhythm.
referring to fig. 3, the second embodiment of the present invention is:
a digital audio-visual system comprises a set-top box and a display screen, wherein the set-top box is connected with the display screen; the set top box comprises a first main board and a second main board; the first mainboard comprises a preset module 1 and a first acquisition module 2;
the second motherboard includes a second obtaining module 3, and the second obtaining module 3 may specifically include a first obtaining unit, a second obtaining unit, and a third obtaining unit;
the preset module 1 is used for acquiring audio data of a song according to a preset acquisition cycle and a preset frequency spectrum display mode;
the first obtaining module 2 is configured to obtain a frequency amplitude of the audio data, and send the frequency amplitude to a second motherboard;
The second obtaining module 3 is configured to obtain a beat value of the audio data, and specifically, perform wavelet transform processing on the audio data through the first obtaining unit to obtain a coefficient value of the audio data; carrying out peak value detection on the detail coefficient value of the audio data to obtain more than two beat points;
acquiring the time interval of two adjacent beat points in the original audio signal through the second acquisition unit; acquiring a time interval with the maximum occurrence probability;
and acquiring a beat value according to the acquisition cycle and the time interval with the maximum occurrence probability through a third acquisition unit.
The display screen is used for displaying the audio data by using a spectrogram of a 3D animation according to the beat value and the frequency amplitude in a preset display mode, wherein the display speed of the audio amplitude is determined according to the beat value.
in conclusion, the song rhythm display and the system thereof provided by the invention can extract the rhythm of the audio data, display the audio data by combining the frequency amplitude, realize vividly and visually displaying the rhythm, the rhythm and the frequency of the music and perfectly embody the rhythm of the music; moreover, the rhythm and the frequency show a linear direct proportional relation, so that the fluctuation of the rhythm of the music is more intuitively reflected; furthermore, various presentation forms for showing the rhythm change of the audio data are provided, the presentation forms are vivid, more choices are provided for people, and brand new audio-visual experience is brought.
The above description is only an embodiment of the present invention, and not intended to limit the scope of the present invention, and all equivalent changes made by using the contents of the present specification and the drawings, or applied directly or indirectly to the related technical fields, are included in the scope of the present invention.

Claims (12)

1. a method for displaying song tempos, comprising:
acquiring audio data of a song according to a preset acquisition period;
acquiring the frequency amplitude of the audio data;
acquiring a beat value of the audio data;
Displaying the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value;
The step of acquiring the beat value of the audio data is specifically as follows:
Acquiring more than two beat points in the audio data;
acquiring the time interval between every two adjacent beat points;
and acquiring the beat value of the audio data according to the acquisition period and the time interval.
2. the method as claimed in claim 1, wherein the audio data includes audio music score, and the tempo value of the audio data is obtained according to the music score.
3. the method of claim 1, wherein the tempo value is linearly proportional to the speed of displaying the audio amplitude.
4. A method of displaying the tempo of a song according to claim 3, wherein said linear scaling relationship is a direct scaling relationship.
5. The method for displaying song tempo according to claim 1, wherein said "displaying said audio data according to said tempo value and frequency amplitude" specifically comprises:
displaying the audio data by using an electrocardiogram, wherein the potential change of the electrocardiogram represents the change of the frequency amplitude value, and the speed of the potential change represents the beat value;
Or using a left-right moving sound column graph to display the audio data, wherein the length of the sound column represents the frequency amplitude, and the moving speed of the sound column represents the beat value;
or displaying the audio data by using a sound column graph with elevation change, wherein the elevation height of the sound column represents the audio amplitude, and the elevation speed of the sound column represents the beat value;
or displaying the audio data by using the flicker change of the light, wherein the brightness of the light represents the audio amplitude, and the flicker speed of the light represents the beat value.
6. The method for displaying song tempos according to claim 1, wherein said step of obtaining more than two beat points in said audio data; acquiring the time interval between every two adjacent beat points; the step value of the audio data obtained according to the acquisition cycle and the time interval "is specifically:
Performing wavelet transform processing on the audio data to obtain a coefficient value of the audio data;
carrying out peak value detection on the coefficient value of the audio data to obtain more than two beat points;
Acquiring the time interval of two adjacent beat points in the original audio signal;
acquiring a time interval with the maximum occurrence probability;
and acquiring a beat value according to the acquisition period and the time interval with the maximum occurrence probability.
7. the method as claimed in claim 6, wherein said step of performing wavelet transform processing on said audio data to obtain coefficient values of the audio data further comprises:
And normalizing the quantization bit number of the audio data.
8. The method for displaying song tempo of claim 6, wherein after said obtaining the tempo value, further comprising:
constructing a beat interval histogram according to the time interval between every two adjacent beat values;
smoothing the beat interval histogram;
and acquiring the time interval with the highest occurrence probability in the smoothed beat interval histogram to obtain a beat value mean value.
9. The method for displaying song tempo according to claim 8, wherein the step of obtaining the time interval with the highest occurrence probability in the smoothed beat interval histogram to obtain the average value of the beat values specifically comprises:
acquiring a time interval with the highest occurrence probability in the beat interval histogram after smoothing processing;
according to the formula: and obtaining a beat value mean value, wherein the beat value is the preset time interval with the highest acquisition cycle/probability.
10. the method for displaying song rhythm of claim 8, wherein after obtaining the mean value of tempo values, further comprising:
acquiring an initial beat point of the audio data;
And acquiring all the beat points of the audio data according to the initial beat point and the beat value mean value.
11. a song tempo display system comprising:
the preset module is used for acquiring audio data of the song according to a preset acquisition cycle;
the first acquisition module is used for acquiring the frequency amplitude of the audio data;
the second acquisition module is used for acquiring the beat value of the audio data;
the display module is used for displaying the audio data according to the beat value and the frequency amplitude, wherein the display speed of the audio amplitude is determined according to the beat value;
The second acquisition module includes:
the first acquisition unit is used for acquiring more than two beat points in the audio data;
a second acquisition unit configured to acquire a time interval between every two adjacent beat points;
and the third acquisition unit is used for acquiring the beat value of the audio data according to the acquisition cycle and the time interval.
12. the display system of song tempos of claim 11,
the first obtaining unit is specifically configured to perform wavelet transform processing on the audio data to obtain a coefficient value of the audio data; carrying out peak value detection on the detail coefficient value of the audio data to obtain more than two beat points;
The second obtaining unit is specifically configured to obtain a time interval between two adjacent beat points in the original audio signal; acquiring a time interval with the maximum occurrence probability;
And the third acquisition unit is specifically used for acquiring the beat value according to the acquisition cycle and the time interval with the maximum occurrence probability.
CN201510827808.1A 2015-11-25 2015-11-25 song rhythm display method and system Active CN105513583B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510827808.1A CN105513583B (en) 2015-11-25 2015-11-25 song rhythm display method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510827808.1A CN105513583B (en) 2015-11-25 2015-11-25 song rhythm display method and system

Publications (2)

Publication Number Publication Date
CN105513583A CN105513583A (en) 2016-04-20
CN105513583B true CN105513583B (en) 2019-12-17

Family

ID=55721512

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510827808.1A Active CN105513583B (en) 2015-11-25 2015-11-25 song rhythm display method and system

Country Status (1)

Country Link
CN (1) CN105513583B (en)

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106373595A (en) * 2016-08-31 2017-02-01 天脉聚源(北京)科技有限公司 Method and device for displaying audio information
CN106340307A (en) * 2016-08-31 2017-01-18 天脉聚源(北京)科技有限公司 Method and device used for displaying audio information
CN106356045A (en) * 2016-08-31 2017-01-25 天脉聚源(北京)科技有限公司 Audio information display method and device
GB201620839D0 (en) * 2016-12-07 2017-01-18 Weav Music Ltd Data format
CN106652981B (en) * 2016-12-28 2019-09-13 广州酷狗计算机科技有限公司 BPM detection method and device
JP6930547B2 (en) * 2017-01-27 2021-09-01 ソニーグループ株式会社 Information processing equipment, information processing methods and their programs
CN107452361B (en) * 2017-08-08 2020-07-07 腾讯音乐娱乐(深圳)有限公司 Song sentence dividing method and device
CN107545883A (en) * 2017-10-13 2018-01-05 广州酷狗计算机科技有限公司 The method and apparatus for determining the rhythm speed grade of music
CN108109611A (en) * 2017-12-11 2018-06-01 大连高马艺术设计工程有限公司 A kind of learning aids system of visual control fundamental tone and overtone and synthetic video
CN108335688B (en) * 2017-12-28 2021-07-06 广州市百果园信息技术有限公司 Method for detecting main beat point in music, computer storage medium and terminal
CN108108457B (en) * 2017-12-28 2020-11-03 广州市百果园信息技术有限公司 Method, storage medium, and terminal for extracting large tempo information from music tempo points
CN108319657B (en) * 2018-01-04 2022-02-01 广州市百果园信息技术有限公司 Method for detecting strong rhythm point, storage medium and terminal
CN108320730B (en) * 2018-01-09 2020-09-29 广州市百果园信息技术有限公司 Music classification method, beat point detection method, storage device and computer device
CN108305604B (en) * 2018-01-30 2020-11-24 浙江省公众信息产业有限公司 Music visualization method and device and computer readable storage medium
CN108335703B (en) * 2018-03-28 2020-10-09 腾讯音乐娱乐科技(深圳)有限公司 Method and apparatus for determining accent position of audio data
CN108495208A (en) * 2018-03-29 2018-09-04 云谷(固安)科技有限公司 Flexible screen speaker and its control method
CA3037908A1 (en) * 2019-03-25 2020-09-25 Kemal S. Ahmed Beat tracking visualization through textual medium
CN110085253A (en) * 2019-05-09 2019-08-02 广州小鹏汽车科技有限公司 A kind of control method, device, vehicle and the storage medium of Music Visualization information
CN110390943B (en) * 2019-06-28 2022-07-08 上海元笛软件有限公司 Audio synthesis method and device, computer equipment and storage medium
CN110717053A (en) * 2019-10-17 2020-01-21 广州酷狗计算机科技有限公司 Picture display method, device, terminal and storage medium based on songs
CN111128100B (en) 2019-12-20 2021-04-20 网易(杭州)网络有限公司 Rhythm point detection method and device and electronic equipment
CN112973092B (en) * 2021-02-02 2022-03-25 随机漫步(上海)体育科技有限公司 Training assisting method and device, storage medium, electronic equipment and bicycle
CN111773620A (en) * 2020-07-01 2020-10-16 随机漫步(上海)体育科技有限公司 Method and device for assisting bicycle training and method and device for training network model
CN111859015A (en) * 2020-07-01 2020-10-30 腾讯音乐娱乐科技(深圳)有限公司 Music response method and related equipment
CN113823325B (en) * 2021-06-03 2024-08-16 腾讯科技(北京)有限公司 Audio rhythm detection method, device, equipment and medium
CN114356470A (en) * 2021-11-26 2022-04-15 惠州华阳通用电子有限公司 Dynamic display method and system of audio playing control
CN114630140A (en) * 2022-03-17 2022-06-14 阿里巴巴(中国)有限公司 Information setting method and device based on audio data

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3743757A (en) * 1971-10-18 1973-07-03 Nippon Musical Instruments Mfg Automatic rhythm instrument with temporary roll beat circuit
JPH08190394A (en) * 1995-01-09 1996-07-23 Taito Corp Method for displaying graphics interlocked to music content on karaoke sing-along machine and device therefor

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030205124A1 (en) * 2002-05-01 2003-11-06 Foote Jonathan T. Method and system for retrieving and sequencing music by rhythmic similarity
JP2006106818A (en) * 2004-09-30 2006-04-20 Toshiba Corp Music retrieval device, music retrieval method and music retrieval program
CN2833663Y (en) * 2005-10-20 2006-11-01 深圳市蔚科电子科技开发有限公司 MP3 player with beating function
US7645929B2 (en) * 2006-09-11 2010-01-12 Hewlett-Packard Development Company, L.P. Computational music-tempo estimation
CN101645268B (en) * 2009-08-19 2012-03-14 李宋 Computer real-time analysis system for singing and playing
CN103578478B (en) * 2013-11-11 2016-08-17 科大讯飞股份有限公司 Obtain the method and system of musical tempo information in real time

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3743757A (en) * 1971-10-18 1973-07-03 Nippon Musical Instruments Mfg Automatic rhythm instrument with temporary roll beat circuit
JPH08190394A (en) * 1995-01-09 1996-07-23 Taito Corp Method for displaying graphics interlocked to music content on karaoke sing-along machine and device therefor

Also Published As

Publication number Publication date
CN105513583A (en) 2016-04-20

Similar Documents

Publication Publication Date Title
CN105513583B (en) song rhythm display method and system
US8907195B1 (en) Method and apparatus for musical training
US9552741B2 (en) Systems and methods for quantifying a sound into dynamic pitch-based graphs
US8853516B2 (en) Audio analysis apparatus
KR20150016225A (en) Automatic conversion of speech into song, rap or other audible expression having target meter or rhythm
JP6759545B2 (en) Evaluation device and program
JP6627482B2 (en) Technique determination device
WO2017090720A1 (en) Technique determining device and recording medium
Müller et al. Interactive fundamental frequency estimation with applications to ethnomusicological research
CN113674723B (en) Audio processing method, computer equipment and readable storage medium
KR101221673B1 (en) Apparatus for practicing electric guitar performance
CN110751935A (en) Method for determining musical instrument playing point and scoring rhythm
US10546566B2 (en) Systems and methods for detecting musical features in audio content
CN104978329A (en) Method for generating musical melody in line drawing form
WO2022172576A1 (en) Sound generation method using machine learning model, method of training machine learning model, sound generation device, training device, sound generation program, and training program
JP5012263B2 (en) Performance clock generating device, data reproducing device, performance clock generating method, data reproducing method and program
CN115619897A (en) Image processing method, image processing device, electronic equipment and storage medium
Chudy et al. On the relation between gesture, tone production and perception in classical cello performance
EP3786942A1 (en) Musical instrument tuning
Gu Recognition algorithm of piano playing music in intelligent background
JP2009014923A (en) Musical performance clock generating device, data reproducing device, musical performance clock generating method, data reproducing method, and program
US10395669B2 (en) Voice analysis apparatus, voice analysis method, and program
WO2017164216A1 (en) Acoustic processing method and acoustic processing device
CN113488007B (en) Information processing method, information processing device, electronic equipment and storage medium
WO2016039465A1 (en) Acoustic analysis device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant