CN106991163A - A kind of song recommendations method based on singer's sound speciality - Google Patents

A kind of song recommendations method based on singer's sound speciality Download PDF

Info

Publication number
CN106991163A
CN106991163A CN201710206783.2A CN201710206783A CN106991163A CN 106991163 A CN106991163 A CN 106991163A CN 201710206783 A CN201710206783 A CN 201710206783A CN 106991163 A CN106991163 A CN 106991163A
Authority
CN
China
Prior art keywords
song
singer
tone color
range
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710206783.2A
Other languages
Chinese (zh)
Inventor
余春艳
苏金池
刘煌
郭文忠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fuzhou University
Original Assignee
Fuzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fuzhou University filed Critical Fuzhou University
Priority to CN201710206783.2A priority Critical patent/CN106991163A/en
Publication of CN106991163A publication Critical patent/CN106991163A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/635Filtering based on additional data, e.g. user or group profiles
    • G06F16/637Administration of user profiles, e.g. generation, initialization, adaptation or distribution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Health & Medical Sciences (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Auxiliary Devices For Music (AREA)

Abstract

The present invention relates to a kind of song recommendations method based on singer's sound speciality, this method using song numbered musical notation and singer sing opera arias band etc. information, song features library is set up, the performance range of song is extracted and constructs voice tone color embedded space, and obtains the voice tone color of original singer and is characterized.Performance range is extracted to the recording file of singing opera arias of singer and voice tone color is characterized, the sound speciality of singer is portrayed;The sound level distribution situation of calculating song, in the performance capability evaluation value of each sound level, thus calculates user and sings range and the matching degree of song range requirement with singer;The sound clip of singer is embedded into tone color embedded space, the tone color similarity with each singer in embedded space is calculated respectively.The present invention can consider the range matching degree and tone color similarity of singer, calculate the recommendation degree for the user per song.

Description

A kind of song recommendations method based on singer's sound speciality
Technical field
It is more particularly to a kind of to be based on singer's sound speciality the present invention relates to the acoustic signal processing method in field of singing Song recommendations method.
Background technology
Music commending system, which is focused on to user, recommends it to may like the song listened, and the recommended technology of use can mainly divide For content-based recommendation and the recommendation based on collaborative filtering.The main audio according to music itself of content-based recommendation algorithm Feature is recommended, the feature such as low-level image feature or melody, rhythm, school, emotion such as including MFCC.Pushing away based on collaborative filtering Algorithm is recommended mainly to be pushed away according to the program request behavior between user or based on playing record, the similitude between user Recommend.
In recent years, the dual thorn for the large-scale true man's music select-elite class program of fast-developing and each shelves applied in mobile Internet Under swashing, music commending system likes this traditional application scenarios of the song listened gradually to migrate from for user's recommendation, and then penetrates into Enter and recommend to like the emerging application scenarios such as the song sung for user.
However, the synchronous migration of method is not recommended in the migration of application scenarios along with music.Exemplified by singing APP, The song that recommendation function in APP is recommended is based on current hit song.But, hit song is not appropriate for all users and drilled Sing.It is possible to that song tone is too high, causes treble portion because user itself sings range scope and sings the limitation of ability Can do not sing up;Song is also possible to be adapted to be gone to deduce with sound that is rough, having explosive force, and user is that a sound is sweet Schoolgirl.
Obviously, new recommendation application scenarios need new recommendation pattern.Under the application scenarios that K is sung, user is not only Listen song, it is often more important that can farthest deduce song.This is the process of a bi-directional matching, on the one hand, needs are examined The speciality of user's own voice is considered, such as the actual tone color for singing range and sound of user;On the other hand, it is necessary to consider song The bent requirement to singing ability, the range scope of such as song claim and what kind of tone color are more suitable for the emotion of the deduction song Deng.
In order to preferably introduce the concept of the song recommendations based on singer's sound speciality, some relevant musicals, people are introduced The theoretical basic conception of sound.
Tone color:Tone color refers to sound in certain attribute acoustically produced, and auditor can judge two accordingly with same Mode is presented, the difference of sound with identical pitch and loudness.
Range:Range has two kinds of the range of total range and other voice or musical instrument.Total range refers to total model of the series of sound Enclose, i.e., the scope from double bass to descant.The range of individual other voice or musical instrument refers to some voice or certain musical instrument whole Scope of the double bass that can be reached in range to descant.The range of musical instrument is relatively fixed, and the range of voice is due to each The inborn vocal cords size of people, length, thickness are different and whether there is the reasons such as vocal music training by system the day after tomorrow, there is larger area Not.
MIDI (Musical Instrument Digital Interface), is the communication of a kind of digital music, musical instrument Standard.MIDI files can flexibly record the information such as the pitch and the duration of a sound of song, be easy to computer carry out pitch analysis with Processing.
CQT is composed, and a kind of tone color frequency physical features, the wave filter group being distributed by centre frequency exponentially believes musical sound Number it is expressed as determining the spectrum energy of music single-tone, the quality factor q of wave filter group keeps constant.
Individual performance ability includes width and the accuracy in pitch control ability in each sound level of singer's range scope etc..Hair Acoustic energy power is the basis of performance ability, medically utilizes the personal sounding range of range of voice dossier and the dynamic model of loudness Enclose.Professional singer then lifts the performance ability of itself by the vocal music training method of system, but common singer typically will not Go to use specific training method.
Therefore, the application be based on above-mentioned analysis, using song numbered musical notation and singer sing opera arias band etc. information, set up song features text Part storehouse, extract song sings range and the tamber characteristic of singer.Recording file of singing opera arias when being given song recitals using user simultaneously With the numbered musical notation information of song, on the premise of sound level completes quality height, the performance range and tamber characteristic of user is extracted.Synthesis is examined Consider the matching degree and user's tone color and singer's tone color in Qu Ku sung in range and Qu Ku between the requirement of song range of user Between similarity, calculate the recommendation degree per song for the user in Qu Ku, and to the high song of user's recommendation recommendation degree.
The content of the invention
In view of this, it is an object of the invention to provide a kind of song recommendations method based on singer's sound speciality, to drilling Range similarity and tone color similarity, song recommendations are analyzed between the person of singing and singer.
The present invention is realized using following scheme:A kind of song recommendations method based on singer's sound speciality, including it is as follows Step:
Step S1:The numbered musical notation information of song in Qu Ku is analyzed, the MIDI pitch consensus sequences of each song is obtained, analyzes The sound level distribution histogram of song, obtains the performance range requirement of each song;
Step S2:Sung opera arias recording file using MELODIA Algorithm Analysis users, obtain the MIDI that singer sings the song The MIDI pitch consensus sequences of the same song obtained in pitch value sequence, acquisition step S1, the benchmark for calculating singer is sung Ability, extracts it and sings range;
Step S3:Time frequency signal sign is extracted to the file of singing opera arias of singer, is input in depth convolutional neural networks to net Network is iterated training, the depth convolutional neural networks and voice tone color embedded space trained;
Step S4:Time frequency signal is extracted according to the file of singing opera arias of singer to characterize, and is entered into what is trained in step S3 In depth convolutional neural networks, the output of network is vectorial corresponding to the 3-dimensional tamber characteristic of voice tone color embedded space, by this 3-dimensional Tamber characteristic vector is characterized as the voice tone color of original singer;
Step S5:The sound clip of singing opera arias of singer is analyzed, the same method using step S4 obtains the insertion of voice tone color One group of 3-dimensional tamber characteristic vector in space, is characterized as singer's voice tone color;
Step S6:Range requirement and the performance range of singer are sung according to song, calculated between user and song Range matching degree;
Step S7:Characterized according to the voice tone color of original singer and singer, calculate the tone color phase of singer and each singer Like degree;
Step S8:According to range matching degree and tone color similarity, the recommendation per song for the user in Qu Ku is calculated Degree.
Further, the step S6 specifically includes following steps:
Step S61:The weight of each sound level is obtained according to song sound level distribution histogram, the weight of each sound level is equal to should The summation of all sound level occurrence numbers in number of times that sound level occurs divided by the song, the definition of calculation formula is specially:
Wherein, num (X) represents the number of times that sound level X occurs in numbered musical notation, XmaxThe maximum MIDI values of note in numbered musical notation are represented, XminRepresent the minimum MIDI values of note in numbered musical notation;
Step S62:Using the sound level distribution situation and user of song in the performance capability evaluation value of each sound level, calculate User sings range and the matching degree of song range requirement, and the definition of the calculation formula of range matching degree is specially:
Wherein, U (X) represents performance capability evaluation value of the user on sound level X.
Further, in the step S7, after the sound clip of singer is embedded into tone color embedded space, calculate respectively The tone color similarity of each singer in singer and embedded space, the definition of the calculation formula of tone color similarity is specially:
Tim_sim (u, s)=1-tanh (μ | | Z1-Z2 | |2)
Wherein, | | Z1-Z2 | |2The Euclidean distance between 2 points is represented, μ is empirical coefficient, and tanh is hyperbolic tangent function.
Further, in the step S8, when carrying out final recommendation, the performance range and Qu Ku of user is considered The tone color similitude of singer, is calculated in Qu Ku per first in the matching degree of range requirement and the tone color and Qu Ku of user of middle song Song is for the recommendation degree of the user, and the definition of the calculation formula of recommendation degree is specially:
Recom (u, s)=cRan_mat (u, s)+(1-c) Tim_sim (u, s)
Wherein, Recom (u, s) represent song s for user u recommendation degree, Ran_mat (u, s) represent song s for Family u range matching degree, Tim_sim (u, s) represents the similarity of user u tone color and song s original singer's tone color, and c takes It is worth for 0.7.
Compared with prior art, the invention has the advantages that:This method using user the record of singing opera arias when giving song recitals The numbered musical notation information of sound file and song, on the premise of sound level completes quality height, extracts the performance range of user.Song is utilized simultaneously Bent numbered musical notation and singer sing opera arias the information such as band, set up song features library, extract the performance range and tamber characteristic of singer.With profit Width calculation user and the matching degree of song with range, but in some cases do not apply to compare, utilize song MIDI pitches Model analysis goes out the sound level distribution histogram of song with user in the performance capability evaluation value of each sound level, calculates user's performance Range and the matching degree of song range requirement, the scope of application are wider also more accurate, while user can also be complete during performance It is apt to it and sings range.And, will be higher-dimension, sequential by depth convolutional network powerful dimensionality reduction ability and feature learning ability Voice spectrum signature is embedded into the tone color embedded space of 3-dimensional, so as to realize tone color similitude in 3-dimensional tone color embedded space Metrizability.Finally consider user sings range and the matching degree and the sound of user of the range requirement of song in Qu Ku Color and the tone color similitude of singer in Qu Ku, calculate the recommendation degree per song for the user in Qu Ku.Can be according to difference Application scenarios, different parameter value values are set.So as to recommend the suitable and high song of recommendation degree to user.
Brief description of the drawings
Fig. 1 is the method flow schematic block diagram of the present invention.
Fig. 2 is user U3 parts song recommendations the results list of the present invention.
Embodiment
The present invention will be further described for table and embodiment below in conjunction with the accompanying drawings.
The present embodiment provides a kind of song recommendations method based on singer's sound speciality, and following step is included as shown in Figure 1 Suddenly:
Step S1:The numbered musical notation information of song in Qu Ku is analyzed, the MIDI pitch consensus sequences of each song is obtained, analyzes The sound level distribution histogram of song, obtains the performance range requirement of each song;
Step S2:Sung opera arias recording file using MELODIA Algorithm Analysis users, obtain the MIDI that singer sings the song The MIDI pitch consensus sequences of the same song obtained in pitch value sequence, acquisition step S1, the benchmark for calculating singer is sung Ability, extracts it and sings range;
Step S3:Time frequency signal sign is extracted to the file of singing opera arias of singer, is input in depth convolutional neural networks to net Network is iterated training, the depth convolutional neural networks and voice tone color embedded space trained;
Step S4:Time frequency signal is extracted according to the file of singing opera arias of singer to characterize, and is entered into what is trained in step S3 In depth convolutional neural networks, the output of network is vectorial corresponding to the 3-dimensional tamber characteristic of voice tone color embedded space, by this 3-dimensional Tamber characteristic vector is characterized as the voice tone color of original singer;
Step S5:The sound clip of singing opera arias of singer is analyzed, the same method using step S4 obtains the insertion of voice tone color One group of 3-dimensional tamber characteristic vector in space, is characterized as singer's voice tone color;
Step S6:Range requirement and the performance range of singer are sung according to song, calculated between user and song Range matching degree;
Step S7:Characterized according to the voice tone color of original singer and singer, calculate the tone color phase of singer and each singer Like degree;
Step S8:According to range matching degree and tone color similarity, the recommendation per song for the user in Qu Ku is calculated Degree.
In the present embodiment, the step S6 specifically includes following steps:
Step S61:The weight of each sound level is obtained according to song sound level distribution histogram, the weight of each sound level is equal to should The summation of all sound level occurrence numbers in number of times that sound level occurs divided by the song, the definition of calculation formula is specially:
Wherein, num (X) represents the number of times that sound level X occurs in numbered musical notation, XmaxThe maximum MIDI values of note in numbered musical notation are represented, XminRepresent the minimum MIDI values of note in numbered musical notation.
Step S62:Using the sound level distribution situation and user of song in the performance capability evaluation value of each sound level, calculate User sings range and the matching degree of song range requirement, and the definition of the calculation formula of range matching degree is specially:
Wherein, U (X) represents performance capability evaluation value of the user on sound level X.
In the present embodiment, in the step S7, after the sound clip of singer is embedded into tone color embedded space, respectively The tone color similarity of singer and each singer in embedded space are calculated, the definition of the calculation formula of tone color similarity is specially:
Tim_sim (u, s)=1-tanh (μ | | Z1-Z2 | |2)
Wherein, | | Z1-Z2 | |2The Euclidean distance between 2 points is represented, μ is empirical coefficient, and tanh is hyperbolic tangent function.
In the present embodiment, in the step S8, when carrying out final recommendation, consider the performance range of user with The tone color similitude of singer, is calculated in Qu Ku in the range of song is required in Qu Ku matching degree and the tone color and Qu Ku of user Per song for the recommendation degree of the user, the definition of the calculation formula of recommendation degree is specially:
Recom (u, s)=cRan_mat (u, s)+(1-c) Tim_sim (u, s)
Wherein, Recom (u, s) represent song s for user u recommendation degree, Ran_mat (u, s) represent song s for Family u range matching degree, Tim_sim (u, s) represents the similarity of user u tone color and song s original singer's tone color, and c takes It is worth for 0.7.
In the present embodiment, example is provided according to above method, specifically includes following steps:
Step 1:The numbered musical notation information of song in Qu Ku is analyzed, the MIDI pitch consensus sequences of each song is obtained, analyzes The sound level distribution histogram of song, obtains the performance range requirement of each song.Comprise the following steps that:
Step 11:Collect the numbered musical notation for arranging that user gives song recitals in Qu Ku, tone mark and roll call in song numbered musical notation etc. Information, corresponding MIDI pitches value sequence is converted to by numbered musical notation information, and according between at the beginning of correspondence accompanying song, continue when Between etc. information, set up the standard MIDI pitch parameters files of the song.
Step 12:Go out the sound level distribution histogram of song, the sound level distribution of song by song MIDI pitch value sequence analyses In statistics with histogram song MIDI pitch value sequences, the number of times that each sound level occurs, and the order according to the series of sound from low to high Arrangement.Therefore, the range part that most left and most right two sound level of sound level distribution histogram is included, to should song performance Range requirement.
Step 2:Sung opera arias recording file using MELODIA Algorithm Analysis users, obtain the MIDI that singer sings the song Pitch value sequence, obtains the MIDI pitch consensus sequences of same song in Qu Ku, and the benchmark for calculating singer sings ability, extracts It sings range.Comprise the following steps that:
Step 21:Sung opera arias recording file, MELODIA algorithm energy automatic detection songs using MELODIA Algorithm Analysis users In main melody fundamental frequency F0, design parameter is set to { " minfqr ":82.0,"maxfqr":1047.0,"voicing": 0.2,"minpeaksalience":0.0}.Fundamental frequency F0 is converted into MIDI pitch value p, conversion formula is specifically defined as:
Step 22:Below with user U1, user U1 is male, with certain performance experience, but too high sound is sung Do not get on.The MIDI pitch consensus sequences of same song in Qu Ku are obtained according to step 11, U1 sound level X are counted in song collection The number of times of middle appearance and singer sing quasi- number of times to evaluate sound level X completion quality, are designated as β (X), are added to the history text of user In part, the performance range historical record of user is updated.The Wilson's confidential interval lower bound of each sound level is calculated, after renewal Sound level completes quality.
Step 23:Sound level after traversal updates completes quality, then calculates benchmark and sings ability α.α calculation formula Definition is specially:
α=mean (β (X))-std (β (X))
Wherein, mean (β (X)) represents the average value of each sound level completion quality in total range, and std (β (X)) represents total sound Each sound level completes the standard deviation of quality in domain.Through statistics, user U1 is sung after 20 songs, the average completion of each sound level Quality is 0.84, and standard deviation is 0.16, and it is 0.68 to calculate its benchmark and sing ability α.
Step 24:Use < α, [Xmin,Xmax],sequ[β(Xmin),β(Xmax)], min (β (X)) > represents drilling for singer Range is sung, is a five-tuple.Wherein, XminIt is for the minimum sound level of MIDI values, XmaxFor the sound level that MIDI values are maximum, sequ [β (Xmin),β(Xmax)] it is XminWith XmaxBetween sound level complete mass-sequential, it is min (β (X)) that tonequality, which completes quality floor value,. It is A3 that analysis, which obtains user U1 more than the minimum sound level that benchmark sings ability, and maximum sound level is G5.And each sound level completes quality Sequential value is [0.86,0.99 ...], wherein it is 0.28 that minimum tonequality, which completes quality floor value,.Finally give drilling for user U1 Singing range is<0.68, [A3, G5], [0.86,0.99 ...], 0.28>.
Step 3:Time frequency signal sign is extracted to the file of singing opera arias of singer, is input in depth convolutional neural networks to network Training is iterated, the depth convolutional neural networks and voice tone color embedded space trained.Comprise the following steps that:
Step 31:Singer is divided into one group of 15 people, the audio of singing opera arias to every suite of song hand carries out framing, CQT is extracted to it special Levy, the CQT coefficients of each frame are 192 dimensions, choose the CQT coefficients of 60 frames, constitute the input matrix of neutral net, matrix size is 60*192。
Step 32:The input matrix obtained in step 31 is input to depth convolutional neural networks, and using training in pairs Method training is iterated to network, the depth convolutional neural networks and corresponding tone color embedded space trained.
Step 4:Time frequency signal is extracted according to the file of singing opera arias of singer to characterize, and is entered into the depth trained in step S3 Spend in convolutional neural networks, the output of network is vectorial corresponding to the 3-dimensional tamber characteristic of voice tone color embedded space, by this 3-dimensional sound Color characteristic vector is characterized as the voice tone color of original singer.Comprise the following steps that:
Step 41:To the sound clip of each singer, using the method for same step 31, extract after CQT features, obtain big The small input matrix for 60*192.
Step 42:The CQT features that step 41 is obtained are input to step 32 and train depth convolutional network, the output of network Corresponding to the 3-dimensional tamber characteristic vector of voice tone color embedded space, using the vectorial voice as original singer of this 3-dimensional tamber characteristic Tone color is characterized.
Step 5:The sound clip of singing opera arias of singer is analyzed, with step S4 methods, is obtained in voice tone color embedded space One group of 3-dimensional tamber characteristic vector, is characterized as singer's voice tone color.Comprise the following steps that:
Step 51:To the sound clip of singer, using the method for same step 31, extract after CQT features, obtaining size is 60*192 input matrix.
Step 52:The CQT features that step 51 is obtained are input to step 32 and train depth convolutional network, the output of network Corresponding to the 3-dimensional tamber characteristic vector of voice tone color embedded space, using the vectorial people's sound as singer of this 3-dimensional tamber characteristic Color table is levied.
Step 6:Range requirement and the performance range of singer are sung according to song, calculated between user and song Range matching degree.Comprise the following steps that:
Step 61:The weight of each sound level is obtained according to song sound level distribution histogram, the weight of each sound level is equal to should The summation of all sound level occurrence numbers in number of times that sound level occurs divided by the song, the definition of calculation formula is specially:
Wherein, num (X) represents the number of times that sound level X occurs in numbered musical notation, XmaxThe maximum MIDI values of note in numbered musical notation are represented, XminRepresent the minimum MIDI values of note in numbered musical notation.
Step 62:The song sung with user U2 performance range and Liu Ruoying《When love is close》Carry out instance analysis.With Family U2, women, its sing range be<0.18,[F4,G5],[0.183,0.409,…],0.08>, i.e. user U2 range [F4, G5] in the range of each sound level completeness be 0.183,0.409 ..., performance ability valuation outside range scope is 0.08.The song Range requirement be [D4, G5].
Calculate in song sound level distribution histogram after the weight of each sound level, range matching degree is equal to user at each Performance capability evaluation value in sound level is multiplied by the weight sum of song sound level distribution histogram.The calculation formula of range matching degree Definition is specially:
Wherein, U (X) represents performance capability evaluation value of the user on sound level X, whether belongs to drilling for user according to sound level X Range is sung, U (X) takes different values.Specific formula for calculation is defined:
Wherein, β (X) represents that user is singing the completion quality of each sound level in range, the sound level of min (β (X)) expression users Complete the minimum value of quality.
Finally give user U2 and sing the song that Liu Ruo English is sung《When love is close》Range matching degree be 0.36.
Step 7:Original singer and singer are respectively obtained according to the method for step 4 and step 5 empty in the insertion of each tone color Between in one group of 3-dimensional tamber characteristic vector, using the Euclidean distance between singer, calculate respectively with it is each in each embedded space The tone color similarity of singer.With singer's subset C8 Liang Jingru, Liu Ruoying, Chen little Chun, schoolmate, Fan Weiqi, Fan Xiaoxuan, TANK, Ren Xianqi } exemplified by.
The definition of the calculation formula of tone color similarity is specially:
Tim_sim (u, s)=1-tanh (μ | | Z1-Z2 | |2)
Wherein, | | Z1-Z2 | |2Represent the Euclidean distance between 2 points, μ is empirical coefficient, take 0.05, tanh for hyperbolic just Cut function.The similitude of singer Liang Jingru and other singers are { { Fan Weiqi:0.53 }, { Liu Ruoying:0.55 }, { Fan Xiaoxuan: 0.55 }, { Chen little Chun:0.22},{TANK:0.37 }, { Ren Xianqi:0.40 } a, { schoolmate:0.24}}.
Step 8:When carrying out final recommendation, consider user sings range and the range requirement of song in Qu Ku Matching degree and user tone color and Qu Ku in singer tone color similitude, calculate Qu Ku in per song for the user's Recommendation degree.The calculating of range matching degree is as shown in abovementioned steps 6, and the calculating of tone color similarity is as shown in abovementioned steps 7.With user U3 carry out instance analysis, user U3, male, its sing range model be<0.58,[A3,G5],[0.835,0.602,…], 0.26>, i.e. each sound level completenesses of the user U3 in the range of range [A3, G5] is 0.835,0.302 ..., outside range scope The valuation of performance ability is 0.26.
The definition of the calculation formula of recommendation degree is specially:
Recom (u, s)=cRan_mat (u, s)+(1-c) Tim_sim (u, s)
Wherein, Recom (u, s) represent song s for user u recommendation degree, Ran_mat (u, s) represent song s for Family u range matching degree, Tim_sim (u, s) represents the similarity of user u tone color and song s original singer's tone color, and c is Constant, 0≤c≤1.C values are 0.7.By range matching degree and tone color similarity of the song s to user u, its recommendation is calculated Degree, finally according to recommendation degree, recommends song to user.
End user's U3 recommendation degree highest recommend song details be:Song《10 years》, singer Chen Yixun, range will Ask [C4, F5].Its range matching degree is 0.84, and tone color similarity is 0.88, final to calculate matching degree 0.85.Fig. 2 gives use Part recommendation results more full family U3.
The foregoing is only presently preferred embodiments of the present invention, all equivalent changes done according to scope of the present invention patent with Modification, should all belong to the covering scope of the present invention.

Claims (4)

1. a kind of song recommendations method based on singer's sound speciality, it is characterised in that:Comprise the following steps:
Step S1:The numbered musical notation information of song in Qu Ku is analyzed, the MIDI pitch consensus sequences of each song is obtained, analyzes song Sound level distribution histogram, obtain the performance range requirement of each song;
Step S2:Sung opera arias recording file using MELODIA Algorithm Analysis users, obtain the MIDI pitches that singer sings the song The MIDI pitch consensus sequences of the same song obtained in value sequence, acquisition step S1, the benchmark for calculating singer sings ability, Extract it and sing range;
Step S3:Time frequency signal sign is extracted to the file of singing opera arias of singer, is input in depth convolutional neural networks to enter network Row iteration is trained, the depth convolutional neural networks and voice tone color embedded space trained;
Step S4:Time frequency signal is extracted according to the file of singing opera arias of singer to characterize, and is entered into the depth trained in step S3 In convolutional neural networks, the output of network is vectorial corresponding to the 3-dimensional tamber characteristic of voice tone color embedded space, by this 3-dimensional tone color Characteristic vector is characterized as the voice tone color of original singer;
Step S5:The sound clip of singing opera arias of singer is analyzed, the same method using step S4 obtains voice tone color embedded space In one group of 3-dimensional tamber characteristic vector, be used as singer's voice tone color characterize;
Step S6:Range requirement and the performance range of singer are sung according to song, the sound between user and song is calculated Domain matching degree;
Step S7:Characterized according to the voice tone color of original singer and singer, calculate the tone color similarity of singer and each singer;
Step S8:According to range matching degree and tone color similarity, the recommendation degree per song for the user in Qu Ku is calculated.
2. a kind of song recommendations method based on singer's sound speciality according to claim 1, it is characterised in that:It is described Step S6 specifically includes following steps:
Step S61:The weight of each sound level is obtained according to song sound level distribution histogram, the weight of each sound level is equal to the sound level The summation of all sound level occurrence numbers in the number of times of appearance divided by the song, the definition of calculation formula is specially:
Wherein, num (X) represents the number of times that sound level X occurs in numbered musical notation, XmaxRepresent the maximum MIDI values of note in numbered musical notation, Xmin Represent the minimum MIDI values of note in numbered musical notation;
Step S62:Using the sound level distribution situation and user of song in the performance capability evaluation value of each sound level, user is calculated Range and the matching degree of song range requirement are sung, the definition of the calculation formula of range matching degree is specially:
Wherein, U (X) represents performance capability evaluation value of the user on sound level X.
3. a kind of song recommendations method based on singer's sound speciality according to claim 1, it is characterised in that:It is described In step S7, after the sound clip of singer is embedded into tone color embedded space, calculate each in singer and embedded space respectively The tone color similarity of singer, the definition of the calculation formula of tone color similarity is specially:
Tim_sim (u, s)=1-tanh (μ | | Z1-Z2 | |2)
Wherein, | | Z1-Z2 | |2The Euclidean distance between 2 points is represented, μ is empirical coefficient, and tanh is hyperbolic tangent function.
4. a kind of song recommendations method based on singer's sound speciality according to claim 1, it is characterised in that:It is described In step S8, when carrying out final recommendation, consider user sings range and of the range requirement of song in Qu Ku The tone color similitude of singer in tone color and Qu Ku with degree and user, calculates the recommendation per song for the user in Qu Ku Spend, the definition of the calculation formula of recommendation degree is specially:
Recom (u, s)=cRan_mat (u, s)+(1-c) Tim_sim (u, s)
Wherein, Recom (u, s) represents recommendation degree of the song s for user u, and Ran_mat (u, s) represents song s for user u Range matching degree, Tim_sim (u, s) represents the similarity of user u tone color and song s original singer's tone color, and c values are 0.7。
CN201710206783.2A 2017-03-31 2017-03-31 A kind of song recommendations method based on singer's sound speciality Pending CN106991163A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710206783.2A CN106991163A (en) 2017-03-31 2017-03-31 A kind of song recommendations method based on singer's sound speciality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710206783.2A CN106991163A (en) 2017-03-31 2017-03-31 A kind of song recommendations method based on singer's sound speciality

Publications (1)

Publication Number Publication Date
CN106991163A true CN106991163A (en) 2017-07-28

Family

ID=59415935

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710206783.2A Pending CN106991163A (en) 2017-03-31 2017-03-31 A kind of song recommendations method based on singer's sound speciality

Country Status (1)

Country Link
CN (1) CN106991163A (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108417228A (en) * 2018-02-02 2018-08-17 福州大学 Voice tone color method for measuring similarity under instrument tamber migration
CN109710797A (en) * 2018-11-14 2019-05-03 腾讯科技(深圳)有限公司 Method for pushing, device, electronic device and the storage medium of audio file
CN109726310A (en) * 2018-11-15 2019-05-07 量子云未来(北京)信息科技有限公司 A kind of determination method, apparatus and storage medium for recommending music track
CN109754820A (en) * 2018-12-07 2019-05-14 百度在线网络技术(北京)有限公司 Target audio acquisition methods and device, storage medium and terminal
CN110083772A (en) * 2019-04-29 2019-08-02 北京小唱科技有限公司 Singer's recommended method and device based on singing skills
CN110111773A (en) * 2019-04-01 2019-08-09 华南理工大学 The more New Method for Instrument Recognition of music signal based on convolutional neural networks
CN110364182A (en) * 2019-08-01 2019-10-22 腾讯音乐娱乐科技(深圳)有限公司 A kind of audio signal processing method and device
CN111475672A (en) * 2020-03-27 2020-07-31 咪咕音乐有限公司 Lyric distribution method, electronic equipment and storage medium
CN111488485A (en) * 2020-04-16 2020-08-04 北京雷石天地电子技术有限公司 Music recommendation method based on convolutional neural network, storage medium and electronic device
CN113140228A (en) * 2021-04-14 2021-07-20 广东工业大学 Vocal music scoring method based on graph neural network
CN113593506A (en) * 2021-08-03 2021-11-02 深圳媲客科技有限公司 Singing scoring-based singing voice evaluation system
CN113744759A (en) * 2021-09-17 2021-12-03 广州酷狗计算机科技有限公司 Tone template customizing method and device, equipment, medium and product thereof
CN113762477A (en) * 2021-09-08 2021-12-07 中山大学 Method for constructing sequence recommendation model and sequence recommendation method
CN113779219A (en) * 2021-09-13 2021-12-10 内蒙古工业大学 Question-answering method for embedding multiple knowledge maps by combining hyperbolic segmented knowledge of text

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106095925A (en) * 2016-06-12 2016-11-09 北京邮电大学 A kind of personalized song recommendations system based on vocal music feature

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106095925A (en) * 2016-06-12 2016-11-09 北京邮电大学 A kind of personalized song recommendations system based on vocal music feature

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108417228B (en) * 2018-02-02 2021-03-30 福州大学 Human voice tone similarity measurement method under musical instrument tone migration
CN108417228A (en) * 2018-02-02 2018-08-17 福州大学 Voice tone color method for measuring similarity under instrument tamber migration
CN109710797B (en) * 2018-11-14 2021-03-26 腾讯科技(深圳)有限公司 Audio file pushing method and device, electronic device and storage medium
CN109710797A (en) * 2018-11-14 2019-05-03 腾讯科技(深圳)有限公司 Method for pushing, device, electronic device and the storage medium of audio file
CN109726310A (en) * 2018-11-15 2019-05-07 量子云未来(北京)信息科技有限公司 A kind of determination method, apparatus and storage medium for recommending music track
CN109754820A (en) * 2018-12-07 2019-05-14 百度在线网络技术(北京)有限公司 Target audio acquisition methods and device, storage medium and terminal
CN110111773A (en) * 2019-04-01 2019-08-09 华南理工大学 The more New Method for Instrument Recognition of music signal based on convolutional neural networks
CN110083772A (en) * 2019-04-29 2019-08-02 北京小唱科技有限公司 Singer's recommended method and device based on singing skills
CN110364182B (en) * 2019-08-01 2022-06-14 腾讯音乐娱乐科技(深圳)有限公司 Sound signal processing method and device
CN110364182A (en) * 2019-08-01 2019-10-22 腾讯音乐娱乐科技(深圳)有限公司 A kind of audio signal processing method and device
CN111475672B (en) * 2020-03-27 2023-12-08 咪咕音乐有限公司 Lyric distribution method, electronic equipment and storage medium
CN111475672A (en) * 2020-03-27 2020-07-31 咪咕音乐有限公司 Lyric distribution method, electronic equipment and storage medium
CN111488485B (en) * 2020-04-16 2023-11-17 北京雷石天地电子技术有限公司 Music recommendation method based on convolutional neural network, storage medium and electronic device
CN111488485A (en) * 2020-04-16 2020-08-04 北京雷石天地电子技术有限公司 Music recommendation method based on convolutional neural network, storage medium and electronic device
CN113140228A (en) * 2021-04-14 2021-07-20 广东工业大学 Vocal music scoring method based on graph neural network
CN113593506A (en) * 2021-08-03 2021-11-02 深圳媲客科技有限公司 Singing scoring-based singing voice evaluation system
CN113762477A (en) * 2021-09-08 2021-12-07 中山大学 Method for constructing sequence recommendation model and sequence recommendation method
CN113762477B (en) * 2021-09-08 2023-06-30 中山大学 Method for constructing sequence recommendation model and sequence recommendation method
CN113779219A (en) * 2021-09-13 2021-12-10 内蒙古工业大学 Question-answering method for embedding multiple knowledge maps by combining hyperbolic segmented knowledge of text
CN113779219B (en) * 2021-09-13 2023-07-21 内蒙古工业大学 Question-answering method for embedding multiple knowledge patterns by combining text hyperbolic segmentation knowledge
CN113744759B (en) * 2021-09-17 2023-09-22 广州酷狗计算机科技有限公司 Tone color template customizing method and device, equipment, medium and product thereof
CN113744759A (en) * 2021-09-17 2021-12-03 广州酷狗计算机科技有限公司 Tone template customizing method and device, equipment, medium and product thereof

Similar Documents

Publication Publication Date Title
CN106991163A (en) A kind of song recommendations method based on singer&#39;s sound speciality
CN103823867B (en) Humming type music retrieval method and system based on note modeling
CN110019931B (en) Audio classification method and device, intelligent equipment and storage medium
CN102654859B (en) Method and system for recommending songs
Typke Music retrieval based on melodic similarity
CN109166564A (en) For the method, apparatus and computer readable storage medium of lyrics text generation melody
CN103824565A (en) Humming music reading method and system based on music note and duration modeling
Saitis et al. Perceptual evaluation of violins: A psycholinguistic analysis of preference verbal descriptions by experienced musicians
CN105575393A (en) Personalized song recommendation method based on voice timbre
CN110111773A (en) The more New Method for Instrument Recognition of music signal based on convolutional neural networks
Streich Music complexity: a multi-faceted description of audio content
CN110209869A (en) A kind of audio file recommended method, device and storage medium
CN106898339B (en) Song chorusing method and terminal
CN106095925A (en) A kind of personalized song recommendations system based on vocal music feature
CN106375780A (en) Method and apparatus for generating multimedia file
Rocha et al. Segmentation and timbre-and rhythm-similarity in Electronic Dance Music
CN106302987A (en) A kind of audio frequency recommends method and apparatus
CN109920446A (en) A kind of audio data processing method, device and computer storage medium
US9037278B2 (en) System and method of predicting user audio file preferences
CN112632318A (en) Audio recommendation method, device and system and storage medium
Elowsson et al. Predicting the perception of performed dynamics in music audio with ensemble learning
Mehrabi et al. Similarity measures for vocal-based drum sample retrieval using deep convolutional auto-encoders
CN105825868A (en) Singer effective range extraction method
Foucard et al. Multi-scale temporal fusion by boosting for music classification.
CN106970950B (en) Similar audio data searching method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170728

RJ01 Rejection of invention patent application after publication