CN101697278A - Method for measuring music emotion distance - Google Patents
Method for measuring music emotion distance Download PDFInfo
- Publication number
- CN101697278A CN101697278A CN200910110671A CN200910110671A CN101697278A CN 101697278 A CN101697278 A CN 101697278A CN 200910110671 A CN200910110671 A CN 200910110671A CN 200910110671 A CN200910110671 A CN 200910110671A CN 101697278 A CN101697278 A CN 101697278A
- Authority
- CN
- China
- Prior art keywords
- music
- emotion
- snatch
- distance
- composition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a method for measuring music emotion similarity, which comprises the following steps: receiving a music segment; analyzing the emotional elements of the music segment; determining the main emotional element of the music segment; repeating the previous steps to process all pieces of music likewise; classifying music segments; calculating the emotion distance between music segments; and determining music emotion similarity according to the emotion distance between the music segments. By calculating the emotion distance between pieces of music, the similarity between the pieces of music can be measured accurately, and thus an effective means is provided for searching pieces of music having similar emotions.
Description
[technical field]
The present invention relates to technical field of information processing, relate in particular to a kind of measure of weighing the music emotion distance of the emotion similarity between the music.
[background technology]
Music has become one of indispensable entertainment information in people's daily life, occupies an important position.The quantity of music is also increasing gradually, and number is more and more huger.Different users often likes different genres of music, has promptly comprised the music of different emotions.And sometimes when having different moods, also need to listen some similar music of mood at that time that suit.How to find suitable music in thousands of music, be the problem of an outbalance.
Music searching all is provided in bigger search website at present, provide a kind of " choosing song " service as the google music, the attribute of music is described from three aspects: rhythm is from releiving strongly, tone is from overcast to loud and sonorous, tone color is from enriching to gradually changing merely, and making up the music that has certain emotion to choose, this retrieval mode often has certain randomness.And for example Baidu's music provides the retrieval service that is called " music mood ", and music is classified according to sentimental, lonely, happy, warm and passion etc.
Above-mentioned music assorting or search method all have bigger subjectivity, can not accurately describe the emotion similarity between the music.
[summary of the invention]
Given this, be necessary to have bigger subjectivity at existing music emotion classification and selection method, can not comparatively accurately describe the problem of the emotion similarity between the music, a kind of balancing method of weighing the music emotion similarity of the emotion similarity between the music is provided.
A kind of balancing method of happy emotion similarity comprises the steps: to receive snatch of music; Analyze the emotion composition of snatch of music: determine the emotion disaggregated model, and analyze the emotion composition of described snatch of music according to described emotion disaggregated model, determine the shared number percent of each emotion composition in the described snatch of music, obtain with described number percent be element, the music emotion vector that is associated with described snatch of music; Determine the main emotion composition of snatch of music: with the emotion composition that accounts for the largest percentage main emotion composition as described snatch of music; Repeat above-mentioned steps all music are carried out same processing; Snatch of music is classified: the identical snatch of music of main emotion composition is divided into a class; Emotion distance between the computational music fragment: to belonging to the distance between the music emotion vector that of a sort snatch of music calculates with each snatch of music is associated; Determine the music emotion similarity according to the emotion distance between snatch of music: emotion distance is big more, and similarity is more little; Emotion distance is more little, and similarity is big more.
By the emotion distance between the computational music, can accurately weigh the emotion similarity between the music, for the music retrieval with similar emotion provides a kind of effective means.
Preferably, described emotion disaggregated model is Hevner emotion ring model or thayer thymopsyche model.
Preferably, the emotion distance between the computational music fragment comprises the Euclidean distance that calculates between the music emotion vector be associated with snatch of music, promptly use D (i, j)
IRepresent i snatch of music X in the I class emotion
iWith j snatch of music X
jDistance:
D(i,j)
I=dis(P
i,P
j)
X wherein
i, X
jThe music of ∈ I class emotion, dis is the function of computed range, P
i, P
jBe respectively X
i, X
jThe emotion composition, s is the emotion classification number in the emotion disaggregated model that adopts:
P
i=[a
i,1,a
i,2,...,a
i,s,]
P
j=[a
j,1,a
j,2,...,a
j,s,]
Then the Euclidean distance between the music emotion vector calculates with following formula:
[description of drawings]
Fig. 1 is the process flow diagram of computational music emotion distance;
Fig. 2 is a Hevner emotion ring synoptic diagram.
[embodiment]
As shown in Figure 1, be the process flow diagram of computational music emotion distance.Specifically comprise step:
S10: receive snatch of music.Snatch of music is meant one section music that the emotion composition is more stable.For several minutes song of a head, its emotion keynote is generally all determined, thereby its emotion composition is more stable, can be used as a snatch of music.But long and divide chapters and sections, the music between each chapters and sections to have different emotions for the time, for example symphony etc. then can not be as a snatch of music.
S20: the emotion composition of analyzing snatch of music: determine the emotion disaggregated model, and analyze the emotion composition of described snatch of music according to described emotion disaggregated model, determine the shared number percent of each emotion composition in the described snatch of music, obtain with described number percent be element, the music emotion vector that is associated with described snatch of music.
The emotion disaggregated model has multiple, according to different standards, multiple different disaggregated model can be arranged.For example common have Hevner emotion ring disaggregated model, a Thayer thymopsyche model.Hevner emotion ring disaggregated model is divided into sacredness, sadness, yearning, lyric, slim and graceful, happy, enthusiasm and 8 kinds of life (as shown in Figure 2) with the emotion of music.Thayer thymopsyche model is divided into the emotion of music vigor-and happy, vigor-sentiment, tranquil-happy and tranquil-sentimental 4 kinds arranged.
After determining the emotion disaggregated model, can analyze the emotion composition of a snatch of music, can obtain based on certain emotion constituent analysis algorithm according to the emotion disaggregated model.For a definite emotion disaggregated model, constitute a music emotion vector as component with all affective styles in this disaggregated model.The analysis that a snatch of music is carried out the emotion composition promptly is the shared number percent of component of determining each affective style, and is that element obtains the music emotion vector that is associated with snatch of music with described number percent.
Based on different emotion disaggregated models, obtaining the music emotion vector has diverse ways.Present embodiment is an example with Hevner emotion ring, describes the process of the music emotion vector that obtains a snatch of music.
As shown in Figure 2, be Hevner emotion ring synoptic diagram.Hevner describes the different emotions attribute of 67 descriptive adjective worlds of art of vocabulary utilization, is included into the type that 8 near synonym bunch expressions belong to close emotional responses according to classification, has constituted an annular according to its mutual relation, is called Hevner emotion ring.Wherein any one link all with its before and after adjacent link on affective logic, have progressive relationship, promptly think adjacent emotion before or after in the emotional change of the rational faculty, can being smoothly transitted into it.
By to the emotional semantic Study on similarity, obtain following emotion similar matrix about music emotion Semantic Similarity tolerance:
Music emotion is natural has a composite attribute, this be the caused people of diversity by the music sensation to the mixed feeling cognitive reaction that music took place, isolated with language value or the linguistic expression this composite attribute that all is beyond expression.It is different that different people is held for the emotion intension of music, but this difference is deferred to the constraint of Hevner emotion ring, and the similarity matrix of emotional semantic can be represented this inherent constraint.
Obtain 8 emotion base unit weight e according to above-mentioned matrix with its 8 row vectors
i(i=1~8).
Be defined as follows computing for emotion music vector:
Additive operation: if A, B is 2 music emotion vectors, a
i, b
i(i=1 2...8) is element in the vector respectively, β
1, β
2Be positive constant, then the additive operation of music emotion vector is defined as
Logical operation:
A∨B=(a
i∨b
i)
1×8
A∧B=(a
i∧b
i)
1×8
And following De Morgan rule is satisfied in logical operation:
1∧A=A
0∨A=A
The additive operation of music emotion vector is to logical operation ∧, and ∨ satisfies following law of distribution:
A+(B∨C)=(A+B)∨(A+C)
A+(B∧C)=(A+B)∧(A+C)
And further define subtraction by above-mentioned definition:
If w=[w so
1, w
2... w
8] be the Fuzzy Distribution of certain song about 8 emotion components in the Hevner emotion ring, then can obtain the formula of following computational music emotion vector, wherein e according to the additive operation of emotion vector
iBe the emotion base vector.
S30: the main emotion composition of determining snatch of music: with the emotion composition that accounts for the largest percentage main emotion composition as described snatch of music.Hevner emotion ring has emotion composition in 8, and therefore different music can be taken the emotion composition as the leading factor with these 8 kinds of emotion compositions.
S40: judge whether handled music is last song, is then to be for further processing, otherwise forward step S10 to.
S50: all music are classified: the identical snatch of music of main emotion composition is divided into a class.Therefore all music can be divided into 8 classes as if being as the criterion, all music can be divided into 4 classes as if being as the criterion with Thayer thymopsyche model with Hevner emotion ring model.
S60: the emotion distance between the computational music: calculate distance between each snatch of music to belonging to of a sort snatch of music.Below with the music emotion vector representation be:
P=[a
1,a
2,...,a
s] (1)
Wherein s is an emotion classification number, a
iRepresent i class emotion composition, i=1 ..., s.According to the emotion composition of music, judge its main emotion constituent class I (maximum composition):
I=1 wherein ..., s.
To I class emotion,, constitute the two-dimensional matrix D of M * M dimension to all music calculating distance each other wherein
I, wherein M is the number of fragment in such.D (i, j)
IRepresent i snatch of music X in the I class emotion
iWith j snatch of music X
jDistance:
D(i,j)
I=dis(P
i,P
j) (3)
X wherein
i, X
jThe music of ∈ I class emotion, dis is the function of computed range, P
i, P
jBe respectively X
i, X
jThe emotion composition:
P
i=[a
i,1,a
i,2,...,a
i,s,] (4)
P
j=[a
j,1,a
j,2,...,a
j,s,] (5)
The employing Euclidean distance is represented the emotion distance between the music in the present embodiment, then:
Therefore, matrix D
IFollowing character is arranged:
D(i,j)
I=D(j,i)
I (7)
D(i,i)
I=0 (8)
Wherein, i=1,2 ..., M; J=1,2 ..., M.
Distance table D
IAnd music has wherein promptly constituted I class database.
To arbitrary song, determine after its main emotion composition classification I, according to distance matrix D
ICan obtain itself and all distances of similar other music.Can find and the very high music of this song emotion similarity apart from after sorting these.
The above embodiment has only expressed several embodiment of the present invention, and it describes comparatively concrete and detailed, but can not therefore be interpreted as the restriction to claim of the present invention.Should be pointed out that for the person of ordinary skill of the art without departing from the inventive concept of the premise, can also make some distortion and improvement, these all belong to protection scope of the present invention.Therefore, the protection domain of patent of the present invention should be as the criterion with claims.
Claims (3)
1. the balancing method of a music emotion similarity is characterized in that, comprises the steps:
Receive snatch of music;
Analyze the emotion composition of snatch of music: determine the emotion disaggregated model, and analyze the emotion composition of described snatch of music according to described emotion disaggregated model, determine the shared number percent of each emotion composition in the described snatch of music, obtain with described number percent be element, the music emotion vector that is associated with described snatch of music;
Determine the main emotion composition of snatch of music: with the emotion composition that accounts for the largest percentage main emotion composition as described snatch of music;
Repeat above-mentioned steps all music are carried out same processing;
Snatch of music is classified: the identical snatch of music of main emotion composition is divided into a class;
Emotion distance between the computational music fragment: to belonging to the distance between the music emotion vector that of a sort snatch of music calculates with each snatch of music is associated;
Determine the music emotion similarity according to the emotion distance between snatch of music: emotion distance is big more, and similarity is more little; Emotion distance is more little, and similarity is big more.
2. the balancing method of music emotion similarity as claimed in claim 1 is characterized in that, described emotion disaggregated model is Hevner emotion ring model or thayer thymopsyche model.
3. the balancing method of music emotion similarity as claimed in claim 1 is characterized in that, the emotion distance between the computational music fragment comprises the Euclidean distance that calculates between the music emotion vector be associated with snatch of music, promptly use D (i, j)
IRepresent i snatch of music X in the I class emotion
iWith j snatch of music X
jDistance:
D(i,j)
I=dis(P
i,P
j)
X wherein
i, X
jThe music of ∈ I class emotion, dis is the function of computed range, P
i, P
jBe respectively X
i, X
jThe emotion composition, s is the emotion classification number in the emotion disaggregated model that adopts:
P
i=[a
i,1,a
i,2,...,a
i,s,]
P
j=[a
j,1,a
j,2,...,a
j,s,]
Then the Euclidean distance between the music emotion vector calculates with following formula:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN200910110671A CN101697278A (en) | 2009-10-16 | 2009-10-16 | Method for measuring music emotion distance |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN200910110671A CN101697278A (en) | 2009-10-16 | 2009-10-16 | Method for measuring music emotion distance |
Publications (1)
Publication Number | Publication Date |
---|---|
CN101697278A true CN101697278A (en) | 2010-04-21 |
Family
ID=42142381
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN200910110671A Pending CN101697278A (en) | 2009-10-16 | 2009-10-16 | Method for measuring music emotion distance |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN101697278A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103116646A (en) * | 2013-02-26 | 2013-05-22 | 浙江大学 | Cloud gene expression programming based music emotion recognition method |
CN109299312A (en) * | 2018-10-18 | 2019-02-01 | 湖南城市学院 | Music rhythm analysis method based on big data |
CN111209445A (en) * | 2018-11-21 | 2020-05-29 | 中国电信股份有限公司 | Method and device for recognizing emotion of terminal user |
CN114756734A (en) * | 2022-03-08 | 2022-07-15 | 上海暖禾脑科学技术有限公司 | Music piece segmentation emotion marking system and method based on machine learning |
-
2009
- 2009-10-16 CN CN200910110671A patent/CN101697278A/en active Pending
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103116646A (en) * | 2013-02-26 | 2013-05-22 | 浙江大学 | Cloud gene expression programming based music emotion recognition method |
CN103116646B (en) * | 2013-02-26 | 2015-10-28 | 浙江大学 | A kind of music emotion recognition method based on cloud gene expression programming |
CN109299312A (en) * | 2018-10-18 | 2019-02-01 | 湖南城市学院 | Music rhythm analysis method based on big data |
CN109299312B (en) * | 2018-10-18 | 2021-11-30 | 湖南城市学院 | Music rhythm analysis method based on big data |
CN111209445A (en) * | 2018-11-21 | 2020-05-29 | 中国电信股份有限公司 | Method and device for recognizing emotion of terminal user |
CN111209445B (en) * | 2018-11-21 | 2023-05-02 | 中国电信股份有限公司 | Method and device for identifying emotion of terminal user |
CN114756734A (en) * | 2022-03-08 | 2022-07-15 | 上海暖禾脑科学技术有限公司 | Music piece segmentation emotion marking system and method based on machine learning |
CN114756734B (en) * | 2022-03-08 | 2023-08-22 | 上海暖禾脑科学技术有限公司 | Music piece subsection emotion marking system and method based on machine learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104866572B (en) | A kind of network short text clustering method | |
Araújo et al. | ifeel: A system that compares and combines sentiment analysis methods | |
US8386482B2 (en) | Method for personalizing information retrieval in a communication network | |
CN108038240A (en) | Based on content, the social networks rumour detection method of user's multiplicity | |
CN101667194A (en) | Automatic abstracting method and system based on user comment text feature | |
CN103365867A (en) | Method and device for emotion analysis of user evaluation | |
CN109670039A (en) | Sentiment analysis method is commented on based on the semi-supervised electric business of tripartite graph and clustering | |
CN106156135A (en) | The method and device of inquiry data | |
CN110134820A (en) | A kind of mixed type individualized music recommended method being incremented by based on feature | |
Zhou et al. | Relevance feature mapping for content-based multimedia information retrieval | |
CN106126605A (en) | A kind of short text classification method based on user's portrait | |
US10089409B2 (en) | Event-triggered data quality verification | |
CN103324758A (en) | News classifying method and system | |
CN102760140A (en) | Incident body-based method for expanding searches | |
CN101697278A (en) | Method for measuring music emotion distance | |
Zhao et al. | Genre-based link prediction in bipartite graph for music recommendation | |
Kim | A semantically enhanced tag-based music recommendation using emotion ontology | |
CN115018255A (en) | Tourist attraction evaluation information quality validity analysis method based on integrated learning data mining technology | |
KR20130103249A (en) | Method of classifying emotion from multi sentence using context information | |
CN105975508B (en) | Personalized meta search engine search result synthesizes sort method | |
Argueta et al. | Unsupervised graph-based patterns extraction for emotion classification | |
CN103279549A (en) | Method and device for acquiring target data of target objects | |
Li et al. | Research of music recommendation system based on user behavior analysis and word2vec user emotion extraction | |
Jayawickrama et al. | Seeking sinhala sentiment: Predicting facebook reactions of sinhala posts | |
Cui et al. | Personalized book recommendation based on ontology and collaborative filtering algorithm |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C02 | Deemed withdrawal of patent application after publication (patent law 2001) | ||
WD01 | Invention patent application deemed withdrawn after publication |
Open date: 20100421 |