CN108735232A - A kind of personality recognition methods and device - Google Patents

A kind of personality recognition methods and device Download PDF

Info

Publication number
CN108735232A
CN108735232A CN201710271654.1A CN201710271654A CN108735232A CN 108735232 A CN108735232 A CN 108735232A CN 201710271654 A CN201710271654 A CN 201710271654A CN 108735232 A CN108735232 A CN 108735232A
Authority
CN
China
Prior art keywords
piece
voice sub
feature information
acoustic feature
emotion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710271654.1A
Other languages
Chinese (zh)
Inventor
谢湘
刘静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN201710271654.1A priority Critical patent/CN108735232A/en
Publication of CN108735232A publication Critical patent/CN108735232A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • G10L25/30Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Child & Adolescent Psychology (AREA)
  • General Health & Medical Sciences (AREA)
  • Hospice & Palliative Care (AREA)
  • Psychiatry (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

The invention discloses a kind of personality recognition methods and device, the method, including:Obtain the sound bite of measurand;The acoustic feature information in each voice sub-piece is extracted respectively, wherein the voice sub-piece is divided to obtain according to the sound bite;For each voice sub-piece, the acoustic feature information in the voice sub-piece is handled using preset algorithm, determines the corresponding type of emotion of voice sub-piece;According to the corresponding type of emotion of each voice sub-piece determined, the personality recognition result of measurand is determined.Using method provided by the invention, point processing such as carry out by the sound bite to measurand, and for each voice sub-piece obtained after decile, acoustic feature information is provided, then the acoustic feature information is handled, not only accurately the personality of measurand is analyzed, moreover it is possible to obtain the personality recognition result of measurand in time.

Description

A kind of personality recognition methods and device
Technical field
The present invention relates to personality analysis technical field more particularly to a kind of personality recognition methods and devices.
Background technology
In recent years, psychological educational circles is formd on personality descriptive model than more consistent common recognition, it is proposed that personality it is big by five Model.Five speciality of big five model are respectively:Nervousness, extropism, opening, compatibility and doing one's duty property, wherein Mei Gete Matter contains six sub- dimensions.Indignation and hostility are to weigh unstrung sub- dimension respectively, reaction be personal experience's indignation and The tendency of correlated condition, such as setback and pain, measurement is that a people generates angry easy degree.Indignation and hostility have height It is intrinsic and low intrinsic.The relatively high people of score shows as being easy to be under fire, especially when experiencing the treatment that oneself is subject to not Animosity can be full of after just, the relatively high people of score can become irascible, indignation and feel to baffle.The relatively low crowd of score, It then can easily control the emotion of oneself, it is not easy to it is under fire and angry, in daily life, the relatively low crowd's meeting of score Close friend is showed, has an amiable disposition, be not easy the image flared up.Generally speaking, unstrung sub- dimension can reflect the choler of people, Indignation and the high people of hostility score show as irritability, and indignation and the relatively low people of hostility score show as not irritability.
The method for measuring big five models, five speciality in the prior art is scaling method, and with indignation and hostility, this sub- dimension is Example illustrates, that is, is directed to this sub- dimension, measurand needs to complete this problem corresponding to sub- dimension, then by measuring people The score of this sub- dimension of member's statistics, obtains the characteristic of this sub- dimension, so as to show whether measurand is with easy Anger personality.In addition, for the accurate anlage for assessing measurand and showing, when to the answer of each problem Using ranking mode, and the range evaluated is generally divided into five grades, i.e., from very different meaning to agreeing to very much.Existing skill Common scale has NEO-PI-R etc. in the scaling method used in art.
However when being measured in the prior art to five features of measurand using scaling method, can generally exist following Problem:(1) in person to person's interactive process, the choler timeliness that measurand is obtained by way of filling in scale is relatively low, It can not timely feedback whether measurand has choler personality;(2) it is only filled in the people of measurand Long Term Contact Whether the accurate reaction measurand of answer ability of the problems in scale has choler personality.
How accurately and in time in conclusion, whether reflection measurand has the technology that choler personality is urgently to be resolved hurrily One of problem.
Invention content
A kind of personality recognition methods of offer of the embodiment of the present invention and device, accurately to analyze the personality of measurand, and Improve the timeliness of character analysis.
The embodiment of the present invention provides a kind of personality recognition methods, including:
Obtain the sound bite of measurand;
The acoustic feature information in each voice sub-piece is extracted respectively, wherein the voice sub-piece is according to institute's predicate Tablet section is divided;
For each voice sub-piece, using preset algorithm to the acoustic feature information in the voice sub-piece at Reason, determines the corresponding type of emotion of voice sub-piece;
According to the corresponding type of emotion of each voice sub-piece determined, the mood personality identification knot of measurand is determined Fruit.
Preferably, each voice sub-piece is directed to, using preset algorithm to the acoustic feature information in the voice sub-piece It is handled, determines the corresponding character type of voice sub-piece, specifically include:
The acoustic feature information is handled using preset algorithm, determines the corresponding mood of the acoustic feature information Measured value;
Compare the emotional measurement value and obtains comparison result with default mood threshold value;
According to comparison result, the corresponding type of emotion of voice sub-piece is determined.
Preferably, being handled the acoustic feature information using preset algorithm, the acoustic feature information pair is determined The emotional measurement value answered, specifically includes:
The acoustic feature information is handled according to following formula, determines the corresponding mood of the acoustic feature information Measured value:
Wherein, x indicates the acoustic feature information;
F (x) indicates the corresponding emotional measurement value of the acoustic feature information;
Indicate optimal Lagrange multiplier vector;
b*Indicate optimal hyperlane intercept;
yiFor predetermined value, value is { -1 ,+1 };
N indicates the quantity of sample in training set;
K (x, z) is gaussian kernel function, andWherein z indicates the acoustic feature information Mean value;σ indicates the standard deviation of the acoustic feature information.
Further, it according to comparison result, determines the corresponding type of emotion of voice sub-piece, specifically includes:
If the comparison result, which is the emotional measurement value, is more than or equal to the default mood threshold value, it is determined that the voice The corresponding type of emotion of sub-piece is negative-morality;Or
If the comparison result, which is the emotional measurement value, is less than the default mood threshold value, it is determined that the voice sub-pieces The corresponding type of emotion of section is positive mood.
Further, according to the corresponding type of emotion of each voice sub-piece determined, the personality of measurand is determined Recognition result specifically includes:
According to the corresponding type of emotion of each voice sub-piece determined, voice that type of emotion is negative-morality is determined The quantity of segment;
Judge whether the type of emotion is more than preset quantity threshold value for the quantity of the voice sub-piece of negative-morality;
If it is, determining that the personality recognition result of measurand is irritability personality;
If it is not, then determining the personality recognition result of measurand for not irritability personality.
The embodiment of the present invention provides a kind of personality identification device, including:
Acquiring unit, the sound bite for obtaining measurand;
Extraction unit, for extracting the acoustic feature information in each voice sub-piece respectively, wherein the voice sub-pieces Section is divided to obtain according to the sound bite;
First determination unit, for being directed to each voice sub-piece, using preset algorithm to the sound in the voice sub-piece It learns characteristic information to be handled, determines the corresponding type of emotion of voice sub-piece;
Second determination unit, for according to the corresponding type of emotion of each voice sub-piece determined, determining tested pair The personality recognition result of elephant.
Preferably, first determination unit, specifically includes:First determination subelement, comparing subunit and second determine Subelement, wherein:
First determination subelement determines institute for being handled the acoustic feature information using preset algorithm State the corresponding emotional measurement value of acoustic feature information;
The comparing subunit obtains comparison result for the emotional measurement value and default mood threshold value;
Second determination subelement, for according to comparison result, determining the corresponding type of emotion of voice sub-piece.
Preferably, first determination subelement, is specifically used for carrying out the acoustic feature information according to following formula Processing, determines the corresponding emotional measurement value of the acoustic feature information:
Wherein, x indicates the acoustic feature information;
F (x) indicates the corresponding emotional measurement value of the acoustic feature information;
Indicate optimal Lagrange multiplier vector;
b*Indicate optimal hyperlane intercept;
yiFor predetermined value, value is { -1 ,+1 };
N indicates the quantity of sample in training set;
K (x, z) is gaussian kernel function, andWherein z indicates the acoustic feature information Mean value;σ indicates the standard deviation of the acoustic feature.
Further, second determination subelement, if being the emotional measurement value specifically for the comparison result More than or equal to the default mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece is negative-morality;Or if institute It is that the emotional measurement value is less than the default mood threshold value to state comparison result, it is determined that the corresponding mood class of the voice sub-piece Type is positive mood.
Further, second determination unit, specifically includes:Third determination subelement, judgment sub-unit, the 4th determine Subelement and the 5th determination subelement, wherein:
The third determination subelement, for according to the corresponding type of emotion of each voice sub-piece determined, determining Type of emotion is the quantity of the voice sub-piece of negative-morality;
The judgment sub-unit, for judging whether the type of emotion is more than for the quantity of the voice sub-piece of negative-morality Preset quantity threshold value;
4th determination subelement, if the judging result for the judgment sub-unit is yes, it is determined that measurand Personality recognition result is irritability personality;
5th determination subelement, if the judging result for the judgment sub-unit is no, it is determined that measurand Personality recognition result is not irritability personality.
Personality recognition methods provided in an embodiment of the present invention and device, after the sound bite for obtaining measurand, to institute It states sound bite to be divided, is then directed to each voice sub-piece obtained after dividing, extracts the sound in the voice sub-piece Characteristic information is learned, according to the acoustic feature information, the acoustic feature information is handled using preset algorithm, determines voice The corresponding type of emotion of segment determines the property of measurand after the completion of waiting for that the corresponding type of emotion of each voice sub-piece determines Lattice recognition result can not only accurately analyze the personality of measurand, moreover it is possible to obtain the personality identification of measurand in time As a result, more improving user experience.
Other features and advantages of the present invention will be illustrated in the following description, also, partly becomes from specification It obtains it is clear that understand through the implementation of the invention.The purpose of the present invention and other advantages can be by the explanations write Specifically noted structure is realized and is obtained in book, claims and attached drawing.
Description of the drawings
Attached drawing described herein is used to provide further understanding of the present invention, and constitutes the part of the present invention, this hair Bright illustrative embodiments and their description are not constituted improper limitations of the present invention for explaining the present invention.In the accompanying drawings:
Fig. 1 a are the flow diagram for the personality recognition methods that the embodiment of the present invention one provides;
Fig. 1 b are the flow signal for the method that the corresponding type of emotion of voice sub-piece is determined in the embodiment of the present invention one Figure;
Fig. 1 c are the flow diagram of the method for the personality recognition result that measurand is determined in the embodiment of the present invention one;
Fig. 2 is the structural schematic diagram of personality identification device provided by Embodiment 2 of the present invention.
Specific implementation mode
A kind of personality recognition methods of offer of the embodiment of the present invention and device, accurately to analyze the personality of measurand, together The timeliness that Shi Tigao analyzes measurand personality.
In invention, it is to be understood that in involved term:
1, emotional measurement value:Emotional measurement value of the present invention is for reflecting whether measurand has irritability personality Evaluation parameter.
2, type of emotion:Type of emotion of the present invention is to reflect whether measurand has irritability personality, institute It states character type and is divided into positive mood and negative-morality.
Below in conjunction with Figure of description, preferred embodiment of the present invention will be described, it should be understood that described herein Preferred embodiment only for the purpose of illustrating and explaining the present invention and is not intended to limit the present invention, and in the absence of conflict, this hair The feature in embodiment and embodiment in bright can be combined with each other.
Embodiment one
As shown in Figure 1a, be the flow diagram of personality recognition methods that the embodiment of the present invention one provides, may include with Lower step:
S11, the sound bite for obtaining measurand.
S12, acoustic feature information in each voice sub-piece is extracted respectively, wherein the voice sub-piece is according to institute State what sound bite was divided.
When it is implemented, in order to improve the accuracy of personality recognition result, after obtaining sound bite, by the voice sheet Section is divided into the voice sub-piece of identical duration, for example, when a length of 10 seconds sound bites can be with when point operation such as carrying out It is halved, trisection can also be carried out, specifically determined by user experience.
Preferably, when being divided to the sound bite, when point operation such as can carry out to the sound bite.Tool Body, it by preset duration etc. point interception sound bite and can be cut after receiving the sound bite of measurand Voice sub-piece after taking, same section is not present in the voice sub-piece after the interception, or when intercepting sound bite, presses The voice sub-piece obtained according to preset duration etc. point interception may exist same section, it is possible thereby to preferably show that personality identifies As a result.For convenience, the embodiment of the present invention is illustrated for point processing such as carrying out to sound bite.
Preferably, the duration of the voice sub-piece after decile can be 2 seconds.I.e.:If a length of 20 when the sound bite Second, then it can carry out ten decile operations.
When it is implemented, it is directed to each voice sub-piece, it, can be with when extracting the acoustic feature information of the voice sub-piece Acoustic feature information is extracted using openSMILE Open-Source Tools.Specifically, the acoustic feature information can be based on 384 dimensions Acoustic feature information.16 Wiki eigens are extracted first with openSMILE, specifically include zero-crossing rate, energy root mean square, base Frequently, harmonic to noise ratio and 12 Jan Vermeer frequency cepstral coefficients and its first-order difference and statistical property, 1 institute of particular content reference table Show.
Preferably, when extracting essential characteristic, can be extracted based on frame level, the essential characteristic and its a scale Divide through 12 statistical functions, finally obtains 16*2*12=384 dimensional features.
Table 1
Preferably, the acoustic feature information is included at least with the next item down:Zero-crossing rate, energy root mean square, fundamental frequency, harmonic wave are made an uproar Acoustic ratio and 12 Jan Vermeer frequency cepstral coefficients.
Specifically, bilingual relationship is in table 1:LLD(low-level descriptors):Bottom describes the factor, Functional:Function, Mean:Mean value, Standard devision:Standard deviation, Kurtosis:Kurtosis, Skewness:Partially Gradient, Extremes:Value, rel.position, range:Maximum value minimum, the position of maximum frame, minimum frame Position, maxima and minima range, Linear regression:offset,slope,MSE:The linear regression of biasing, The linear regression of slope, mean square error linear return.
S13, it is directed to each voice sub-piece, the acoustic feature information in the voice sub-piece is carried out using preset algorithm Processing, determines the corresponding type of emotion of voice sub-piece.
It, can be with when it is implemented, the corresponding type of emotion of voice sub-piece can be determined according to method shown in Fig. 1 b Include the following steps:
S131, the acoustic feature information is handled using preset algorithm, determines that the acoustic feature information corresponds to Emotional measurement value.
When it is implemented, the embodiment of the present invention uses SVM (the Support Vector with gaussian kernel function Machine, support vector machines) algorithm handles the acoustic feature information, determine that the acoustic feature information is corresponding Emotional measurement value.
It should be noted that the basic principle of SVM algorithm is:Classification prediction is carried out to data based on Statistical Learning Theory. It is dedicated to finding the generalization ability that structuring least risk further increases learning machine, to reach empiric risk and fiducial range Minimum, it is final so that in the case where statistical sample amount is fewer, can also obtain good learning effect.The present invention is real The SVM algorithm that example uses is applied as nonlinear support vector machines.I.e.:Kernel function is applied in SVM algorithm, uses one first The data for becoming linearly inseparable of changing commanders are mapped to a new higher dimensional space from original space data are become linear separability, then The rule of the method for linear classification study training data is used in new higher dimensional space.Kernel function is applied in support vector machines (SVM) in, exactly a new feature is mapped to using a nonlinear original input space-theorem in Euclid space Rn that changes commanders that becomes Space-Hilbert space H can be thus that the hypersurface inside original space becomes in new Hilbert space Hyperplane.
Specifically, gaussian kernel function provided in an embodiment of the present invention is one kind in kernel function.Certainly it can also use Polynomial kernel function applies it in SVM algorithm and executes personality recognition methods provided by the invention.
Emotional measurement value described in S132, comparison obtains comparison result with default mood threshold value.
S133, according to comparison result, determine the corresponding type of emotion of voice sub-piece.
Preferably, when executing step S133, if the comparison result is the emotional measurement value be more than or equal to it is described Default mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece is negative-morality;Or
If the comparison result, which is the emotional measurement value, is less than the default mood threshold value, it is determined that the voice sub-pieces The corresponding type of emotion of section is positive mood.
Preferably, the corresponding type of emotion of voice sub-piece can also be determined using following methods:If the comparison As a result it is that the emotional measurement value is equal to the prediction mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece is negative Mood;Or if the comparison result, which is the emotional measurement value, is not equal to the default mood threshold value, it is determined that the voice The corresponding type of emotion of sub-piece is positive mood.Specifically, the default mood threshold value can be -1.
Specifically, the positive mood can be, but not limited to include neutral (neutral) and joy (happy);It is described negative Mood can be, but not limited to include hot anger (angry) and cold anger (cold violence).
For a better understanding of the present invention, the embodiment of the present invention one with preset algorithm be the support with gaussian kernel function to It is illustrated for amount machine SVM algorithm, it, can be true according to formula (1) for each voice sub-piece when executing step S131 Determine the corresponding emotional measurement value of the acoustic feature information:
Wherein, N indicates the quantity of the acoustic feature,Indicate optimal Lagrange multiplier vector, b*Indicate optimal super Plane intercept, yiFor setting value, value is { -1 ,+1 }.
In formula, K (x, z) is gaussian kernel function, shown in expression formula reference formula (2):
X indicates the acoustic feature information in formula (1);F (x) indicates the corresponding emotional measurement of the acoustic feature information Value, z indicate the mean value of the acoustic feature information;σ indicates the standard deviation of the acoustic feature information.
It, will be in the acoustic feature information input to the formula (1) based on above-mentioned formula (1), you can determine described The corresponding emotional measurement value of acoustic feature information.
When it is implemented, before executing personality recognition methods provided by the invention, it is also necessary to being imputed in advance in the present invention The relevant parameter that method is related to is determined, and substantially process is:Sound bite (training set data) is obtained first from corpus, After obtaining sound bite, the sound bite is subjected to decile, for the voice sub-piece after each decile, determines voice Then the corresponding type of emotion of segment determines the personality recognition result of measurand, i.e., according to type of emotion:According to known tested The personality recognition result of object can determine measurand current speech sub-piece after obtaining the voice sub-piece of measurand Type of emotion, i.e.,:It is input in preset algorithm with the voice sub-piece of known type of emotion in advance, based on this determination preset algorithm Middle relevant parameter.
For a better understanding of the present invention, it is illustrated by taking the SVM algorithm with gaussian kernel function as an example, is executing this hair Before bright, it is thus necessary to determine that a in above-mentioned formula (1)i *, z, σ and b*Occurrence.It, can be by based on sound bite in corpus The phonetic decision such as is subjected at point operation according to the method for step S12 of the present invention, can determine that the voice sub-piece corresponds in advance Type of emotion.Then it is directed to each voice sub-piece, extracts the acoustic feature information in the voice sub-piece, you can to obtain X in formula (1).It is unknown due to there is 4 parameters in formula (1), it is therefore desirable to utilize at least four sound bite and its right The type of emotion answered determines the parameter in formula (1), for example, if type of emotion is positive mood, the f (x) in formula (1) It is+1, if type of emotion is negative-morality, the f (x) in formula (1) is -1.So far, it may be determined that go out with gaussian kernel function SVM algorithm in relevant parameter.
Further, when determining the relevant parameter of SVM algorithm according to the voice sub-piece of known type of emotion, in advance will The voice sub-piece of input is labeled, and NEO-PI-R methods mark voice sub-piece specifically may be used, in order to ensure personality The corresponding type of emotion of voice sub-piece of the accuracy of recognition result, selection should not be single.
Preferably, in the relevant parameter in determining formula (1), it can also be using the method for crossing over many times verification come really It is fixed, for example, the number of cross validation can be, but not limited to be 5 times.It is possible thereby to avoid the randomness of parameter determined.
When it is implemented, being illustrated by taking 5 cross validations as an example, 5 voice sub-pieces are obtained first from corpus, The corresponding type of emotion of sound bite is determined respectively, is then directed to the acoustics spy that each voice sub-piece extracts the voice sub-piece Reference ceases, and composing training collection, the training set includes training sample, and the quantity for the sample for being included is N.The trained sample This is the corresponding acoustic feature information of voice sub-piece and type of emotion.Based on the basic principle of 5 cross validations, from training set In 4 groups of training samples of any selection as training data, another group is used as test data, carries out a cross validation and obtains one group Then the parameter value of SVM algorithm repeats 4 times, respectively obtains the parameter value of 4 groups of SVM algorithms, according to according to 5 groups of obtained SVM The parameter value of algorithm determines parameter value of the optimal class value as final SVM algorithm again.
Preferably, in order to improve the accuracy of personality recognition result, in SVM algorithm provided in an embodiment of the present invention, may be used also To introduce penalty coefficient C, then the parameter value in SVM algorithm is determined according to formula (3):
Wherein N indicates training set total sample number amount;aiIt indicates,:Lagrange multiplier vector, hi,hjFor sample in training set The corresponding value of corresponding personality recognition result, value be,:{ -1 ,+1 }, shown in the constraints reference formula (4) of formula (3):
When it is implemented, if the corresponding personality recognition result of sample is irritability personality, corresponding value in training set It is -1;Otherwise its corresponding value is+1.When determining penalty coefficient C, grid data service may be used and select best punishment system Then number determines optimal Lagrange multiplier vector in SVM algorithm further according to formula (3) and (4)Then it chooses to be located at and open Section (0, C)Calculate optimal hyperlane interceptBased on this, by introduce penalty coefficient, So that when executing personality recognition methods provided in an embodiment of the present invention, the SVM algorithm that can be determined according to penalty coefficient obtains More accurate personality recognition result is obtained, user experience is improved.
Preferably, in order to improve the accuracy of personality recognition result, in SVM algorithm provided in an embodiment of the present invention, may be used also To introduce kernel function gamma parameter γ, i.e., the Gaussian kernel in the SVM algorithm used in the embodiment of the present invention with gaussian kernel function Function can introduce gamma parameter γ, that is, the gaussian kernel function expression formula for introducing gamma parameters is:
When it is implemented, gamma parameter γ can be determined using grid data service, based on this using formula (5) as new Gaussian kernel function substitute into formula (1), then determine the type of emotion of the voice sub-piece, it is to be determined go out each voice sub-piece After corresponding type of emotion, the personality recognition result of measurand is further determined that so that according to the Gauss with gamma parameters The SVM algorithm of kernel function determines more accurate personality recognition result, improves user experience.
Preferably, the corresponding emotional measurement value of the acoustic feature information can also be determined using following preset algorithm:ELM (Extreme Learning Machine, extreme learning machine) algorithm, GMM (Gaussian Mixture Model, Gaussian Mixture Model), ANN (Artificial Neural Machine, artificial neural network) algorithms and LR (Logistic Regression, logistic regression) algorithm.
The corresponding type of emotion of each voice sub-piece that S14, basis are determined determines the personality identification knot of measurand Fruit.
Preferably, can be according to method shown in Fig. 1 c according to the corresponding mood class of each voice sub-piece determined Type determines the personality recognition result of measurand, includes the following steps:
The corresponding type of emotion of each voice sub-piece that S141, basis are determined determines that type of emotion is negative-morality The quantity of voice sub-piece.
S142, judge whether the type of emotion is more than preset quantity threshold value for the quantity of the voice sub-piece of negative-morality, If so, thening follow the steps S143;It is no to then follow the steps S144.
S143, determine that the personality recognition result of measurand is irritability personality.
S144, determine the personality recognition result of measurand for not irritability personality.
When it is implemented, the quantity for working as the voice sub-piece for determining that the type of emotion is negative-morality is more than preset quantity When threshold value, then assert that measurand has the characteristics that irritability, excited type personality easily show angry and indignation, temper ratio It is more irascible;When determine the type of emotion be negative-morality voice sub-piece quantity be less than or equal to preset quantity threshold value When, then assert that measurand has the characteristics that not irritability, mood high order smooth pattern personality show as being not easy ignition and angry, temper ratio It is more amiable.
The personality recognition methods that the embodiment of the present invention one provides, after the sound bite for obtaining measurand, by institute's predicate Tablet section such as carries out at point operation, is then directed to each voice sub-piece after decile, and the acoustics extracted in the voice sub-piece is special Reference ceases, and according to the acoustic feature information, handles the acoustic feature information using preset algorithm, determines the voice sub-piece Corresponding type of emotion, after the completion of waiting for that the corresponding type of emotion of each voice sub-piece determines, if it is determined that go out the mood class Type is that the quantity of the voice sub-piece of negative-morality is more than preset quantity threshold value, it is determined that the personality recognition result of measurand is easy Anger personality;Otherwise determine that the personality recognition result of measurand for not irritability personality, can not only accurately analyze tested pair The personality of elephant, moreover it is possible to which the personality recognition result for obtaining measurand in time more improves user experience.
Embodiment two
Based on same inventive concept, a kind of personality identification device is additionally provided in the embodiment of the present invention, due to above-mentioned apparatus The principle solved the problems, such as is similar to personality recognition methods, therefore the implementation of above-mentioned apparatus may refer to the implementation of method, repetition Place repeats no more.
As shown in Fig. 2, for the structural schematic diagram of personality identification device provided by Embodiment 2 of the present invention, including:It obtains single Member 21, extraction unit 22, the first determination unit 23 and the second determination unit 24, wherein:
Acquiring unit 21, the sound bite for obtaining measurand;
Extraction unit 22, for extracting the acoustic feature information in each voice sub-piece respectively, wherein the voice is sub Segment is divided to obtain according to the sound bite;
First determination unit 23, for being directed to each voice sub-piece, using preset algorithm in the voice sub-piece Acoustic feature information is handled, and determines the corresponding type of emotion of voice sub-piece;
Second determination unit 24, for according to the corresponding type of emotion of each voice sub-piece determined, determining tested The personality recognition result of object.
When it is implemented, first determination unit 23, specifically includes:First determination subelement, comparing subunit and Two determination subelements, wherein:
First determination subelement determines institute for being handled the acoustic feature information using preset algorithm State the corresponding emotional measurement value of acoustic feature information;
The comparing subunit obtains comparison result for the emotional measurement value and default mood threshold value;
Second determination subelement, for according to comparison result, determining the corresponding type of emotion of voice sub-piece.
Preferably, first determination subelement, is specifically used for carrying out the acoustic feature information according to following formula Processing, determines the corresponding emotional measurement value of the acoustic feature information:
Wherein, x indicates the acoustic feature information;
F (x) indicates the corresponding emotional measurement value of the acoustic feature information;
Indicate optimal Lagrange multiplier vector;
b*Indicate optimal hyperlane intercept;
yiFor setting value, value is { -1 ,+1 };
N indicates the quantity of sample in training set;
K (x, z) is gaussian kernel function, andWherein z indicates the acoustic feature information Mean value;σ indicates the standard deviation of the acoustic feature information.
Preferably, second determination subelement, if being that the emotional measurement value is big specifically for the comparison result In equal to the default mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece is negative-morality;Or it is if described Comparison result is that the emotional measurement value is less than the default mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece For positive mood.
Preferably, second determination unit 24, specifically includes:Third determination subelement, judgment sub-unit, the 4th determine Subelement and the 5th determination subelement, wherein:
The third determination subelement, for according to the corresponding type of emotion of each voice sub-piece determined, determining Type of emotion is the quantity of the voice sub-piece of negative-morality;
The judgment sub-unit, for judging whether the type of emotion is more than for the quantity of the voice sub-piece of negative-morality Preset quantity threshold value;
4th determination subelement, if the judging result for the judgment sub-unit is yes, it is determined that measurand Personality recognition result is irritability personality;
5th determination subelement, if the judging result for the judgment sub-unit is no, it is determined that measurand Personality recognition result is not irritability personality.
Preferably, the acoustic feature information is included at least with the next item down:Zero-crossing rate, energy root mean square, fundamental frequency, harmonic wave are made an uproar Acoustic ratio and 12 Jan Vermeer frequency cepstral coefficients.
For convenience of description, above each section is divided by function describes respectively for each module (or unit).Certainly, exist Implement the function of each module (or unit) can be realized in same or multiple softwares or hardware when the present invention.
The personality identification device that embodiments herein is provided can be realized by computer program.Those skilled in the art It should be appreciated that above-mentioned module dividing mode is only one kind in numerous module dividing modes, if being divided into other moulds Block or non-division module all should be within the protection domains of the application as long as personality identification device has above-mentioned function.
It should be understood by those skilled in the art that, the embodiment of the present invention can be provided as method, system or computer program Product.Therefore, complete hardware embodiment, complete software embodiment or reality combining software and hardware aspects can be used in the present invention Apply the form of example.Moreover, the present invention can be used in one or more wherein include computer usable program code computer The computer program production implemented in usable storage medium (including but not limited to magnetic disk storage, CD-ROM, optical memory etc.) The form of product.
The present invention be with reference to according to the method for the embodiment of the present invention, the flow of equipment (system) and computer program product Figure and/or block diagram describe.It should be understood that can be realized by computer program instructions every first-class in flowchart and/or the block diagram The combination of flow and/or box in journey and/or box and flowchart and/or the block diagram.These computer programs can be provided Instruct the processor of all-purpose computer, special purpose computer, Embedded Processor or other programmable data processing devices to produce A raw machine so that the instruction executed by computer or the processor of other programmable data processing devices is generated for real The device for the function of being specified in present one flow of flow chart or one box of multiple flows and/or block diagram or multiple boxes.
These computer program instructions, which may also be stored in, can guide computer or other programmable data processing devices with spy Determine in the computer-readable memory that mode works so that instruction generation stored in the computer readable memory includes referring to Enable the manufacture of device, the command device realize in one flow of flow chart or multiple flows and/or one box of block diagram or The function of being specified in multiple boxes.
These computer program instructions also can be loaded onto a computer or other programmable data processing device so that count Series of operation steps are executed on calculation machine or other programmable devices to generate computer implemented processing, in computer or The instruction executed on other programmable devices is provided for realizing in one flow of flow chart or multiple flows and/or block diagram one The step of function of being specified in a box or multiple boxes.
Although preferred embodiments of the present invention have been described, it is created once a person skilled in the art knows basic Property concept, then additional changes and modifications can be made to these embodiments.So it includes excellent that the following claims are intended to be interpreted as It selects embodiment and falls into all change and modification of the scope of the invention.
Obviously, various changes and modifications can be made to the invention without departing from essence of the invention by those skilled in the art God and range.In this way, if these modifications and changes of the present invention belongs to the range of the claims in the present invention and its equivalent technologies Within, then the present invention is also intended to include these modifications and variations.

Claims (10)

1. a kind of personality recognition methods, which is characterized in that including:
Obtain the sound bite of measurand;
The acoustic feature information in each voice sub-piece is extracted respectively, wherein the voice sub-piece is according to the voice sheet What Duan Jinhang was divided;
For each voice sub-piece, the acoustic feature information in the voice sub-piece is handled using preset algorithm, really The fixed corresponding type of emotion of voice sub-piece;
According to the corresponding type of emotion of each voice sub-piece determined, the personality recognition result of measurand is determined.
2. the method as described in claim 1, which is characterized in that each voice sub-piece is directed to, using preset algorithm to the language Acoustic feature information in phone segment is handled, and is determined the corresponding type of emotion of voice sub-piece, is specifically included:
The acoustic feature information is handled using preset algorithm, determines the corresponding emotional measurement of the acoustic feature information Value;
Compare the emotional measurement value and obtains comparison result with default mood threshold value;
According to comparison result, the corresponding type of emotion of voice sub-piece is determined.
3. method as claimed in claim 2, which is characterized in that using preset algorithm to the acoustic feature information at Reason, determines the corresponding emotional measurement value of the acoustic feature information, specifically includes:
The acoustic feature information is handled according to following formula, determines the corresponding emotional measurement of the acoustic feature information Value:
Wherein, x indicates the acoustic feature information;
F (x) indicates the corresponding emotional measurement value of the acoustic feature information;
ai *Indicate optimal Lagrange multiplier vector;
b*Indicate optimal hyperlane intercept;
yiFor predetermined value, value is { -1 ,+1 };
N indicates the quantity of sample in training set;
K (x, z) is gaussian kernel function, andWherein z indicates the mean value of the acoustic feature information; σ indicates the standard deviation of the acoustic feature.
4. method as claimed in claim 2, which is characterized in that according to comparison result, determine the corresponding feelings of voice sub-piece Thread type, specifically includes:
If the comparison result, which is the emotional measurement value, is more than or equal to the default mood threshold value, it is determined that the voice sub-pieces The corresponding type of emotion of section is negative-morality;Or
If the comparison result, which is the emotional measurement value, is less than the default mood threshold value, it is determined that the voice sub-piece pair The type of emotion answered is positive mood.
5. method as claimed in claim 4, which is characterized in that according to the corresponding mood class of each voice sub-piece determined Type determines the personality recognition result of measurand, specifically includes:
According to the corresponding type of emotion of each voice sub-piece determined, determine that type of emotion is the voice sub-piece of negative-morality Quantity;
Judge whether the type of emotion is more than preset quantity threshold value for the quantity of the voice sub-piece of negative-morality;
If it is, determining that the personality recognition result of measurand is irritability personality;
If it is not, then determining the personality recognition result of measurand for not irritability personality.
6. a kind of personality identification device, which is characterized in that including:
Acquiring unit, the sound bite for obtaining measurand;
Extraction unit, for extracting the acoustic feature information in each voice sub-piece respectively, wherein the voice sub-piece is It is divided according to the sound bite;
First determination unit, it is special to the acoustics in the voice sub-piece using preset algorithm for being directed to each voice sub-piece Reference breath is handled, and determines the corresponding type of emotion of voice sub-piece;
Second determination unit, for according to the corresponding type of emotion of each voice sub-piece determined, determining measurand Personality recognition result.
7. device as claimed in claim 6, which is characterized in that first determination unit specifically includes:First determines that son is single Member, comparing subunit and the second determination subelement, wherein:
First determination subelement determines the sound for being handled the acoustic feature information using preset algorithm Learn the corresponding emotional measurement value of characteristic information;
The comparing subunit obtains comparison result for the emotional measurement value and default mood threshold value;
Second determination subelement, for according to comparison result, determining the corresponding type of emotion of voice sub-piece.
8. device as claimed in claim 7, which is characterized in that first determination subelement is specifically used for according to following public affairs Formula handles the acoustic feature information, determines the corresponding emotional measurement value of the acoustic feature information:
Wherein, x indicates the acoustic feature information;
F (x) indicates the corresponding emotional measurement value of the acoustic feature information;
ai *Indicate optimal Lagrange multiplier vector;
b*Indicate optimal hyperlane intercept;
yiFor setting value, value is { -1 ,+1 };
N indicates the quantity of sample in training set;
K (x, z) is gaussian kernel function, andWherein z indicates the mean value of the acoustic feature information; σ indicates the standard deviation of the acoustic feature information.
9. device as claimed in claim 7, which is characterized in that
Second determination subelement, if specifically for the comparison result be the emotional measurement value be more than or equal to it is described pre- If mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece is negative-morality;Or if the comparison result is institute It states emotional measurement value and is less than the default mood threshold value, it is determined that the corresponding type of emotion of voice sub-piece is positive mood.
10. device as claimed in claim 6, which is characterized in that second determination unit specifically includes:Third determines son Unit, judgment sub-unit, the 4th determination subelement and the 5th determination subelement, wherein:
The third determination subelement, for according to the corresponding type of emotion of each voice sub-piece determined, determining mood Type is the quantity of the voice sub-piece of negative-morality;
The judgment sub-unit is preset for judging the type of emotion for whether the quantity of the voice sub-piece of negative-morality is more than Amount threshold;
4th determination subelement, if the judging result for the judgment sub-unit is yes, it is determined that the personality of measurand Recognition result is irritability personality;
5th determination subelement, if the judging result for the judgment sub-unit is no, it is determined that the personality of measurand Recognition result is not irritability personality.
CN201710271654.1A 2017-04-24 2017-04-24 A kind of personality recognition methods and device Pending CN108735232A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710271654.1A CN108735232A (en) 2017-04-24 2017-04-24 A kind of personality recognition methods and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710271654.1A CN108735232A (en) 2017-04-24 2017-04-24 A kind of personality recognition methods and device

Publications (1)

Publication Number Publication Date
CN108735232A true CN108735232A (en) 2018-11-02

Family

ID=63934087

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710271654.1A Pending CN108735232A (en) 2017-04-24 2017-04-24 A kind of personality recognition methods and device

Country Status (1)

Country Link
CN (1) CN108735232A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109451188A (en) * 2018-11-29 2019-03-08 平安科技(深圳)有限公司 Method, apparatus, computer equipment and the storage medium of the self-service response of otherness
WO2021027117A1 (en) * 2019-08-15 2021-02-18 平安科技(深圳)有限公司 Speech emotion recognition method and appartus, and computer-readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20080086791A (en) * 2007-03-23 2008-09-26 엘지전자 주식회사 Feeling recognition system based on voice
CN102623009A (en) * 2012-03-02 2012-08-01 安徽科大讯飞信息技术股份有限公司 Abnormal emotion automatic detection and extraction method and system on basis of short-time analysis
CN102831891A (en) * 2011-06-13 2012-12-19 富士通株式会社 Processing method and system for voice data
JP2015099304A (en) * 2013-11-20 2015-05-28 日本電信電話株式会社 Sympathy/antipathy location detecting apparatus, sympathy/antipathy location detecting method, and program
CN106250855A (en) * 2016-08-02 2016-12-21 南京邮电大学 A kind of multi-modal emotion identification method based on Multiple Kernel Learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20080086791A (en) * 2007-03-23 2008-09-26 엘지전자 주식회사 Feeling recognition system based on voice
CN102831891A (en) * 2011-06-13 2012-12-19 富士通株式会社 Processing method and system for voice data
CN102623009A (en) * 2012-03-02 2012-08-01 安徽科大讯飞信息技术股份有限公司 Abnormal emotion automatic detection and extraction method and system on basis of short-time analysis
JP2015099304A (en) * 2013-11-20 2015-05-28 日本電信電話株式会社 Sympathy/antipathy location detecting apparatus, sympathy/antipathy location detecting method, and program
CN106250855A (en) * 2016-08-02 2016-12-21 南京邮电大学 A kind of multi-modal emotion identification method based on Multiple Kernel Learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
张向君: "《信号分析与数据统计学习》", 28 February 2009, 哈尔滨工程大学出版社 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109451188A (en) * 2018-11-29 2019-03-08 平安科技(深圳)有限公司 Method, apparatus, computer equipment and the storage medium of the self-service response of otherness
WO2021027117A1 (en) * 2019-08-15 2021-02-18 平安科技(深圳)有限公司 Speech emotion recognition method and appartus, and computer-readable storage medium

Similar Documents

Publication Publication Date Title
CN108597492B (en) Phoneme synthesizing method and device
Tsanas et al. Objective automatic assessment of rehabilitative speech treatment in Parkinson's disease
Bone et al. Robust unsupervised arousal rating: A rule-based framework withknowledge-inspired vocal features
Frauenfelder et al. Simulating the time course of spoken word recognition: An analysis of lexical competition in TRACE
Cummins et al. Generalized two-stage rank regression framework for depression score prediction from speech
CN110147732A (en) Refer to vein identification method, device, computer equipment and storage medium
CN110033022A (en) Processing method, device and the storage medium of text
CN110085211A (en) Speech recognition exchange method, device, computer equipment and storage medium
US11763174B2 (en) Learning material recommendation method, learning material recommendation device, and learning material recommendation program
Kurz-Milcke et al. Heuristic decision making
CN111382573A (en) Method, apparatus, device and storage medium for answer quality assessment
Fontes et al. Classification system of pathological voices using correntropy
WO2010021723A1 (en) Content and quality assessment method and apparatus for quality searching
US20230252315A1 (en) Adjusting mental state to improve task performance
CN109685104B (en) Determination method and device for recognition model
JP2018045062A (en) Program, device and method automatically grading from dictation voice of learner
Dsouza et al. Chat with bots intelligently: A critical review & analysis
Velmurugan et al. Developing a fidelity evaluation approach for interpretable machine learning
Shibata et al. Analytic automated essay scoring based on deep neural networks integrating multidimensional item response theory
CN108735232A (en) A kind of personality recognition methods and device
KR101745874B1 (en) System and method for a learning course automatic generation
Sundarprasad Speech emotion detection using machine learning techniques
Loughran et al. Feature selection for speaker verification using genetic programming
Banerjee et al. Relation extraction using multi-encoder lstm network on a distant supervised dataset
CN113705792A (en) Personalized recommendation method, device, equipment and medium based on deep learning model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20181102