CN109087670A - Mood analysis method, system, server and storage medium - Google Patents

Mood analysis method, system, server and storage medium Download PDF

Info

Publication number
CN109087670A
CN109087670A CN201811005214.2A CN201811005214A CN109087670A CN 109087670 A CN109087670 A CN 109087670A CN 201811005214 A CN201811005214 A CN 201811005214A CN 109087670 A CN109087670 A CN 109087670A
Authority
CN
China
Prior art keywords
target
intonation
data
mood
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811005214.2A
Other languages
Chinese (zh)
Other versions
CN109087670B (en
Inventor
申王萍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Wingtech Electronic Technology Co Ltd
Original Assignee
Xian Wingtech Electronic Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Wingtech Electronic Technology Co Ltd filed Critical Xian Wingtech Electronic Technology Co Ltd
Priority to CN201811005214.2A priority Critical patent/CN109087670B/en
Publication of CN109087670A publication Critical patent/CN109087670A/en
Application granted granted Critical
Publication of CN109087670B publication Critical patent/CN109087670B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit

Abstract

The embodiment of the invention discloses a kind of mood analysis method, system, server and storage mediums, wherein the described method includes: language keywords and target intonation included in the voice data that identification is got;According to the language keywords and target intonation of acquisition, analyzes and determine phonetic feature point;Target emotion model, and the spotting phonetic feature point on target emotion model are generated based on the phonetic feature point;Target emotion model is matched with the standard mood model in standard mood model library, to adjust calibrated target voice characteristic point on target emotion model, and records the delta data of target voice characteristic point;By the delta data in database intonation characteristic and Psychological behavioral Characteristic data match, and user emotion or emotional change data are exported according to matching result.It can reach the emotional change of objective, accurate analysis user as a result, and help the purpose of user management mood.

Description

Mood analysis method, system, server and storage medium
Technical field
The present invention relates to data analysis technique field more particularly to a kind of mood analysis method, system, server and storages Medium.
Background technique
The emotional expression of people is for personal abnormal important, if cannot be released in due course, over time, psychology and physiology Variation will be generated.In today of rapid economic development, treatment for specific group, people are more and more unable to do what one wishes.Therefore Being badly in need of one kind can help the specific group there are emotional problems objectively to analyze mood by intelligent method, manage mood.
It is most common at present to understand by assigning computer system identification by affection computation machine technology, express and adapt to The emotional ability of people, to help user objectively to analyze mood, manages mood to establish harmonious man-machine environment.Common analysis The method of user emotion, which has, analyzes user emotion by speech recognition technology and face recognition technology.Only using speech recognition technology It is the sensitive word by including in identification voice to analyze user emotion, accuracy is low.When using face recognition technology, by adopting Collect the expressive features of face, and then analyzes user emotion by analysis expressive features.But people face expression typically last for when Between it is shorter in addition one dodge and die, be not easy to acquire, and the expression of people expresses opposite mood sometimes, so that leading to sometimes Face discriminance analysis user emotion is crossed to be inaccurate.Accordingly, it is difficult to reach the emotional change by objective and accurate analysis user, Help the purpose of user management mood.
Summary of the invention
The embodiment of the invention provides a kind of mood analysis method, system, server and storage mediums, pass through visitor to reach The emotional change of accurately analysis user is seen, the purpose of user management mood is helped.
In a first aspect, the embodiment of the invention provides a kind of mood analysis methods, comprising:
Identify language keywords and target intonation included in the voice data got;
According to the language keywords and target intonation of acquisition, analyzes and determine phonetic feature point, wherein the voice Characteristic point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation;
Target emotion model, and the spotting voice on the target emotion model are generated based on the phonetic feature point Characteristic point;
The target emotion model is matched with the standard mood model in standard mood model library, described in adjustment Calibrated target voice characteristic point on target emotion model, and record the delta data of target voice characteristic point;
By the delta data in database intonation characteristic and Psychological behavioral Characteristic data match, and root User emotion or emotional change data are exported according to matching result.
Second aspect, the embodiment of the invention provides a kind of mood analysis systems, comprising:
Identification module, included language keywords and target intonation in the voice data got for identification;
Characteristic point analysis module is analyzed for the language keywords and target intonation according to acquisition and determines voice Characteristic point, wherein the phonetic feature point be the language keywords and target intonation acceptance of the bid know user emotion keyword and Intonation;
Modeling module, for generating basic target emotion model based on the phonetic feature point analyzed and determined, and The spotting phonetic feature point on the target emotion model;
Adjust logging modle, for by the standard mood model in the target emotion model and standard mood model library into Row matching, to adjust calibrated target voice characteristic point on the target emotion model, and records target voice characteristic point Delta data;
Mood analysis module, for by the intonation characteristic and Psychological behavioral Characteristic in the delta data and database Data are matched, and export user emotion or emotional change data according to matching result.
The third aspect, the embodiment of the invention also provides a kind of servers, comprising:
One or more processors;
Memory, for storing one or more programs,
When one or more of programs are executed by one or more of processors, so that one or more of processing Device realizes the mood analysis method as described in any in the embodiment of the present invention.
Fourth aspect, the embodiment of the invention also provides a kind of computer readable storage mediums, are stored thereon with computer Program realizes the mood method as described in any in the embodiment of the present invention when program is executed by processor.
A kind of mood analysis method, system, server and storage medium provided in an embodiment of the present invention, according to the language extracted Sound characteristic point establishes target emotion model, and the spotting phonetic feature point on target emotion model, filters out energy to reach The purpose of enough phonetic feature points for accurately showing user emotion, then by the mark in target emotion model and standard mood model library Quasi- mood model is matched, adjust target emotion model on calibrated target voice characteristic point, make target emotion model with Master pattern is closer, the subsequent intonation characteristic according in target voice characteristic point delta data and database and psychological row It is characterized data to be matched, fast and accurately analyzes user emotion variation.
Detailed description of the invention
Fig. 1 is a kind of flow diagram for mood analysis method that the embodiment of the present invention one provides;
Fig. 2 is a kind of flow diagram of mood analysis method provided by Embodiment 2 of the present invention;
Fig. 3 is a kind of structural schematic diagram for mood analysis system that the embodiment of the present invention three provides;
Fig. 4 is a kind of structure chart for server that the embodiment of the present invention four provides.
Specific embodiment
The present invention is described in further detail with reference to the accompanying drawings and examples.It is understood that this place is retouched The specific embodiment stated is used only for explaining the present invention rather than limiting the invention.It also should be noted that in order to just Only the parts related to the present invention are shown in description, attached drawing rather than entire infrastructure.
Embodiment one
Fig. 1 is the flow chart for the mood analysis method that the embodiment of the present invention one provides, and the present embodiment is applicable to presence The specific group of emotional problems objectively analyzes and manages mood, and this method can be executed by mood analysis system, the system Such as it can be only fitted in server.The method specifically includes:
Included language keywords and target intonation in the voice data that S110, identification are got.
Wherein it is possible to acquire the voice data of user by preset rules, illustratively, according to preset time rule, User's voice data interior for a period of time or voice data in different time periods in acquisition user one day are acquired, specifically, can To be acquired by corresponding voice collecting software.
For the voice data acquired, can be identified by speech recognition technology semantic content included by voice data and Intonation, wherein intonation includes volume, word speed and tone of voice data etc..Illustratively, acoustic model and language mould can be passed through Type analysis identifies the semantic content namely content of text in voice data.
According to speech recognition as a result, extracting language keywords and target language from the semantic content and intonation identified Adjust, wherein the target language stealthily substitute in the volume for including voice data, word speed, tone and respective variation tendency at least one Kind.Illustratively, it using the meaningless word in participle dictionary removal semantic content, while extracting and can be shown that user emotion Language keywords;For the intonation of identification, screening wherein meets the conduct target intonation of preset condition, illustratively, by volume It is more than a certain more than maximum preset threshold value and screening as a kind of target intonation lower than minimum preset threshold, or by word speed Preset threshold is also used as target intonation.
S120, the language keywords and target intonation according to acquisition are analyzed and are determined phonetic feature point, wherein institute Predicate sound characteristic point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation.
The language keywords and target language tune got are further analyzed and screened, determination can wherein clearly show that The keyword and intonation of user emotion are as phonetic feature point, wherein phonetic feature point includes that keyword feature point and intonation are special Sign point.Illustratively, language keywords can be screened by the mood sensitive word dictionary established in advance, and will filtered out Keyword is determined as keyword feature point, wherein mood sensitive word dictionary includes the vocabulary often said under the various different moods of user. It, can be using the obvious point of variation tendency as intonation spy since target intonation is usually to be shown in the form of waveform diagram Levy point, such as the point that word speed is accelerated suddenly.
S130, target emotion model is generated based on the phonetic feature point, and demarcates mesh on the target emotion model Poster sound characteristic point.
Target emotion model is generated according to determining phonetic feature point, so as to according to target emotion model analysis user's feelings Thread.The spotting phonetic feature point on target emotion model, wherein target voice characteristic point can be the language determined in S120 Feature a part more outstanding in sound characteristic point, it is thus achieved that the further screening to user emotion feature, so that user Emotional characteristics it is more obvious.
S140, the target emotion model is matched with the standard mood model in standard mood model library, to adjust Calibrated target voice characteristic point on the whole target emotion model, and record the delta data of target voice characteristic point.
Target emotion model is matched with the standard mood model in standard mood model library, that is to say target feelings The phonetic feature point demarcated on thread model is matched with the characteristic point in standard mood model, including keyword and intonation Match, according to the data tendency of matching result and each target voice characteristic point to calibrated target voice on target emotion model Characteristic point is finely adjusted, and records the delta data of target voice characteristic point.Illustratively, by target voice characteristic point from position A It is adjusted to position B, records the situation of change of target voice characteristic point each feature between position A and B, such as Speed variation, Intonation variation and volume change etc..
S150, by the delta data and database intonation characteristic and Psychological behavioral Characteristic data carry out Match, and user emotion or emotional change data are exported according to matching result.
According to the intonation characteristic and Psychological behavioral Characteristic number in the delta data and database of target voice characteristic point According to matching result, export user emotion or emotional change data.Illustratively, if word speed frequency is fast, give great volume, intonation becomes Change big, then it is assumed that user emotion is more eager;If word speed frequency is slow, volume is moderate, intonation is gentle, then it is assumed that user emotion compares Steadily.
In the present embodiment, target emotion model is established according to the phonetic feature point extracted, and on target emotion model Then spotting phonetic feature point will with achieving the purpose that filter out the phonetic feature point that can accurately show user emotion Target emotion model is matched with the mood model in standard mood model library, adjusts calibrated mesh on target emotion model Poster sound characteristic point keeps target emotion model and master pattern closer, it is subsequent according to target voice characteristic point delta data with Intonation characteristic and Psychological behavioral Characteristic data in database are matched, and user emotion variation is fast and accurately analyzed.
Embodiment two
Fig. 2 is a kind of flow diagram of mood analysis method provided by Embodiment 2 of the present invention, and the present embodiment is above-mentioned It is optimized on the basis of embodiment, which comprises
Included language keywords and target intonation in the voice data that S210, identification are got.
S220, the language keywords and target intonation according to acquisition are analyzed and are determined phonetic feature point, wherein institute Predicate sound characteristic point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation.
S230, convolution algorithm is obtained as a result, based on the convolution algorithm knot to phonetic feature point progress convolution algorithm Fruit generates target emotion model, and the extreme point in the convolution algorithm result is demarcated as target voice characteristic point.
In the present embodiment, convolution algorithm is carried out to the phonetic feature point that S220 is obtained and obtains convolution algorithm as a result, simultaneously basis The data tendency of each characteristic point in convolution algorithm result generates target emotion model, illustratively, the target emotion mould of generation Type can be a waveform diagram.Further, user emotion can be clearly showed that in the phonetic feature point in order to screen acquisition The corresponding phonetic feature point of extreme point in the illustraton of model such as wave crest and wave trough position is demarcated as target voice by characteristic point Characteristic point.
S240, the target emotion model is matched with the standard mood model in standard mood model library, to adjust Calibrated target voice characteristic point on the whole target emotion model, and record the delta data of target voice characteristic point.
S250, by the delta data and database intonation characteristic and Psychological behavioral Characteristic data carry out Match, and user emotion or emotional change data are exported according to matching result.
Further, in the present embodiment, it in order to improve the accuracy that user emotion is analyzed, is calculated by big data analysis To update the intonation characteristic and Psychological behavioral Characteristic data in the database.Illustratively, periodically acquisition interconnection Network data, and analytical calculation is carried out to collected internet data, constantly update the intonation characteristic in the database With Psychological behavioral Characteristic data, wherein the database is local data base or internet database.
In the present embodiment, target emotion model, and root are established by carrying out convolution algorithm to the phonetic feature point of acquisition According to the extreme value spotting phonetic feature point in calculated result, reach the mesh that screening clearly shows that user emotion phonetic feature point , while improving the accuracy of user emotion analysis, periodically the intonation characteristic and Psychology and behavior more in new database Characteristic is convenient for user emotion management so that mood analysis is more acurrate.
Embodiment three
Fig. 3 is a kind of structural schematic diagram for mood analysis system that the embodiment of the present invention three provides, as shown in figure 3, described System includes:
Identification module 310, included language keywords and target intonation in the voice data got for identification;
Characteristic point analysis module 320 is analyzed for the language keywords and target intonation according to acquisition and determines language Sound characteristic point, wherein the phonetic feature point is the keyword that user emotion is known in the language keywords and the acceptance of the bid of target intonation And intonation;
Modeling module 330, for generating basic target emotion model based on the phonetic feature point analyzed and determined, And the spotting phonetic feature point on the target emotion model;
Logging modle 340 is adjusted, for by the standard mood mould in the target emotion model and standard mood model library Type is matched, and to adjust calibrated target voice characteristic point on the target emotion model, and records target voice feature The delta data of point;
Mood analysis module 350, for by the intonation characteristic and Psychology and behavior in the delta data and database Characteristic is matched, and exports user emotion or emotional change data according to matching result.
In the present embodiment, characteristic point analysis module determines phonetic feature point according to the recognition result of identification module, models mould Root tuber establishes target emotion model according to phonetic feature point, and adjustment logging modle is finely adjusted target emotion model, and records mesh The variation of poster sound characteristic point, mood analysis module analyze the mood and feelings of user according to the delta data of target voice characteristic point Thread variation.It can reach the emotional change by objective, accurate analysis user, as a result, to help the mesh of user management mood 's.
On the basis of the above embodiments, the identification module includes:
Acquisition unit, for acquiring the voice data of user according to preset rules;
Recognition unit, for identification semantic content and intonation included by the voice data;
Extraction unit, for extracting language keyword and target intonation from the semantic content and intonation identified, wherein The target language is stealthily substituted the volume for including voice data, word speed, tone and respective variation tendency.
On the basis of the above embodiments, the modeling module includes:
Modeling unit obtains convolution algorithm as a result, based on the volume for carrying out convolution algorithm to the phonetic feature point Product operation result generates target emotion model.
On the basis of the above embodiments, the modeling module includes:
Unit is demarcated, for the extreme point in the convolution algorithm result to be demarcated as target voice characteristic point.
On the basis of the above embodiments, the system also includes:
Big data handles update module, for calculating the intonation characteristic updated in the database by big data analysis According to Psychological behavioral Characteristic data, wherein the database be local data base or internet database.
Mood provided by any embodiment of the invention point can be performed in mood analysis system provided by the embodiment of the present invention Analysis method has the corresponding functional module of execution method and beneficial effect.
Example IV
Fig. 4 is the structural schematic diagram for the server that the embodiment of the present invention four provides.Fig. 4, which is shown, to be suitable for being used to realizing this hair The block diagram of the exemplary servers 12 of bright embodiment.The server 12 that Fig. 4 is shown is only an example, should not be to the present invention The function and use scope of embodiment bring any restrictions.
As shown in figure 4, server 12 is showed in the form of universal computing device.The component of server 12 may include but not Be limited to: one or more processor or processing unit 16, system storage 28 connect different system components (including system Memory 28 and processing unit 16) bus 18.
Bus 18 indicates one of a few class bus structures or a variety of, including memory bus or Memory Controller, Peripheral bus, graphics acceleration port, processor or the local bus using any bus structures in a variety of bus structures.It lifts For example, these architectures include but is not limited to industry standard architecture (ISA) bus, microchannel architecture (MAC) Bus, enhanced isa bus, Video Electronics Standards Association (VESA) local bus and peripheral component interconnection (PCI) bus.
Server 12 typically comprises a variety of computer system readable media.These media can be and any can be serviced The usable medium that device 12 accesses, including volatile and non-volatile media, moveable and immovable medium.
System storage 28 may include the computer system readable media of form of volatile memory, such as arbitrary access Memory (RAM) 30 and/or cache memory 32.Server 12 may further include other removable/nonremovable , volatile/non-volatile computer system storage medium.Only as an example, storage system 34 can be used for reading and writing not removable Dynamic, non-volatile magnetic media (Fig. 4 do not show, commonly referred to as " hard disk drive ").Although not shown in fig 4, it can provide Disc driver for being read and write to removable non-volatile magnetic disk (such as " floppy disk "), and to removable anonvolatile optical disk The CD drive of (such as CD-ROM, DVD-ROM or other optical mediums) read-write.In these cases, each driver can To be connected by one or more data media interfaces with bus 18.Memory 28 may include at least one program product, The program product has one group of (for example, at least one) program module, these program modules are configured to perform each implementation of the invention The function of example.
Program/utility 40 with one group of (at least one) program module 42 can store in such as memory 28 In, such program module 42 include but is not limited to operating system, one or more application program, other program modules and It may include the realization of network environment in program data, each of these examples or certain combination.Program module 42 is usual Execute the function and/or method in embodiment described in the invention.
Server 12 can also be logical with one or more external equipments 14 (such as keyboard, sensing equipment, display 24 etc.) Letter, can also be enabled a user to one or more equipment interact with the server 12 communicate, and/or with make the server The 12 any equipment (such as network interface card, modem etc.) that can be communicated with one or more of the other calculating equipment communicate. This communication can be carried out by input/output (I/O) interface 22.Also, server 12 can also pass through network adapter 20 With one or more network (such as local area network (LAN), wide area network (WAN) and/or public network, such as internet) communication. As shown, network adapter 20 is communicated by bus 18 with other modules of server 12.It should be understood that although not showing in figure Out, can in conjunction with server 12 use other hardware and/or software module, including but not limited to: microcode, device driver, Redundant processing unit, external disk drive array, RAID system, tape drive and data backup storage system etc..
Processing unit 16 by the program that is stored in system storage 28 of operation, thereby executing various function application and Data processing, such as realize that mood provided by the embodiment of the present invention analyzes classification method, comprising:
Identify language keywords and target intonation included in the voice data got;
According to the language keywords and target intonation of acquisition, analyzes and determine phonetic feature point, wherein the voice Characteristic point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation;
Target emotion model, and the spotting phonetic feature on target emotion model are generated based on the phonetic feature point Point;
Target emotion model is matched with the standard mood model in standard mood model library, to adjust target emotion Calibrated target voice characteristic point on model, and record the delta data of target voice characteristic point;
By the delta data in database intonation feature and Psychological behavioral Characteristic match, and according to matching tie Fruit exports user emotion or emotional change data.
In one embodiment, the program that processing unit 16 is stored in system storage 28 by operation, realization Mood analysis method, further includes:
The voice data of user is acquired according to preset rules;
Identify semantic content included by the voice data and intonation;
Language keywords and target intonation are extracted from the semantic content and intonation identified, wherein the target intonation At least one of volume, word speed, tone and respective variation tendency including voice data.
In one embodiment, the program that processing unit 16 is stored in system storage 28 by operation, realization Mood analysis method, further includes:
Convolution algorithm is carried out to the phonetic feature point and obtains convolution algorithm as a result, generating based on the convolution algorithm result Target emotion model.
In one embodiment, the program that processing unit 16 is stored in system storage 28 by operation, realization Mood analysis method, further includes:
Extreme point in the convolution algorithm result is demarcated as target voice characteristic point.
In one embodiment, the program that processing unit 16 is stored in system storage 28 by operation, realization Mood analysis method, further includes:
It is calculated by big data analysis to update the intonation characteristic and Psychological behavioral Characteristic data in the database, Wherein, the database is local data base or internet database.
Embodiment five
The embodiment of the present invention five additionally provides a kind of computer readable storage medium, is stored thereon with computer program, should The mood analysis method as provided by the embodiment of the present invention is realized when program is executed by processor, comprising:
Identify language keywords and target intonation included in the voice data got;
According to the language keywords and target intonation of acquisition, analyzes and determine phonetic feature point, wherein the voice Characteristic point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation;
Target emotion model, and the spotting phonetic feature on target emotion model are generated based on the phonetic feature point Point;
Target emotion model is matched with the standard mood model in standard mood model library, to adjust target emotion Calibrated target voice characteristic point on model, and record the delta data of target voice characteristic point;
By the delta data in database intonation feature and Psychological behavioral Characteristic match, and according to matching tie Fruit exports user emotion or emotional change data.
In one embodiment, it can also be realized when which is executed by processor:
The voice data of user is acquired according to preset rules;
Identify semantic content included by the voice data and intonation;
Language keywords and target intonation are extracted from the semantic content and intonation identified, wherein the target intonation At least one of volume, word speed, tone and respective variation tendency including voice data.
In one embodiment, it can also be realized when which is executed by processor:
Convolution algorithm is carried out to the phonetic feature point and obtains convolution algorithm as a result, generating based on the convolution algorithm result Target emotion model.
In one embodiment, it can also be realized when which is executed by processor:
Extreme point in the convolution algorithm result is demarcated as target voice characteristic point.
In one embodiment, it can also be realized when which is executed by processor:
It is calculated by big data analysis to update the intonation characteristic and Psychological behavioral Characteristic data in the database, Wherein, the database is local data base or internet database.
The computer storage medium of the embodiment of the present invention, can be using any of one or more computer-readable media Combination.Computer-readable medium can be computer-readable signal media or computer readable storage medium.It is computer-readable Storage medium for example may be-but not limited to-the system of electricity, magnetic, optical, electromagnetic, infrared ray or semiconductor, device or Device, or any above combination.The more specific example (non exhaustive list) of computer readable storage medium includes: tool There are electrical connection, the portable computer diskette, hard disk, random access memory (RAM), read-only memory of one or more conducting wires (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, portable compact disc read-only memory (CD- ROM), light storage device, magnetic memory device or above-mentioned any appropriate combination.In this document, computer-readable storage Medium can be any tangible medium for including or store program, which can be commanded execution system, device or device Using or it is in connection.
Computer-readable signal media may include in a base band or as carrier wave a part propagate data-signal, Wherein carry computer-readable program code.The data-signal of this propagation can take various forms, including but unlimited In electromagnetic signal, optical signal or above-mentioned any appropriate combination.Computer-readable signal media can also be that computer can Any computer-readable medium other than storage medium is read, which can send, propagates or transmit and be used for By the use of instruction execution system, device or device or program in connection.
The program code for including on computer-readable medium can transmit with any suitable medium, including --- but it is unlimited In wireless, electric wire, optical cable, RF etc. or above-mentioned any appropriate combination.
The computer for executing operation of the present invention can be write with one or more programming languages or combinations thereof Program code, described program design language include object oriented program language-such as Java, Smalltalk, C++, Further include conventional procedural programming language-such as " C " language or similar programming language.Program code can be with It fully executes, partly execute on the user computer on the user computer, being executed as an independent software package, portion Divide and partially executes or executed on a remote computer or server completely on the remote computer on the user computer.? Be related in the situation of remote computer, remote computer can pass through the network of any kind --- including local area network (LAN) or Wide area network (WAN)-be connected to subscriber computer, or, it may be connected to outer computer (such as mentioned using Internet service It is connected for quotient by internet).
Note that the above is only a better embodiment of the present invention and the applied technical principle.It will be appreciated by those skilled in the art that The invention is not limited to the specific embodiments described herein, be able to carry out for a person skilled in the art it is various it is apparent variation, It readjusts and substitutes without departing from protection scope of the present invention.Therefore, although being carried out by above embodiments to the present invention It is described in further detail, but the present invention is not limited to the above embodiments only, without departing from the inventive concept, also It may include more other equivalent embodiments, and the scope of the invention is determined by the scope of the appended claims.

Claims (10)

1. a kind of mood analysis method, which is characterized in that the described method includes:
Identify language keywords and target intonation included in the voice data got;
According to the language keywords and target intonation of acquisition, analyzes and determine phonetic feature point, wherein the phonetic feature Point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation;
Target emotion model, and the spotting phonetic feature on the target emotion model are generated based on the phonetic feature point Point;
The target emotion model is matched with the standard mood model in standard mood model library, to adjust the target Calibrated target voice characteristic point in mood model, and record the delta data of target voice characteristic point;
By the delta data in database intonation characteristic and Psychological behavioral Characteristic data match, and according to User emotion or emotional change data are exported with result.
2. the method according to claim 1, wherein described identify language included in the voice data got Say keyword and target intonation, comprising:
The voice data of user is acquired according to preset rules;
Identify semantic content included by the voice data and intonation;
Language keywords and target intonation are extracted from the semantic content and intonation identified, wherein the target language, which is stealthily substituted, to be included At least one of volume, word speed, tone and respective variation tendency of voice data.
3. the method according to claim 1, wherein based on the phonetic feature point generate target emotion model, Include:
Convolution algorithm is carried out to the phonetic feature point and obtains convolution algorithm as a result, generating target based on the convolution algorithm result Mood model.
4. according to the method described in claim 3, it is characterized in that, on target emotion model spotting phonetic feature point, Include:
Extreme point in the convolution algorithm result is demarcated as target voice characteristic point.
5. the method according to claim 1, wherein the method also includes:
It is calculated by big data analysis to update the intonation characteristic and Psychological behavioral Characteristic data in the database, In, the database is local data base or internet database.
6. a kind of mood analysis system, which is characterized in that the system comprises:
Identification module, included language keywords and target intonation in the voice data got for identification;
Characteristic point analysis module is analyzed for the language keywords and target intonation according to acquisition and determines phonetic feature Point, wherein the phonetic feature point is the keyword and intonation that user emotion is known in the language keywords and the acceptance of the bid of target intonation;
Modeling module, for generating basic target emotion model based on the phonetic feature point analyzed and determined, and in institute State spotting phonetic feature point on target emotion model;
Logging modle is adjusted, for carrying out the standard mood model in the target emotion model and standard mood model library Match, to adjust calibrated target voice characteristic point on the target emotion model, and records the variation of target voice characteristic point Data;
Mood analysis module, in the delta data and database intonation characteristic and Psychological behavioral Characteristic data into Row matching, and user emotion or emotional change data are exported according to matching result.
7. according to right want 6 described in system, which is characterized in that the identification module includes:
Acquisition unit, for acquiring the voice data of user according to preset rules;
Recognition unit, for identification semantic content and intonation included by the voice data;
Extraction unit, for extracting language keyword and target intonation from the semantic content and intonation identified, wherein described Target language is stealthily substituted at least one of the volume for including voice data, word speed, tone and respective variation tendency.
8. according to right want 6 described in system, which is characterized in that the system also includes:
Big data handles update module, for calculated by big data analysis update intonation characteristic in the database with Psychological behavioral Characteristic data, wherein the database is local data base or internet database.
9. a kind of server, which is characterized in that the server includes:
One or more processors;
Memory, for storing one or more programs,
When one or more of programs are executed by one or more of processors, so that one or more of processors are real Now such as mood analysis method as claimed in any one of claims 1 to 5.
10. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the program is by processor Such as mood analysis method as claimed in any one of claims 1 to 5 is realized when execution.
CN201811005214.2A 2018-08-30 2018-08-30 Emotion analysis method, system, server and storage medium Active CN109087670B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811005214.2A CN109087670B (en) 2018-08-30 2018-08-30 Emotion analysis method, system, server and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811005214.2A CN109087670B (en) 2018-08-30 2018-08-30 Emotion analysis method, system, server and storage medium

Publications (2)

Publication Number Publication Date
CN109087670A true CN109087670A (en) 2018-12-25
CN109087670B CN109087670B (en) 2021-04-20

Family

ID=64840278

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811005214.2A Active CN109087670B (en) 2018-08-30 2018-08-30 Emotion analysis method, system, server and storage medium

Country Status (1)

Country Link
CN (1) CN109087670B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110110135A (en) * 2019-04-17 2019-08-09 西安极蜂天下信息科技有限公司 Voice characteristics data library update method and device
CN110399837A (en) * 2019-07-25 2019-11-01 深圳智慧林网络科技有限公司 User emotion recognition methods, device and computer readable storage medium
CN110534135A (en) * 2019-10-18 2019-12-03 四川大学华西医院 A method of emotional characteristics are assessed with heart rate response based on language guidance
CN111354377A (en) * 2019-06-27 2020-06-30 深圳市鸿合创新信息技术有限责任公司 Method and device for recognizing emotion through voice and electronic equipment
CN111447124A (en) * 2020-04-02 2020-07-24 张瑞华 Intelligent household control method and intelligent control equipment based on biological feature recognition
CN112037821A (en) * 2019-06-03 2020-12-04 阿里巴巴集团控股有限公司 Visual representation method and device of voice emotion and computer storage medium
CN112037820A (en) * 2019-05-16 2020-12-04 杭州海康威视数字技术股份有限公司 Security alarm method, device, system and equipment
CN112071304A (en) * 2020-09-08 2020-12-11 深圳市天维大数据技术有限公司 Semantic analysis method and device
CN113284494A (en) * 2021-05-25 2021-08-20 平安普惠企业管理有限公司 Voice assistant recognition method, device, equipment and computer readable storage medium
CN113468983A (en) * 2021-06-15 2021-10-01 杭州海康威视系统技术有限公司 Emotion analysis method, device, equipment and storage medium

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101789990A (en) * 2009-12-23 2010-07-28 宇龙计算机通信科技(深圳)有限公司 Method and mobile terminal for judging emotion of opposite party in conservation process
CN102099853A (en) * 2009-03-16 2011-06-15 富士通株式会社 Apparatus and method for recognizing speech emotion change
CN103593054A (en) * 2013-11-25 2014-02-19 北京光年无限科技有限公司 Question-answering system combining emotion recognition and output
CN104050965A (en) * 2013-09-02 2014-09-17 广东外语外贸大学 English phonetic pronunciation quality evaluation system with emotion recognition function and method thereof
CN104102627A (en) * 2014-07-11 2014-10-15 合肥工业大学 Multi-mode non-contact emotion analyzing and recording system
US20150286858A1 (en) * 2015-03-18 2015-10-08 Looksery, Inc. Emotion recognition in video conferencing
CN105159979A (en) * 2015-08-27 2015-12-16 广东小天才科技有限公司 Good friend recommendation method and apparatus
CN106548788A (en) * 2015-09-23 2017-03-29 中国移动通信集团山东有限公司 A kind of intelligent emotion determines method and system
WO2018061839A1 (en) * 2016-09-29 2018-04-05 株式会社村田製作所 Transmission device, transmission method, and transmission program
CN107948417A (en) * 2017-11-22 2018-04-20 周燕红 A kind of method, apparatus, terminal and the storage medium of voice data monitoring
CN108010516A (en) * 2017-12-04 2018-05-08 广州势必可赢网络科技有限公司 A kind of semanteme independent voice mood characteristic recognition method and device
CN108010512A (en) * 2017-12-05 2018-05-08 广东小天才科技有限公司 The acquisition methods and recording terminal of a kind of audio

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102099853A (en) * 2009-03-16 2011-06-15 富士通株式会社 Apparatus and method for recognizing speech emotion change
CN101789990A (en) * 2009-12-23 2010-07-28 宇龙计算机通信科技(深圳)有限公司 Method and mobile terminal for judging emotion of opposite party in conservation process
CN104050965A (en) * 2013-09-02 2014-09-17 广东外语外贸大学 English phonetic pronunciation quality evaluation system with emotion recognition function and method thereof
CN103593054A (en) * 2013-11-25 2014-02-19 北京光年无限科技有限公司 Question-answering system combining emotion recognition and output
CN104102627A (en) * 2014-07-11 2014-10-15 合肥工业大学 Multi-mode non-contact emotion analyzing and recording system
US20150286858A1 (en) * 2015-03-18 2015-10-08 Looksery, Inc. Emotion recognition in video conferencing
CN105159979A (en) * 2015-08-27 2015-12-16 广东小天才科技有限公司 Good friend recommendation method and apparatus
CN106548788A (en) * 2015-09-23 2017-03-29 中国移动通信集团山东有限公司 A kind of intelligent emotion determines method and system
WO2018061839A1 (en) * 2016-09-29 2018-04-05 株式会社村田製作所 Transmission device, transmission method, and transmission program
CN107948417A (en) * 2017-11-22 2018-04-20 周燕红 A kind of method, apparatus, terminal and the storage medium of voice data monitoring
CN108010516A (en) * 2017-12-04 2018-05-08 广州势必可赢网络科技有限公司 A kind of semanteme independent voice mood characteristic recognition method and device
CN108010512A (en) * 2017-12-05 2018-05-08 广东小天才科技有限公司 The acquisition methods and recording terminal of a kind of audio

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陈鹏展 等: "基于语音信号与文本信息的双模态情感识别", 《华东交通大学学报》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110110135A (en) * 2019-04-17 2019-08-09 西安极蜂天下信息科技有限公司 Voice characteristics data library update method and device
CN112037820B (en) * 2019-05-16 2023-09-05 杭州海康威视数字技术股份有限公司 Security alarm method, device, system and equipment
CN112037820A (en) * 2019-05-16 2020-12-04 杭州海康威视数字技术股份有限公司 Security alarm method, device, system and equipment
CN112037821A (en) * 2019-06-03 2020-12-04 阿里巴巴集团控股有限公司 Visual representation method and device of voice emotion and computer storage medium
CN111354377A (en) * 2019-06-27 2020-06-30 深圳市鸿合创新信息技术有限责任公司 Method and device for recognizing emotion through voice and electronic equipment
CN110399837A (en) * 2019-07-25 2019-11-01 深圳智慧林网络科技有限公司 User emotion recognition methods, device and computer readable storage medium
CN110399837B (en) * 2019-07-25 2024-01-05 深圳智慧林网络科技有限公司 User emotion recognition method, device and computer readable storage medium
CN110534135A (en) * 2019-10-18 2019-12-03 四川大学华西医院 A method of emotional characteristics are assessed with heart rate response based on language guidance
CN111447124A (en) * 2020-04-02 2020-07-24 张瑞华 Intelligent household control method and intelligent control equipment based on biological feature recognition
CN112071304A (en) * 2020-09-08 2020-12-11 深圳市天维大数据技术有限公司 Semantic analysis method and device
CN112071304B (en) * 2020-09-08 2024-03-15 深圳市天维大数据技术有限公司 Semantic analysis method and device
CN113284494A (en) * 2021-05-25 2021-08-20 平安普惠企业管理有限公司 Voice assistant recognition method, device, equipment and computer readable storage medium
CN113284494B (en) * 2021-05-25 2023-12-01 北京基智科技有限公司 Voice assistant recognition method, device, equipment and computer readable storage medium
CN113468983A (en) * 2021-06-15 2021-10-01 杭州海康威视系统技术有限公司 Emotion analysis method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN109087670B (en) 2021-04-20

Similar Documents

Publication Publication Date Title
CN109087670A (en) Mood analysis method, system, server and storage medium
Ancilin et al. Improved speech emotion recognition with Mel frequency magnitude coefficient
CN106683680B (en) Speaker recognition method and device, computer equipment and computer readable medium
CN109243491B (en) Method, system and storage medium for emotion recognition of speech in frequency spectrum
CN107767869B (en) Method and apparatus for providing voice service
US8676574B2 (en) Method for tone/intonation recognition using auditory attention cues
Pawar et al. Convolution neural network based automatic speech emotion recognition using Mel-frequency Cepstrum coefficients
WO2021000408A1 (en) Interview scoring method and apparatus, and device and storage medium
CN110880329B (en) Audio identification method and equipment and storage medium
CN112259106A (en) Voiceprint recognition method and device, storage medium and computer equipment
CN107633851B (en) Discrete speech emotion recognition method, device and system based on emotion dimension prediction
Özseven et al. SPeech ACoustic (SPAC): A novel tool for speech feature extraction and classification
CN113707173B (en) Voice separation method, device, equipment and storage medium based on audio segmentation
Chatterjee et al. Auditory model-based design and optimization of feature vectors for automatic speech recognition
KR20220044446A (en) Method and apparatus for testing vehicle-mounted voice device, electronic device and storage medium
Staudacher et al. Fast fundamental frequency determination via adaptive autocorrelation
CN115273904A (en) Angry emotion recognition method and device based on multi-feature fusion
Chi et al. Robust emotion recognition by spectro-temporal modulation statistic features
Ton-That et al. Speech emotion recognition using a fuzzy approach
CN114595692A (en) Emotion recognition method, system and terminal equipment
CN115240696B (en) Speech recognition method and readable storage medium
CN110321404B (en) Vocabulary entry selection method and device for vocabulary learning, electronic equipment and storage medium
Batzorig et al. Speech recognition in Mongolian language using a neural network with pre-processing technique
CN113761206A (en) Intelligent information query method, device, equipment and medium based on intention recognition
Hajihashemi et al. Novel time-frequency based scheme for detecting sound events from sound background in audio segments

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant