KR20110017559A - Method and apparatus for analyzing emotion - Google Patents
Method and apparatus for analyzing emotion Download PDFInfo
- Publication number
- KR20110017559A KR20110017559A KR1020090075078A KR20090075078A KR20110017559A KR 20110017559 A KR20110017559 A KR 20110017559A KR 1020090075078 A KR1020090075078 A KR 1020090075078A KR 20090075078 A KR20090075078 A KR 20090075078A KR 20110017559 A KR20110017559 A KR 20110017559A
- Authority
- KR
- South Korea
- Prior art keywords
- voice
- emotional state
- information
- feature
- emotion
- Prior art date
Links
- 230000008451 emotion Effects 0.000 title claims abstract description 129
- 238000000034 method Methods 0.000 title claims abstract description 22
- 238000004458 analytical method Methods 0.000 claims abstract description 114
- 230000002996 emotional effect Effects 0.000 claims abstract description 97
- 238000004891 communication Methods 0.000 claims abstract description 61
- 230000001419 dependent effect Effects 0.000 claims description 11
- 238000009825 accumulation Methods 0.000 claims description 7
- 230000002093 peripheral effect Effects 0.000 claims description 7
- 208000019901 Anxiety disease Diseases 0.000 claims description 5
- 230000036506 anxiety Effects 0.000 claims description 5
- 238000012417 linear regression Methods 0.000 claims description 5
- 238000005070 sampling Methods 0.000 claims description 5
- 238000000605 extraction Methods 0.000 claims description 2
- 239000000284 extract Substances 0.000 abstract description 23
- 238000010586 diagram Methods 0.000 description 6
- 238000010295 mobile communication Methods 0.000 description 5
- 238000004590 computer program Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000012986 modification Methods 0.000 description 4
- 230000004048 modification Effects 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 239000000470 constituent Substances 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- SHXWCVYOXRDMCX-UHFFFAOYSA-N 3,4-methylenedioxymethamphetamine Chemical compound CNC(C)CC1=CC=C2OCOC2=C1 SHXWCVYOXRDMCX-UHFFFAOYSA-N 0.000 description 1
- 241000169170 Boreogadus saida Species 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000007654 immersion Methods 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 238000011002 quantification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W88/00—Devices specially adapted for wireless communication networks, e.g. terminals, base stations or access point devices
- H04W88/18—Service support devices; Network management devices
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Child & Adolescent Psychology (AREA)
- General Health & Medical Sciences (AREA)
- Hospice & Palliative Care (AREA)
- Psychiatry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Networks & Wireless Communication (AREA)
- Telephonic Communication Services (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
One embodiment of the invention relates to a method and apparatus for emotion analysis. More specifically, in the case of extracting emotions on the basis of speech, various emotions are generated through an adaptive emotion analysis algorithm structure that gradually increases the accuracy of emotion analysis by accumulating individual voice feature DB considering the voice feature of a specific speaker. Emotion analysis method and apparatus that can analyze emotions.
Speech recognition is the quantification by analyzing the vibrations of people's phonemes, syllables, words, etc., due to the different characteristics of human accent and pitch. It's a way of finding out. Emotion analysis technology for determining human emotional state is a technique for analyzing human emotion by generating different signals according to the determined emotional state.
However, even though the voice-based emotion analysis service is commercially available, it is not activated at present because the accuracy of the voice emotion analysis is low and the user's convenience is required to improve performance. In other words, in order to increase the field of application of emotion analysis and increase its utilization, it is necessary to develop an algorithm for improving the accuracy of emotion analysis to improve service reliability, and to expand the user base and diversify the service by providing various life necessity services.
In order to solve the above problems, an embodiment of the present invention is to provide a method and apparatus for analyzing emotions that can analyze a variety of accurate emotions for a voice call subscriber.
In order to achieve the above object, an embodiment of the present invention is to sample a voice from a voice call between communication terminals, extract a first voice feature from the sampled voice, and extract the first voice feature. A DB accumulator for accumulating the voice feature DB stored by matching the emotional state in the memory; And an emotional state analyzing unit configured to extract a second speech feature from a voice input to the communication terminal requesting the emotional state, and analyze the emotional state by using the extracted second speech feature and the accumulated speech feature DB. An emotional analysis apparatus is provided.
In addition, according to another object of the invention, the voice sampling step of sampling the voice from the voice call between the communication terminal; Extracting a first voice feature from the sampled voice; A voice feature DB accumulating step of accumulating a voice feature DB stored by matching an emotional state with the extracted first voice feature; A second voice feature extraction step of extracting a second voice feature from a voice input to the communication terminal requesting an emotional state; And an emotional state analyzing step of analyzing an emotional state using the extracted second speech feature and the accumulated speech feature DB.
As described above, according to an exemplary embodiment of the present invention, voice features of all voice call subscribers can be extracted and accumulated in a DB, and more accurate and various emotions of the voice call subscriber can be analyzed through the accumulated DB. It has an effect.
In addition, according to an embodiment of the present invention, it is possible to accumulate a DB matching the emotional state for each individual by extracting the voice feature from the voice of the emotional state is not confirmed from the voice calls of all mobile communication subscribers.
Hereinafter, some embodiments of the present invention will be described in detail through exemplary drawings. In adding reference numerals to the components of each drawing, it should be noted that the same reference numerals are assigned to the same components as much as possible even though they are shown in different drawings. In the following description of the present invention, a detailed description of known functions and configurations incorporated herein will be omitted when it may make the subject matter of the present invention rather unclear.
In addition, in describing the component of this invention, terms, such as 1st, 2nd, A, B, (a), (b), can be used. These terms are only for distinguishing the components from other components, and the nature, order or order of the components are not limited by the terms. If a component is described as being "connected", "coupled" or "connected" to another component, that component may be directly connected to or connected to that other component, but there may be another configuration between each component. It is to be understood that the elements may be "connected", "coupled" or "connected".
1 is a block diagram schematically showing an emotion analysis system according to an embodiment of the present invention.
Emotion analysis system according to an embodiment of the present invention includes a
The
The wired /
The
The
The
The
The
2 is a block diagram schematically illustrating an apparatus for analyzing emotions according to an embodiment of the present invention.
The
The
The DB
The
The
The speaker-dependent voice
The peripheral
The
3 is a flowchart illustrating a emotion analysis method according to an embodiment of the present invention.
The
The
The
The
In FIG. 3, the
As described above, the emotion analysis method according to an embodiment of the present invention described in FIG. 3 may be implemented in a program and recorded in a computer-readable recording medium. The computer-readable recording medium having recorded thereon a program for implementing the emotion analysis method according to an embodiment of the present invention includes all kinds of recording devices storing data that can be read by a computer system. Examples of such computer-readable recording media include ROM, RAM, CD-ROM, magnetic tape, floppy disk, optical data storage, etc., and also implemented in the form of a carrier wave (e.g., transmission over the Internet) . The computer readable recording medium may also be distributed over a networked computer system so that computer readable code is stored and executed in a distributed manner. In addition, functional programs, code, and code segments for implementing an embodiment of the present invention may be easily inferred by programmers skilled in the art to which an embodiment of the present invention belongs.
4 is an exemplary diagram for transmitting the analyzed emotional state of the called party to the caller according to an embodiment of the present invention.
4 illustrates an example of a voice additional service procedure during a call in which a caller (customer) may analyze the other party's emotion through a voice call with the other party (the called party) using the
For example, the
In the above description, all elements constituting the embodiments of the present invention are described as being combined or operating in combination, but the present invention is not necessarily limited to the embodiments. In other words, within the scope of the present invention, all of the components may be selectively operated in combination with one or more. In addition, although all of the components may be implemented in one independent hardware, each or all of the components may be selectively combined to perform some or all functions combined in one or a plurality of hardware. It may be implemented as a computer program having a. Codes and code segments constituting the computer program may be easily inferred by those skilled in the art. Such a computer program may be stored in a computer readable storage medium and read and executed by a computer, thereby implementing embodiments of the present invention. The storage medium of the computer program may include a magnetic recording medium, an optical recording medium, a carrier wave medium, and the like.
In addition, the terms "comprise", "comprise" or "having" described above mean that the corresponding component may be included, unless otherwise stated, and thus excludes other components. It should be construed that it may further include other components instead. All terms, including technical and scientific terms, have the same meanings as commonly understood by one of ordinary skill in the art unless otherwise defined. Terms commonly used, such as terms defined in the dictionary, should be interpreted to coincide with the contextual meaning of the related art, and shall not be construed in an ideal or overly formal sense unless explicitly defined in the present invention.
The foregoing description is merely illustrative of the technical idea of the present invention, and various changes and modifications may be made by those skilled in the art without departing from the essential characteristics of the present invention. Therefore, the embodiments disclosed in the present invention are not intended to limit the technical idea of the present invention but to describe the present invention, and the scope of the technical idea of the present invention is not limited by these embodiments. The protection scope of the present invention should be interpreted by the following claims, and all technical ideas within the equivalent scope should be interpreted as being included in the scope of the present invention.
As described above, the present invention is applied to an application service capable of analyzing various emotions through voice recognition, and extracts voice features from voices whose emotion states are not confirmed from voice calls of all mobile communication subscribers. It is a useful invention that not only can accumulate DB matching a state, but also generate an effect of analyzing a more accurate emotional state using the accumulated DB.
1 is a block diagram schematically showing an emotion analysis system according to an embodiment of the present invention;
2 is a block diagram schematically showing an emotion analysis apparatus according to an embodiment of the present invention;
3 is a flowchart illustrating a emotion analysis method according to an embodiment of the present invention;
4 is an exemplary diagram for transmitting the analyzed emotional state of the called party to the caller according to an embodiment of the present invention.
<Description of Symbols for Main Parts of Drawings>
110: communication terminal 120: wired and wireless communication network
130: emotion analysis device 210: DB accumulation unit
220: emotional state analysis unit 230: message transmission unit
Claims (16)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020090075078A KR20110017559A (en) | 2009-08-14 | 2009-08-14 | Method and apparatus for analyzing emotion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020090075078A KR20110017559A (en) | 2009-08-14 | 2009-08-14 | Method and apparatus for analyzing emotion |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20110017559A true KR20110017559A (en) | 2011-02-22 |
Family
ID=43775525
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020090075078A KR20110017559A (en) | 2009-08-14 | 2009-08-14 | Method and apparatus for analyzing emotion |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20110017559A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013137512A1 (en) * | 2012-03-13 | 2013-09-19 | 주식회사 이루온 | Emotion-based advertisement system and method |
CN104754110A (en) * | 2013-12-31 | 2015-07-01 | 广州华久信息科技有限公司 | Machine voice conversation based emotion release method mobile phone |
US9972341B2 (en) | 2014-01-22 | 2018-05-15 | Samsung Electronics Co., Ltd. | Apparatus and method for emotion recognition |
KR20190022151A (en) * | 2017-08-25 | 2019-03-06 | 강원대학교산학협력단 | Non-contact biosignal detecting method and apparatus thereof |
KR20190069992A (en) * | 2017-12-12 | 2019-06-20 | 세종대학교산학협력단 | Method and system for recognizing emotions based on speaking style |
KR20200143991A (en) | 2019-06-17 | 2020-12-28 | 주식회사 오니온파이브 | Answer recommendation system and method based on text content and emotion analysis |
WO2021162489A1 (en) * | 2020-02-12 | 2021-08-19 | Samsung Electronics Co., Ltd. | Method and voice assistance apparatus for providing an intelligence response |
-
2009
- 2009-08-14 KR KR1020090075078A patent/KR20110017559A/en not_active Application Discontinuation
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013137512A1 (en) * | 2012-03-13 | 2013-09-19 | 주식회사 이루온 | Emotion-based advertisement system and method |
CN104754110A (en) * | 2013-12-31 | 2015-07-01 | 广州华久信息科技有限公司 | Machine voice conversation based emotion release method mobile phone |
US9972341B2 (en) | 2014-01-22 | 2018-05-15 | Samsung Electronics Co., Ltd. | Apparatus and method for emotion recognition |
KR20190022151A (en) * | 2017-08-25 | 2019-03-06 | 강원대학교산학협력단 | Non-contact biosignal detecting method and apparatus thereof |
KR20190069992A (en) * | 2017-12-12 | 2019-06-20 | 세종대학교산학협력단 | Method and system for recognizing emotions based on speaking style |
KR20200143991A (en) | 2019-06-17 | 2020-12-28 | 주식회사 오니온파이브 | Answer recommendation system and method based on text content and emotion analysis |
WO2021162489A1 (en) * | 2020-02-12 | 2021-08-19 | Samsung Electronics Co., Ltd. | Method and voice assistance apparatus for providing an intelligence response |
US11741954B2 (en) | 2020-02-12 | 2023-08-29 | Samsung Eleotronics Co., Ltd. | Method and voice assistance apparatus for providing an intelligence response |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11811970B2 (en) | Voice and speech recognition for call center feedback and quality assurance | |
US10645214B1 (en) | Identical conversation detection method and apparatus | |
KR20110017559A (en) | Method and apparatus for analyzing emotion | |
CN105489221B (en) | A kind of audio recognition method and device | |
US8804918B2 (en) | Method and system for using conversational biometrics and speaker identification/verification to filter voice streams | |
EP2523441A1 (en) | A Mass-Scale, User-Independent, Device-Independent, Voice Message to Text Conversion System | |
US8051134B1 (en) | Systems, methods, and programs for evaluating audio messages | |
US20110004473A1 (en) | Apparatus and method for enhanced speech recognition | |
US20150310877A1 (en) | Conversation analysis device and conversation analysis method | |
CN102780819A (en) | Method of voice recognition of contact for mobile terminal | |
CN107886951B (en) | Voice detection method, device and equipment | |
JP2013011830A (en) | Abnormal state detection device, telephone set, abnormal state detection method, and program | |
CN108831456A (en) | It is a kind of by speech recognition to the method, apparatus and system of video marker | |
CN103856626A (en) | Customization method and device of individual voice | |
JP2010103751A (en) | Method for preventing prohibited word transmission, telephone for preventing prohibited word transmission, and server for preventing prohibited word transmission | |
JP6268916B2 (en) | Abnormal conversation detection apparatus, abnormal conversation detection method, and abnormal conversation detection computer program | |
JP5988077B2 (en) | Utterance section detection apparatus and computer program for detecting an utterance section | |
US9875236B2 (en) | Analysis object determination device and analysis object determination method | |
EP2913822A1 (en) | Speaker recognition method | |
CN113194210A (en) | Voice call access method and device | |
KR100463706B1 (en) | A system and a method for analyzing human emotion based on voice recognition through wire or wireless network | |
CN108040185B (en) | A kind of method and apparatus identifying harassing call | |
US10237399B1 (en) | Identical conversation detection method and apparatus | |
CN113593580B (en) | Voiceprint recognition method and device | |
US20090326940A1 (en) | Automated voice-operated user support |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
N231 | Notification of change of applicant | ||
WITN | Withdrawal due to no request for examination |