WO2022025025A1 - Emotion analysis system and emotion analysis device - Google Patents

Emotion analysis system and emotion analysis device Download PDF

Info

Publication number
WO2022025025A1
WO2022025025A1 PCT/JP2021/027638 JP2021027638W WO2022025025A1 WO 2022025025 A1 WO2022025025 A1 WO 2022025025A1 JP 2021027638 W JP2021027638 W JP 2021027638W WO 2022025025 A1 WO2022025025 A1 WO 2022025025A1
Authority
WO
WIPO (PCT)
Prior art keywords
emotion
biological reaction
change
index value
subject
Prior art date
Application number
PCT/JP2021/027638
Other languages
French (fr)
Japanese (ja)
Inventor
渉三 神谷
Original Assignee
株式会社I’mbesideyou
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社I’mbesideyou filed Critical 株式会社I’mbesideyou
Priority to JP2022539467A priority Critical patent/JPWO2022025025A1/ja
Publication of WO2022025025A1 publication Critical patent/WO2022025025A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management

Definitions

  • the present invention relates to an emotion analysis system and an emotion analysis device, and more particularly to a system and an apparatus for analyzing emotions of a user participating in an online session.
  • Patent Documents 1 to 4 Conventionally, there is known a technique of comparing a normal (expressionless) facial expression of a subject with a current facial expression to determine the degree of emotion of the subject (for example, Patent Documents 1 to 4). reference). Further, a technique for recognizing emotions in consideration of individuality from voice and images is also known (see, for example, Patent Document 5).
  • Patent Document 4 a reference parameter generated based on the operator's calm voice is stored in advance, and the emotion parameter is compared with the emotion parameter generated based on the operator's voice acquired by the voice acquisition unit. It is disclosed to specify the degree of emotion. Further, in view of the fact that there are individual differences in the appearance of the difference between the reference parameter and the emotion parameter, the average value of the emotion parameter is obtained for each operator, and the difference between the reference parameter and the average value is obtained. It is also disclosed that the emotional degree is relatively specified for each operator based on the ratio of the difference between the reference parameter and the emotional parameter.
  • Japanese Unexamined Patent Publication No. 2011-154665 Japanese Unexamined Patent Publication No. 2012-8949 Japanese Unexamined Patent Publication No. 2013-300 Japanese Unexamined Patent Publication No. 2015-141428 Japanese Unexamined Patent Publication No. 2001-83984
  • the subject A is determined to have a current emotion X level of 3 based on the difference from his / her normal facial expression
  • the subject B is also determined to have a current emotion X level of 3. .
  • the meaning of emotion X (the degree of emotion in the true sense) is different between the level 3 of the subject A who easily causes emotion X and the level 3 of the subject B who does not easily generate emotion X. Since the techniques described in Patent Documents 1 to 3 cannot evaluate the degree of emotion in the true sense of the word, it is not possible to objectively compare the degree of emotion between different subjects.
  • changes in the biological reaction of the subject are analyzed based on the moving images obtained for the subject, and based on the analyzed changes in the biological reaction, the changes in the biological reaction are analyzed.
  • the evaluation criteria leveled among a plurality of subjects the degree of emotions adjusted according to the likelihood of the same emotions being generated by the subjects is evaluated.
  • the degree of emotion is evaluated in consideration of the susceptibility to generate different emotions for each subject, and therefore the degree of true emotions regarding the subject is evaluated. This makes it possible to objectively compare the degree of emotion between different subjects.
  • FIG. 1 is a diagram showing an overall configuration example of an emotion analysis system according to the present embodiment.
  • the emotion analysis system of the present embodiment is referred to as an emotion analysis device 100, a plurality of user terminals 200 -1 , 200-2 , ... (Hereinafter, unless otherwise specified, it is simply referred to as a user terminal 200. ) And a session management server 300.
  • the emotion analysis device 100, the user terminal 200, and the session management server 300 are connected via a communication network 500 such as the Internet or a mobile phone network.
  • the emotion analysis system of the present embodiment is caused by a change in emotions of the subject based on a moving image obtained about the subject (participant of the online session), for example, in an environment where an online session is performed by a plurality of participants. It is a system that analyzes the changes in the biological reaction that occur and evaluates the degree of emotion of the subject according to the evaluation criteria leveled among the plurality of subjects based on the analyzed changes in the biological reaction.
  • the online session is, for example, an online conference, an online class, an online chat, or the like, in which a plurality of user terminals 200 installed at a plurality of locations are connected to the emotion analysis device 100 and the session management server 300 via a communication network 500.
  • a moving image can be exchanged between a plurality of user terminals 200 through an emotion analysis device 100 and a session management server 300.
  • An application program (hereinafter referred to as a session application) necessary for exchanging moving images in an online session is installed in a plurality of user terminals 200.
  • the moving image handled in the online session includes a face image (actually, an image of a body part other than the face and a background image) and a voice of a user (participant in the online session) who uses the user terminal 200. ..
  • the user's face image and voice are acquired by a camera and a microphone provided in the user terminal 200 or connected to the user terminal 200, and transmitted to the session management server 300.
  • the face image and voice of each user transmitted to the session management server 300 are acquired by the emotion analysis device 100, and are transmitted from the emotion analysis device 100 to the session application of each user terminal 200.
  • the moving image transmitted from the user terminal 200 may be acquired by the emotion analysis device 100 and transferred from the emotion analysis device 100 to the session management server 300.
  • the moving image may be transmitted from the user terminal 200 to both the emotion analysis device 100 and the session management server 300.
  • moving images include images such as materials shared and viewed by multiple users.
  • the material image to be viewed by the user is transmitted from any user terminal 200 to the session management server 300. Then, the material image transmitted to the session management server 300 is acquired by the emotion analysis device 100, and is transmitted from the emotion analysis device 100 to the session application of each user terminal 200.
  • the face image or the document image of the plurality of users is displayed on the display in each of the plurality of user terminals 200, and the voices of the plurality of users are output from the speaker.
  • the face image and the material image can be switched to display only one of them on the display screen, or the display area can be divided into the face image and the material image. Can be displayed at the same time. Further, it is possible to display the image of one of a plurality of users on the full screen, or to display the image of a part or all of the users on a small screen.
  • the camera on / off and the microphone on / off by the function of the session application installed in the user terminal 200.
  • the camera is turned off in the user terminal 200-1
  • the face image taken by the camera of the user terminal 200-1 is transmitted to the session management server 300 and the emotion analysis device 100, but each of them is transmitted from the emotion analysis device 100. It is not transmitted to the user terminal 200.
  • the microphone is turned off in the user terminal 200-1
  • the sound collected by the microphone of the user terminal 200-1 is transmitted to the session management server 300 and the emotion analysis device 100, but is transmitted from the emotion analysis device 100. It is not transmitted to each user terminal 200.
  • FIG. 2 is a block diagram showing a functional configuration example of the emotion analysis device 100 according to the present embodiment.
  • the emotion analysis device 100 of the present embodiment includes a moving image acquisition unit 11, a biological reaction analysis unit 12, and an emotion evaluation unit 13 as functional configurations. Further, the emotion analysis device 100 of the present embodiment includes a moving image storage unit 101 as a storage medium.
  • Each of the above functional blocks 11 to 13 can be configured by any of hardware, DSP (Digital Signal Processor), and software.
  • DSP Digital Signal Processor
  • each of the above functional blocks 11 to 13 is actually configured to include a computer CPU, RAM, ROM, etc., and is a program stored in a recording medium such as RAM, ROM, a hard disk, or a semiconductor memory. Is realized by the operation of.
  • the moving image acquisition unit 11 acquires moving images (face image, voice, material image) transmitted from each user terminal 200 during the online session from the session management server 300.
  • the moving image acquisition unit 11 stores the moving image acquired from each user terminal 200 in the moving image storage unit 101 in association with information that can identify each user (for example, a user ID).
  • the moving image acquisition unit 11 acquires the face image from the session management server 300, including the face image displayed on the display of each user terminal 200 and the face image not being displayed. Further, it does not matter whether the sound acquired from the session management server 300 is set to be output from the speaker of each user terminal 200 (whether the microphone is set to on or off). No. That is, the moving image acquisition unit 11 acquires audio from the session management server 300, including audio being output from the speaker of each user terminal 200 and audio being non-output.
  • the biological reaction analysis unit 12 outputs a moving image (whether or not it is a face image displayed on the screen of the user terminal 200, whether or not it is a face image displayed on the screen of the user terminal 200, from the speaker of the user terminal 200) acquired by the moving image acquisition unit 11 and stored in the moving image storage unit 101. Based on (whether or not it is the voice inside), we analyze the changes in biological reactions caused by changes in emotions for each of the multiple participants.
  • the biological reaction analysis unit 12 separates the moving image acquired by the moving image acquisition unit 11 into a set of facial images (a collection of frame images) and voice, and analyzes changes in the biological reaction from each.
  • the biological reaction analysis unit 12 analyzes the user's face image using the frame image separated from the moving image acquired by the moving image acquisition unit 11, and thereby at least one of the facial expression, the line of sight, the pulse, and the movement of the face. Analyze changes in biological reactions related to one. In addition, the biological reaction analysis unit 12 analyzes changes in the biological reaction regarding at least one of the user's speech content and voice quality by analyzing the voice separated from the moving image acquired by the moving image acquisition unit 11.
  • the biological reaction analysis unit 12 calculates a biological reaction index value that reflects the content of the change in the biological reaction by quantifying the change in the biological reaction according to a predetermined standard.
  • the biological reaction analysis unit 12 identifies a facial region from the frame image for each frame image, and analyzes which facial expression element the facial expression corresponds to according to an image analysis model machine-learned in advance. do. Then, based on the analysis result, whether or not a facial expression change occurs between consecutive frame images, and if a facial expression change occurs, whether it is a positive facial expression change or a negative facial expression change, and how large it is. It analyzes whether the facial expression change is occurring, and calculates the facial expression change index value according to the analysis result.
  • Facial expression elements are, for example, neutral / calm / happy / surprised / sad / angry / fearful / disgust. be. Of these, joy and surprise are positive facial expression elements, and sadness, anger, fear, and disgust are negative facial expression elements.
  • the biological reaction analysis unit 12 determines between consecutive frame images depending on whether at least one of the facial expression element determined for each frame image and the facial expression score calculated for each frame image has changed from the previous frame. Determine if the facial expression has changed.
  • the biological reaction analysis unit 12 determines that the facial expression change has occurred when the score change amount from the previous frame is equal to or more than a predetermined threshold value when there is no change in the facial expression element of the maximum score. good.
  • the magnitude of the facial expression change can be determined by the amount of change from the previous frame of the facial expression score.
  • the biological reaction analysis unit 12 causes a positive facial expression change when the facial expression score of the positive facial expression increases from the previous frame and when the negative facial expression of the previous frame changes to the positive facial expression of the current frame. It is determined that it is.
  • the biological reaction analysis unit 12 causes a negative facial expression change when the facial expression score of the negative facial expression increases from the previous frame and when the positive facial expression of the previous frame changes to the negative facial expression of the current frame. It is determined that it is.
  • the biological reaction analysis unit 12 uses the direction of facial expression change (positive ⁇ positive, positive ⁇ negative, negative ⁇ positive, negative ⁇ negative) and the magnitude of facial expression change as explanatory variables, and the facial expression change index value as the objective variable.
  • the facial expression change index value is calculated using a predetermined function. In this function, for example, when the facial expression is reversed (positive ⁇ negative, negative ⁇ positive), the absolute value of the facial expression change index value is larger than when it is not reversed, and the greater the degree of facial expression change, the larger the facial expression change.
  • Negative ⁇ Negative can be a function that has a negative value.
  • the facial expression change may be analyzed every predetermined time interval (for example, every 500 milliseconds). This also applies to the analysis of the change in the line of sight, the analysis of the change in the pulse, and the analysis of the change in the movement of the face described below.
  • Analysis of changes in the line of sight is performed, for example, as follows. That is, the biological reaction analysis unit 12 identifies the eye region from the frame image for each frame image and analyzes the direction (line of sight) of both eyes. Then, the biological reaction analysis unit 12 calculates the line-of-sight change index value according to the analysis result of the line-of-sight change. For example, the biological reaction analysis unit 12 calculates the angle of the line of sight from the front for each frame image, and calculates the moving average or the moving variance between a plurality of frames of the angle as the line-of-sight change index value.
  • the biological reaction analysis unit 12 may analyze where the user is looking.
  • the change in the line of sight is also related to the degree of concentration of the user. For example, it analyzes whether the speaker's face being displayed, the shared material being displayed, or the outside of the screen is being viewed. In addition, it may be possible to analyze whether the movement of the line of sight is large or small, and whether the movement is frequent or infrequent. Then, the biological reaction analysis unit 12 calculates the line-of-sight change index value according to the analysis result of the line-of-sight change.
  • the biological reaction analysis unit 12 uses the place of viewing (speaker's face, shared material, outside the screen), the magnitude of eye movement, and the frequency of eye movement as explanatory variables, and the eye change index value.
  • the line-of-sight change index value is calculated using a predetermined function with. This function is, for example, a function in which the absolute value of the line-of-sight change index value changes depending on the place of viewing, and the absolute value of the line-of-sight change index value increases as the movement of the line of sight increases and the frequency of the movement of the line of sight increases. It is possible.
  • Analysis of pulse changes is performed, for example, as follows. That is, for each frame image, the face area is specified from the frame image. Then, using a trained image analysis model that captures the numerical value of the face color information (G in RGB), the change in the G color on the face surface is analyzed. By arranging the results along the time axis, a waveform showing the change in color information is formed, and the pulse is specified from this waveform. When a person is nervous, the pulse becomes faster, and when he / she feels calm, the pulse becomes slower.
  • the biological reaction analysis unit 12 calculates a pulse change index value according to the analysis result of the pulse change. For example, the biological reaction analysis unit 12 calculates the moving average or the moving variance of the pulse values specified for each frame as the pulse change index value.
  • Analysis of changes in facial movement is performed, for example, as follows. That is, the biological reaction analysis unit 12 identifies a face region from the frame image for each frame image and analyzes the orientation of the face. Then, the biological reaction analysis unit 12 calculates a face orientation change index value according to the analysis result of the face orientation change. For example, the biological reaction analysis unit 12 calculates the difference in the orientation of the face from the true state for each frame image by roll pitch yaw, and determines the moving average or the moving variance of the difference between a plurality of frames as the face orientation change index value. Calculated as.
  • the biological reaction analysis unit 12 may analyze where the user is looking. For example, it analyzes whether the speaker's face being displayed, the shared material being displayed, or the outside of the screen is being viewed. In addition, it may be possible to analyze whether the movement of the face is large or small, and whether the movement is frequent or infrequent. The movement of the face and the movement of the line of sight may be combined and analyzed. For example, it may be possible to analyze whether the speaker's face being displayed is viewed straight, whether the speaker is viewed with an upper eye or a lower eye, or whether the speaker is viewed from an angle. The biological reaction analysis unit 12 calculates a face orientation change index value according to the analysis result of the face orientation change.
  • the biological reaction analysis unit 12 determines the viewing location (speaker's face, shared materials, outside the screen), the direction in which the location is viewed, the magnitude of facial movement, and the frequency of facial movement. Is used as an explanatory variable, and a predetermined function with the face orientation change index value as the objective variable is used to calculate the face orientation change index value.
  • This function for example, changes the absolute value of the face orientation change index value depending on the place of viewing and the direction in which the person is looking. It is possible to make the function so that the absolute value of the value becomes large.
  • the content of the statement is analyzed as follows, for example. That is, the biological reaction analysis unit 12 converts the voice into a character string by performing a known voice recognition process on the voice for a specified time (for example, a time of about 30 to 150 seconds), and morphologically analyzes the character string. By doing so, words unnecessary for expressing conversation such as particles and acronyms are removed. Then, the remaining words are vectorized by the TF-IDF (Term Frequency-Inverse Document Frequency) method, etc., and based on the characteristics of the vector, it is analyzed whether a positive emotional change is occurring or a negative emotional change is occurring. , Calculate the remark content index value according to the analysis result.
  • TF-IDF Term Frequency-Inverse Document Frequency
  • what kind of remark content is used by using a database or the like that stores information relating the vector feature amount and the remark content type based on the vector feature calculated according to the remark content. To estimate. Then, it is possible to calculate the statement content index value by using a predetermined function using the estimation result as an explanatory variable and the statement content index value as the objective variable.
  • the biological reaction analysis unit 12 matches the words extracted from the content of remarks within the specified time with a dictionary (definition of whether each word is positive or negative), and the number of appearances of positive words and negative words. Count the number of appearances. Then, the biological reaction analysis unit 12 calculates the remark content index value by using a predetermined function using each count value as an explanatory variable and the remark content index value as the objective variable.
  • Voice quality analysis is performed as follows, for example. That is, the biological reaction analysis unit 12 identifies the acoustic characteristics of the voice by performing a known voice analysis process on the voice for a specified time (for example, a time of about 30 to 150 seconds). Then, the voice quality change index value is calculated based on the value representing the acoustic feature. For example, the biological reaction analysis unit 12 calculates MFCC (mel frequency cepstrum coefficient) as an acoustic feature of speech, and calculates a moving average or a moving variance for each predetermined time interval of the MFCC as a voice quality change index value.
  • MFCC is an example and is not limited to this.
  • the biological reaction analysis unit 12 determines whether a positive voice quality change is occurring, a negative voice quality change is occurring, and how loud the voice quality change is occurring, based on the acoustic characteristics of the voice. It may be analyzed and the voice quality change index value according to the analysis result may be calculated. For example, similar to the analysis of facial expressions, it analyzes which emotional element of neutrality / calmness / joy / surprise / sadness / anger / fear / disgust corresponds to the voice according to a voice analysis model machine-learned in advance. do. Then, based on the analysis result, whether or not the emotional change occurs at a predetermined time interval, and if the emotional change occurs, whether it is a positive emotional change or a negative emotional change, and how large it is. It analyzes whether or not the emotional change is occurring, and calculates the voice quality change index value according to the analysis result.
  • the biological reaction analysis unit 12 uses at least one of the facial expression change index value, the line-of-sight change index value, the pulse change index value, the face orientation change index value, the speech content index value, and the voice quality change index value calculated as described above.
  • the biological reaction index value is calculated.
  • the biological reaction index value is calculated by weighting the facial expression change index value, the line-of-sight change index value, the pulse change index value, the face orientation change index value, the speech content index value, and the voice quality change index value.
  • the emotion evaluation unit 13 evaluates the degree of emotion of the subject according to the evaluation criteria leveled among the plurality of subjects based on the change in the biological reaction analyzed for the subject by the biological reaction analysis unit 12. For example, the emotion evaluation unit 13 has an emotional response based on an evaluation standard leveled among a plurality of subjects based on the change in the biological reaction (biological reaction index value) analyzed for the subject by the biological reaction analysis unit 12. Calculate the absolute value.
  • the emotional response absolute value calculated by the emotional evaluation unit 13 is, for example, a value obtained by adjusting the biological reaction index value calculated by the biological reaction analysis unit 12 according to the likelihood of the same emotion occurring by the subject.
  • the emotion evaluation unit 13 calculates the absolute emotional response value by multiplying the biological reaction index value calculated by the biological reaction analysis unit 12 by a weight value according to the frequency of causing the same emotion.
  • the emotion evaluation unit 13 has a facial expression change index value, a line-of-sight change index value, a pulse change index value, a face orientation change index value, a speech content index value, and a voice quality change index value calculated by the biological reaction analysis unit 12. Emotions by multiplying each of the above (hereinafter, may be abbreviated as each index value) or at least one index value in each index value by a weight value according to the frequency of causing the same emotion.
  • the absolute reaction value may be calculated.
  • the index value any of the biological reaction index value, each index value, and at least one index value used for calculating the absolute emotional response value is referred to as a calculation target index value.
  • the emotion evaluation unit 13 calculates the absolute emotional response value according to a function such that the weight value becomes smaller as the same emotion is more likely to occur, and the weight value becomes larger as the same emotion is less likely to occur.
  • the frequency of causing the same emotion counts, for example, the number of times that approximately the same value occurs based on a plurality of calculated index values calculated for each predetermined time interval (for example, every 500 milliseconds) during an online session. It can be obtained by doing.
  • the term "almost the same value” as used herein means not only a value having exactly the same value but also a value having a predetermined width that allows a predetermined difference to be regarded as the same value. It should be noted that not only the calculated target index value calculated in one online session but also the calculated target index value calculated in a plurality of online sessions may be used to determine the frequency of causing the same emotion.
  • the emotion evaluation unit 13 has, for example, information representing each emotion (division of a calculation target index value having a predetermined width) and a weight value set according to the frequency of generating the same emotion for each emotion.
  • the table information associated with and is stored in advance. Then, the emotion evaluation unit 13 extracts the weight value corresponding to the analysis result by collating the analysis result by the biological reaction analysis unit 12 with the table information, and multiplies the weight value by the calculation target index value. Calculate the absolute emotional response value.
  • a function designed to calculate the weight value based on the information representing each emotion may be used.
  • the degree of emotions considering the susceptibility to different emotions for each subject is evaluated by using the emotional response absolute value calculated based on the frequency of causing the same emotions. Therefore, it is possible to evaluate the degree of emotion in the true sense of the subject, and it is possible to objectively compare the degree of emotion between different subjects.
  • the processing of the biological reaction analysis unit 12 and the emotion evaluation unit 13 described above may be performed in real time when the moving image acquisition unit 11 acquires the moving images of a plurality of subjects, or the moving image storage unit 101. It may be performed after the fact by using the moving image stored in.
  • the emotion evaluation unit 13 is a degree of emotion based on the magnitude of the difference in the current biological reaction to the biological reaction in normal times, and the emotion is adjusted according to the likelihood of the same emotion being generated by the subject.
  • the degree may be evaluated.
  • the biological reaction in normal times can be, for example, a major biological reaction analyzed based on the past biological reaction distribution of the same subject.
  • the emotion evaluation unit 13 uses the calculated index value calculated by the biological reaction analysis unit 12 to determine the magnitude of the difference in the current biological reaction to the biological reaction in normal times and the susceptibility to the same emotion by the subject.
  • the absolute value of emotional response is calculated by adjusting accordingly.
  • the absolute emotional response value calculated in this way is a value representing the degree of emotion based on the magnitude of the difference in the current biological response to the biological response in normal times, and the subject is likely to generate the same emotion or occurs. It is a value adjusted according to the degree of difficulty.
  • the method of determining the frequency of causing the same emotion described in the above embodiment is an example, and the present invention is not limited to this.
  • the same emotion based on the emotional element (neutral, calm, joy, surprise, sadness, anger, fear, disgust) analyzed by the biological reaction analysis unit 12 at predetermined time intervals during the online session. By counting the number of times the element occurs, the frequency with which the same emotion is generated may be determined.
  • joy and surprise are defined as "pleasant” emotions
  • sadness, anger, fear and disgust are defined as “unpleasant” emotions, and every predetermined time interval during an online session.
  • the frequency of generating "pleasant” emotions and the frequency of generating "unpleasant” emotions may be determined.
  • the emotion evaluation unit 13 is, for example, information representing each emotion (each emotion element). And the table information corresponding to the weight value set according to the frequency of causing the same emotion for each emotion is stored in advance. Then, the emotion evaluation unit 13 extracts the weight value corresponding to the analysis result by collating the analysis result by the biological reaction analysis unit 12 with the table information, and multiplies the weight value by the score corresponding to the emotion element. By doing so, the calculation target index value is calculated, and further, the emotional response absolute value is calculated from the calculation target index value.
  • a function designed to calculate the weight value based on the information representing each emotion may be used.
  • the emotional evaluation unit 13 may obtain, for example, information representing each emotion (comfort / discomfort classification). ) And the weight value set according to the frequency of causing the same emotion for each emotion are stored in advance. Then, the emotion evaluation unit 13 extracts the weight value corresponding to the analysis result by collating the analysis result by the biological reaction analysis unit 12 with the table information, and the emotion element whose weight value is defined as pleasant or unpleasant.
  • the calculation target index value is calculated by multiplying the score corresponding to, and the emotional response absolute value is calculated from the calculation target index value.
  • a function designed to calculate the weight value based on the information representing each emotion may be used.
  • the frequency of generating the same emotion is used as a measure for expressing the susceptibility to the same emotion
  • the present invention is not limited to this.
  • the nature or personality of the subject may be used in place of or in addition to the frequency with which the same emotions occur.
  • the target audience for emotion analysis is not limited to participants in online sessions. That is, any person who is in an environment where a moving image can be acquired can analyze emotions as a target person of the present embodiment.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Strategic Management (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Human Resources & Organizations (AREA)
  • Operations Research (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Data Mining & Analysis (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

The present invention comprises: a moving image acquisition unit 11 which acquires a moving image obtained for a subject; a vital reaction analysis unit 12 which analyzes, on the basis of the moving image acquired by the moving image acquisition unit 11, a change in vital reaction generated due to a change in emotion of the subject; and an emotion evaluation unit 13 which evaluates the degree of emotion adjusted according to how easily the same emotion occurs in the subject on the basis of the change in vital reaction for the subject analyzed by the vital reaction analysis unit 12, wherein, the change in vital reaction for the subject is analyzed on the basis of the moving image obtained for the subject, an emotion reaction absolute value, which is based on an evaluation criterion leveled between a plurality of subjects according to how easily the same emotion occurs in the subject, is calculated on the basis of the analyzed change in vital reaction, and thus the degree of emotion is enabled to be evaluated in a true sense for the subject, and the degrees of emotion between different subjects can be objectively contrasted.

Description

感情解析システムおよび感情解析装置Emotion analysis system and emotion analysis device
 本発明は、感情解析システムおよび感情解析装置に関し、特にオンラインセッションに参加しているユーザの感情を解析するシステムおよび装置に関するものである。 The present invention relates to an emotion analysis system and an emotion analysis device, and more particularly to a system and an apparatus for analyzing emotions of a user participating in an online session.
 従来、対象者の平常時(無表情時)の表情と現在の表情とを比較して、対象者の感情の度合いを判定するようにした技術が知られている(例えば、特許文献1~4参照)。また、音声および画像から個性を考慮した感情を認識する技術も知られている(例えば、特許文献5参照)。 Conventionally, there is known a technique of comparing a normal (expressionless) facial expression of a subject with a current facial expression to determine the degree of emotion of the subject (for example, Patent Documents 1 to 4). reference). Further, a technique for recognizing emotions in consideration of individuality from voice and images is also known (see, for example, Patent Document 5).
 特許文献4には、オペレータの平静時の音声に基づいて生成された基準パラメータをあらかじめ格納しておき、音声取得部により取得されたオペレータの音声に基づいて生成される感情パラメータと比較することによって感情度合を特定することが開示されている。また、特許文献4には、基準パラメータと感情パラメータとの差異の出方については個人差があることに鑑みて、オペレータ毎に感情パラメータの平均値を求め、基準パラメータと平均値との差と、基準パラメータと感情パラメータとの差との割合に基づいて、オペレータ毎に相対的に感情度合を特定することも開示されている。 In Patent Document 4, a reference parameter generated based on the operator's calm voice is stored in advance, and the emotion parameter is compared with the emotion parameter generated based on the operator's voice acquired by the voice acquisition unit. It is disclosed to specify the degree of emotion. Further, in Patent Document 4, in view of the fact that there are individual differences in the appearance of the difference between the reference parameter and the emotion parameter, the average value of the emotion parameter is obtained for each operator, and the difference between the reference parameter and the average value is obtained. It is also disclosed that the emotional degree is relatively specified for each operator based on the ratio of the difference between the reference parameter and the emotional parameter.
特開2011-154665号公報Japanese Unexamined Patent Publication No. 2011-154665 特開2012-8949号公報Japanese Unexamined Patent Publication No. 2012-8949 特開2013-300号公報Japanese Unexamined Patent Publication No. 2013-300 特開2015-141428号公報Japanese Unexamined Patent Publication No. 2015-141428 特開2001-83984号公報Japanese Unexamined Patent Publication No. 2001-83984
 しかしながら、上記特許文献1~3に記載の技術では、平常時の表情から現在の表情がどの程度変化しているかによって感情の度合いを評価しているだけである。そのため、この評価情報は、異なる対象者間で感情の度合いを客観的に対比するための情報としては用いることができないという問題があった。 However, in the techniques described in Patent Documents 1 to 3, the degree of emotion is only evaluated by how much the current facial expression changes from the normal facial expression. Therefore, there is a problem that this evaluation information cannot be used as information for objectively comparing the degree of emotion between different subjects.
 例えば、対象者Aが自身の平常時の表情との違いから現在の感情Xの度合いがレベル3と判定され、対象者Bについても同様に現在の感情Xの度合いがレベル3と判定されたとする。しかしながら、感情Xを生起しやすい対象者Aのレベル3と、感情Xを生起しにくい対象者Bのレベル3とでは感情Xの意味合い(真の意味での感情の度合い)が違う。上記特許文献1~3に記載の技術では、このような真の意味での感情の度合いを評価することができないため、異なる対象者間で感情の度合いを客観的に対比することができない。 For example, it is assumed that the subject A is determined to have a current emotion X level of 3 based on the difference from his / her normal facial expression, and the subject B is also determined to have a current emotion X level of 3. .. However, the meaning of emotion X (the degree of emotion in the true sense) is different between the level 3 of the subject A who easily causes emotion X and the level 3 of the subject B who does not easily generate emotion X. Since the techniques described in Patent Documents 1 to 3 cannot evaluate the degree of emotion in the true sense of the word, it is not possible to objectively compare the degree of emotion between different subjects.
 これに対し、上記特許文献4に記載の技術では、感情の出る大きさに個人差があることに鑑みて、オペレータ毎に感情パラメータの平均値を求め、この平均値を利用して、オペレータ毎に相対的に感情度合を特定している。しかしながら、この特許文献4に記載の技術においても、オペレータ毎の感情の生起しやすさを考慮した感情度合の特定は行っておらず、上述の問題を解消することはできない。 On the other hand, in the technique described in Patent Document 4, in view of individual differences in the magnitude of emotions, the average value of emotion parameters is obtained for each operator, and the average value is used for each operator. The degree of emotion is relatively specified. However, even in the technique described in Patent Document 4, the degree of emotion is not specified in consideration of the ease of occurrence of emotion for each operator, and the above-mentioned problem cannot be solved.
 本発明は、対象者に関する真の意味での感情の度合いを評価可能とすることで、異なる対象者間で感情の度合いを客観的に対比することができるようにすることを目的とする。 It is an object of the present invention to be able to objectively compare the degree of emotion between different subjects by making it possible to evaluate the degree of true emotions regarding the subject.
 上記した課題を解決するために、本発明の感情解析システムでは、対象者について得られる動画像に基づいて、対象者について生体反応の変化を解析し、解析された生体反応の変化に基づいて、複数の対象者間で平準化された評価基準に従って、対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価するようにしている。 In order to solve the above-mentioned problems, in the emotion analysis system of the present invention, changes in the biological reaction of the subject are analyzed based on the moving images obtained for the subject, and based on the analyzed changes in the biological reaction, the changes in the biological reaction are analyzed. According to the evaluation criteria leveled among a plurality of subjects, the degree of emotions adjusted according to the likelihood of the same emotions being generated by the subjects is evaluated.
 上記のように構成した本発明によれば、対象者ごとに異なる感情の生起しやすさを考慮した感情の度合の評価が行われるので、対象者に関する真の意味での感情の度合いを評価することが可能となり、異なる対象者間で感情の度合いを客観的に対比することができる。 According to the present invention configured as described above, the degree of emotion is evaluated in consideration of the susceptibility to generate different emotions for each subject, and therefore the degree of true emotions regarding the subject is evaluated. This makes it possible to objectively compare the degree of emotion between different subjects.
本実施形態による感情解析システムの全体構成例を示す図である。It is a figure which shows the whole structure example of the emotion analysis system by this embodiment. 本実施形態による感情解析装置の機能構成例を示すブロック図である。It is a block diagram which shows the functional composition example of the emotion analysis apparatus by this embodiment.
 以下、本発明の一実施形態を図面に基づいて説明する。図1は、本実施形態による感情解析システムの全体構成例を示す図である。図1に示すように、本実施形態の感情解析システムは、感情解析装置100、複数のユーザ端末200-1,200-2,・・・(以下、特に区別しないときは単にユーザ端末200と記す)およびセッション管理サーバ300を備えて構成される。これらの感情解析装置100、ユーザ端末200およびセッション管理サーバ300は、インターネットや携帯電話網などの通信ネットワーク500を介して接続される。 Hereinafter, an embodiment of the present invention will be described with reference to the drawings. FIG. 1 is a diagram showing an overall configuration example of an emotion analysis system according to the present embodiment. As shown in FIG. 1, the emotion analysis system of the present embodiment is referred to as an emotion analysis device 100, a plurality of user terminals 200 -1 , 200-2 , ... (Hereinafter, unless otherwise specified, it is simply referred to as a user terminal 200. ) And a session management server 300. The emotion analysis device 100, the user terminal 200, and the session management server 300 are connected via a communication network 500 such as the Internet or a mobile phone network.
 本実施形態の感情解析システムは、例えば複数人の参加者でオンラインセッションが行われる環境において、対象者(オンラインセッションの参加者)について得られる動画像に基づいて、対象者について感情の変化に起因して起こる生体反応の変化を解析し、解析された生体反応の変化に基づいて、複数の対象者間で平準化された評価基準に従って対象者の感情の度合いを評価するシステムである。 The emotion analysis system of the present embodiment is caused by a change in emotions of the subject based on a moving image obtained about the subject (participant of the online session), for example, in an environment where an online session is performed by a plurality of participants. It is a system that analyzes the changes in the biological reaction that occur and evaluates the degree of emotion of the subject according to the evaluation criteria leveled among the plurality of subjects based on the analyzed changes in the biological reaction.
 オンラインセッションは、例えばオンライン会議、オンライン授業、オンラインチャットなどであり、複数の場所に設置された複数のユーザ端末200を通信ネットワーク500を介して感情解析装置100およびセッション管理サーバ300に接続し、当該感情解析装置100およびセッション管理サーバ300を通じて複数のユーザ端末200間で動画像をやり取りできるようにしたものである。複数のユーザ端末200には、オンラインセッションで動画像をやり取りするために必要なアプリケーションプログラム(以下、セッションアプリという)がインストールされている。 The online session is, for example, an online conference, an online class, an online chat, or the like, in which a plurality of user terminals 200 installed at a plurality of locations are connected to the emotion analysis device 100 and the session management server 300 via a communication network 500. A moving image can be exchanged between a plurality of user terminals 200 through an emotion analysis device 100 and a session management server 300. An application program (hereinafter referred to as a session application) necessary for exchanging moving images in an online session is installed in a plurality of user terminals 200.
 オンラインセッションで扱う動画像には、ユーザ端末200を使用するユーザ(オンラインセッションの参加者)の顔画像(実際には、顔以外の身体の部位や背景の画像も含まれる)や音声が含まれる。ユーザの顔画像と音声は、ユーザ端末200に備えられた、またはユーザ端末200に接続されたカメラおよびマイクにより取得され、セッション管理サーバ300に送信される。そして、セッション管理サーバ300に送信された各ユーザの顔画像と音声は感情解析装置100によって取得され、感情解析装置100から各ユーザ端末200のセッションアプリに送信される。なお、ユーザ端末200から送信された動画像を感情解析装置100にて取得し、これを感情解析装置100からセッション管理サーバ300に転送するようにしてもよい。あるいは、ユーザ端末200から動画像を感情解析装置100およびセッション管理サーバ300の両方に送信するようにしてもよい。 The moving image handled in the online session includes a face image (actually, an image of a body part other than the face and a background image) and a voice of a user (participant in the online session) who uses the user terminal 200. .. The user's face image and voice are acquired by a camera and a microphone provided in the user terminal 200 or connected to the user terminal 200, and transmitted to the session management server 300. Then, the face image and voice of each user transmitted to the session management server 300 are acquired by the emotion analysis device 100, and are transmitted from the emotion analysis device 100 to the session application of each user terminal 200. The moving image transmitted from the user terminal 200 may be acquired by the emotion analysis device 100 and transferred from the emotion analysis device 100 to the session management server 300. Alternatively, the moving image may be transmitted from the user terminal 200 to both the emotion analysis device 100 and the session management server 300.
 また、動画像には、複数のユーザが共有して閲覧する資料などの画像も含まれる。ユーザが閲覧する資料画像は、何れかのユーザ端末200からセッション管理サーバ300に送信される。そして、セッション管理サーバ300に送信された資料画像は感情解析装置100によって取得され、感情解析装置100から各ユーザ端末200のセッションアプリに送信される。 In addition, moving images include images such as materials shared and viewed by multiple users. The material image to be viewed by the user is transmitted from any user terminal 200 to the session management server 300. Then, the material image transmitted to the session management server 300 is acquired by the emotion analysis device 100, and is transmitted from the emotion analysis device 100 to the session application of each user terminal 200.
 以上の動作により、複数のユーザ端末200のそれぞれにおいて、複数のユーザの顔画像または資料画像がディスプレイに表示され、複数のユーザの音声がスピーカから出力される。ここで、ユーザ端末200にインストールされているセッションアプリの機能により、ディスプレイの画面上に顔画像と資料画像とを切り替えて何れか一方のみを表示させたり、表示領域を分けて顔画像と資料画像とを同時に表示させたりすることが可能である。また、複数人のユーザのうち1人の画像を全画面表示させたり、一部または全部のユーザの画像を小画面に分割して表示させたりすることが可能である。 By the above operation, the face image or the document image of the plurality of users is displayed on the display in each of the plurality of user terminals 200, and the voices of the plurality of users are output from the speaker. Here, by the function of the session application installed in the user terminal 200, the face image and the material image can be switched to display only one of them on the display screen, or the display area can be divided into the face image and the material image. Can be displayed at the same time. Further, it is possible to display the image of one of a plurality of users on the full screen, or to display the image of a part or all of the users on a small screen.
 また、ユーザ端末200にインストールされているセッションアプリの機能により、カメラのオン/オフを切り替えたり、マイクのオン/オフを切り替えたりすることも可能である。例えば、ユーザ端末200-1においてカメラをオフにした場合、ユーザ端末200-1のカメラにより撮影された顔画像はセッション管理サーバ300および感情解析装置100に送信されるが、感情解析装置100から各ユーザ端末200に送信されない。同様に、ユーザ端末200-1においてマイクをオフにした場合、ユーザ端末200-1のマイクにより集音された音声はセッション管理サーバ300および感情解析装置100に送信されるが、感情解析装置100から各ユーザ端末200に送信されない。 It is also possible to switch the camera on / off and the microphone on / off by the function of the session application installed in the user terminal 200. For example, when the camera is turned off in the user terminal 200-1 , the face image taken by the camera of the user terminal 200-1 is transmitted to the session management server 300 and the emotion analysis device 100, but each of them is transmitted from the emotion analysis device 100. It is not transmitted to the user terminal 200. Similarly, when the microphone is turned off in the user terminal 200-1 , the sound collected by the microphone of the user terminal 200-1 is transmitted to the session management server 300 and the emotion analysis device 100, but is transmitted from the emotion analysis device 100. It is not transmitted to each user terminal 200.
 図2は、本実施形態による感情解析装置100の機能構成例を示すブロック図である。図2に示すように、本実施形態の感情解析装置100は、機能構成として、動画像取得部11、生体反応解析部12および感情評価部13を備えている。また、本実施形態の感情解析装置100は、記憶媒体として、動画像記憶部101を備えている。 FIG. 2 is a block diagram showing a functional configuration example of the emotion analysis device 100 according to the present embodiment. As shown in FIG. 2, the emotion analysis device 100 of the present embodiment includes a moving image acquisition unit 11, a biological reaction analysis unit 12, and an emotion evaluation unit 13 as functional configurations. Further, the emotion analysis device 100 of the present embodiment includes a moving image storage unit 101 as a storage medium.
 上記各機能ブロック11~13は、ハードウェア、DSP(Digital Signal Processor)、ソフトウェアの何れによっても構成することが可能である。例えばソフトウェアによって構成する場合、上記各機能ブロック11~13は、実際にはコンピュータのCPU、RAM、ROMなどを備えて構成され、RAMやROM、ハードディスクまたは半導体メモリ等の記録媒体に記憶されたプログラムが動作することによって実現される。 Each of the above functional blocks 11 to 13 can be configured by any of hardware, DSP (Digital Signal Processor), and software. For example, when configured by software, each of the above functional blocks 11 to 13 is actually configured to include a computer CPU, RAM, ROM, etc., and is a program stored in a recording medium such as RAM, ROM, a hard disk, or a semiconductor memory. Is realized by the operation of.
 動画像取得部11は、オンラインセッション中に各ユーザ端末200から送信される動画像(顔画像、音声、資料画像)をセッション管理サーバ300から取得する。動画像取得部11は、各ユーザ端末200から取得した動画像を、各ユーザを識別可能な情報(例えば、ユーザID)に関連付けて動画像記憶部101に記憶させる。 The moving image acquisition unit 11 acquires moving images (face image, voice, material image) transmitted from each user terminal 200 during the online session from the session management server 300. The moving image acquisition unit 11 stores the moving image acquired from each user terminal 200 in the moving image storage unit 101 in association with information that can identify each user (for example, a user ID).
 セッション管理サーバ300から取得する顔画像は、各ユーザ端末200の画面上に表示されるように設定されているものか否か(カメラがオンに設定されているかオフに設定されているか)は問わない。すなわち、動画像取得部11は、各ユーザ端末200のディスプレイに表示中の顔画像および非表示中の顔画像を含めて、顔画像をセッション管理サーバ300から取得する。また、セッション管理サーバ300から取得する音声は、各ユーザ端末200のスピーカから出力されるように設定されているものか否か(マイクがオンに設定されているかオフに設定されているか)は問わない。すなわち、動画像取得部11は、各ユーザ端末200のスピーカから出力中の音声および非出力中の音声を含めて、音声をセッション管理サーバ300から取得する。 It does not matter whether the face image acquired from the session management server 300 is set to be displayed on the screen of each user terminal 200 (whether the camera is set to on or off). No. That is, the moving image acquisition unit 11 acquires the face image from the session management server 300, including the face image displayed on the display of each user terminal 200 and the face image not being displayed. Further, it does not matter whether the sound acquired from the session management server 300 is set to be output from the speaker of each user terminal 200 (whether the microphone is set to on or off). No. That is, the moving image acquisition unit 11 acquires audio from the session management server 300, including audio being output from the speaker of each user terminal 200 and audio being non-output.
 生体反応解析部12は、動画像取得部11により取得され動画像記憶部101に記憶された動画像(ユーザ端末200の画面上に表示中の顔画像か否か、ユーザ端末200のスピーカから出力中の音声か否かは問わない)に基づいて、複数人の参加者のそれぞれについて、感情の変化に起因して起こる生体反応の変化を解析する。本実施形態において生体反応解析部12は、動画像取得部11により取得された動画像を顔画像のセット(フレーム画像の集まり)と音声とに分離し、それぞれから生体反応の変化を解析する。 The biological reaction analysis unit 12 outputs a moving image (whether or not it is a face image displayed on the screen of the user terminal 200, whether or not it is a face image displayed on the screen of the user terminal 200, from the speaker of the user terminal 200) acquired by the moving image acquisition unit 11 and stored in the moving image storage unit 101. Based on (whether or not it is the voice inside), we analyze the changes in biological reactions caused by changes in emotions for each of the multiple participants. In the present embodiment, the biological reaction analysis unit 12 separates the moving image acquired by the moving image acquisition unit 11 into a set of facial images (a collection of frame images) and voice, and analyzes changes in the biological reaction from each.
 例えば、生体反応解析部12は、動画像取得部11により取得された動画像から分離したフレーム画像を用いてユーザの顔画像を解析することにより、表情、目線、脈拍、顔の動きの少なくとも1つに関する生体反応の変化を解析する。また、生体反応解析部12は、動画像取得部11により取得された動画像から分離した音声を解析することにより、ユーザの発言内容、声質の少なくとも1つに関する生体反応の変化を解析する。 For example, the biological reaction analysis unit 12 analyzes the user's face image using the frame image separated from the moving image acquired by the moving image acquisition unit 11, and thereby at least one of the facial expression, the line of sight, the pulse, and the movement of the face. Analyze changes in biological reactions related to one. In addition, the biological reaction analysis unit 12 analyzes changes in the biological reaction regarding at least one of the user's speech content and voice quality by analyzing the voice separated from the moving image acquired by the moving image acquisition unit 11.
 人は感情が変化すると、それが表情、目線、脈拍、顔の動き、発言内容、声質などの生体反応の変化となって現れる。本実施形態では、ユーザの生体反応の変化を解析することを通じて、ユーザの感情の変化を解析する。本実施形態において生体反応解析部12は、生体反応の変化を所定の基準に従って数値化することにより、生体反応の変化の内容を反映させた生体反応指標値を算出する。 When a person's emotions change, it appears as changes in biological reactions such as facial expressions, eyes, pulse, facial movements, speech content, and voice quality. In this embodiment, changes in the user's emotions are analyzed by analyzing changes in the user's biological reaction. In the present embodiment, the biological reaction analysis unit 12 calculates a biological reaction index value that reflects the content of the change in the biological reaction by quantifying the change in the biological reaction according to a predetermined standard.
 表情の変化の解析は、例えば以下のようにして行う。すなわち、生体反応解析部12は、フレーム画像ごとに、フレーム画像の中から顔の領域を特定し、事前に機械学習させた画像解析モデルに従って、顔の表情がどの表情要素に該当するかを解析する。そして、その解析結果に基づいて、連続するフレーム画像間で表情変化が起きているか否か、表情変化が起きている場合はそれがポジティブな表情変化かネガティブな表情変化か、およびどの程度の大きさの表情変化が起きているかを解析し、その解析結果に応じた表情変化指標値を算出する。 Analysis of changes in facial expressions is performed, for example, as follows. That is, the biological reaction analysis unit 12 identifies a facial region from the frame image for each frame image, and analyzes which facial expression element the facial expression corresponds to according to an image analysis model machine-learned in advance. do. Then, based on the analysis result, whether or not a facial expression change occurs between consecutive frame images, and if a facial expression change occurs, whether it is a positive facial expression change or a negative facial expression change, and how large it is. It analyzes whether the facial expression change is occurring, and calculates the facial expression change index value according to the analysis result.
 顔の表情要素は、例えば、中立(neutral)/落ち着き(calm)/喜び(happy)/驚き(surprised)/悲しみ(sad)/怒り(angry)/恐れ(fearful)/嫌悪感(disgust)などである。このうち、喜びおよび驚きはポジティブな表情要素であり、悲しみ、怒り、恐れ、嫌悪感はネガティブな表情要素である。 Facial expression elements are, for example, neutral / calm / happy / surprised / sad / angry / fearful / disgust. be. Of these, joy and surprise are positive facial expression elements, and sadness, anger, fear, and disgust are negative facial expression elements.
 生体反応解析部12は、各フレーム画像における顔の表情について、複数の表情要素ごとに合計100となるスコアを算出する。例えば、中立=10、落ち着き=10、喜び=30、驚き=20、悲しみ=10、怒り=10、恐れ=5、嫌悪感=5といったように、各表情要素に該当する可能性の高さに応じたスコアを表情要素ごとに算出する。そして、例えばスコアが最大の表情要素を、そのフレーム画像における顔の表情として決定する。以下では、フレーム画像ごとに決定される顔の表情のスコア(複数の表情要素について算出されたスコアのうち最大のスコア)を「表情スコア」という。 The biological reaction analysis unit 12 calculates a total score of 100 for each of the plurality of facial expression elements for the facial expression in each frame image. For example, neutrality = 10, calmness = 10, joy = 30, surprise = 20, sadness = 10, anger = 10, fear = 5, disgust = 5, and so on. The corresponding score is calculated for each facial expression element. Then, for example, the facial expression element having the maximum score is determined as the facial expression in the frame image. In the following, the facial expression score (the maximum score calculated for a plurality of facial expression elements) determined for each frame image is referred to as a “facial expression score”.
 生体反応解析部12は、このようにしてフレーム画像ごとに決定される表情要素およびフレーム画像ごとに算出される表情スコアの少なくとも一方が前フレームから変化したか否かによって、連続するフレーム画像間で表情変化が起きているか否かを判定する。ここで、生体反応解析部12は、最大スコアの表情要素に変化がない場合に、前フレームからのスコア変化量が所定の閾値以上の場合に表情変化が起きていると判定するようにしてもよい。表情変化の大きさは、表情スコアの前フレームからの変化量によって判定することが可能である。 The biological reaction analysis unit 12 determines between consecutive frame images depending on whether at least one of the facial expression element determined for each frame image and the facial expression score calculated for each frame image has changed from the previous frame. Determine if the facial expression has changed. Here, even if the biological reaction analysis unit 12 determines that the facial expression change has occurred when the score change amount from the previous frame is equal to or more than a predetermined threshold value when there is no change in the facial expression element of the maximum score. good. The magnitude of the facial expression change can be determined by the amount of change from the previous frame of the facial expression score.
 また、生体反応解析部12は、ポジティブな表情の表情スコアが前フレームから増加した場合、および、前フレームのネガティブな表情から現フレームのポジティブな表情に変化した場合に、ポジティブな表情変化が起きていると判定する。一方、生体反応解析部12は、ネガティブな表情の表情スコアが前フレームから増加した場合、および、前フレームのポジティブな表情から現フレームのネガティブな表情に変化した場合に、ネガティブな表情変化が起きていると判定する。 In addition, the biological reaction analysis unit 12 causes a positive facial expression change when the facial expression score of the positive facial expression increases from the previous frame and when the negative facial expression of the previous frame changes to the positive facial expression of the current frame. It is determined that it is. On the other hand, the biological reaction analysis unit 12 causes a negative facial expression change when the facial expression score of the negative facial expression increases from the previous frame and when the positive facial expression of the previous frame changes to the negative facial expression of the current frame. It is determined that it is.
 さらに、生体反応解析部12は、表情変化の方向(ポジティブ→ポジティブ、ポジティブ→ネガティブ、ネガティブ→ポジティブ、ネガティブ→ネガティブ)と、表情変化の大きさとを説明変数とし、表情変化指標値を目的変数とする所定の関数を用いて、表情変化指標値を算出する。この関数は、例えば、表情が逆転する場合(ポジティブ→ネガティブ、ネガティブ→ポジティブ)には逆転しない場合に比べて表情変化指標値の絶対値が大きくなり、かつ、表情変化の程度が大きいほど表情変化指標値の絶対値が大きくなるような関数で、表情がポジティブな方向に変化する場合(ポジティブ→ポジティブ、ネガティブ→ポジティブ)は正の値となり、表情がネガティブな方向に変化する場合(ポジティブ→ネガティブ、ネガティブ→ネガティブ)は負の値となるような関数とすることが可能である。 Further, the biological reaction analysis unit 12 uses the direction of facial expression change (positive → positive, positive → negative, negative → positive, negative → negative) and the magnitude of facial expression change as explanatory variables, and the facial expression change index value as the objective variable. The facial expression change index value is calculated using a predetermined function. In this function, for example, when the facial expression is reversed (positive → negative, negative → positive), the absolute value of the facial expression change index value is larger than when it is not reversed, and the greater the degree of facial expression change, the larger the facial expression change. A function that increases the absolute value of the index value, and when the facial expression changes in the positive direction (positive → positive, negative → positive), it becomes a positive value, and when the facial expression changes in the negative direction (positive → negative). , Negative → Negative) can be a function that has a negative value.
 ここでは、連続するフレーム画像間での表情変化を解析する例について説明したが、所定の時間区間ごと(例えば、500ミリ秒ごと)に表情変化を解析するようにしてもよい。これは、以下に述べる目線の変化の解析、脈拍の変化の解析、顔の動きの変化の解析についても同様である。 Here, an example of analyzing a facial expression change between consecutive frame images has been described, but the facial expression change may be analyzed every predetermined time interval (for example, every 500 milliseconds). This also applies to the analysis of the change in the line of sight, the analysis of the change in the pulse, and the analysis of the change in the movement of the face described below.
 目線の変化の解析は、例えば以下のようにして行う。すなわち、生体反応解析部12は、フレーム画像ごとに、フレーム画像の中から目の領域を特定し、両目の向き(目線)を解析する。そして、生体反応解析部12は、目線の変化の解析結果に応じた目線変化指標値を算出する。例えば、生体反応解析部12は、フレーム画像ごとに正面からの視線の角度を算出し、当該角度の複数フレーム間の移動平均または移動分散を目線変化指標値として算出する。 Analysis of changes in the line of sight is performed, for example, as follows. That is, the biological reaction analysis unit 12 identifies the eye region from the frame image for each frame image and analyzes the direction (line of sight) of both eyes. Then, the biological reaction analysis unit 12 calculates the line-of-sight change index value according to the analysis result of the line-of-sight change. For example, the biological reaction analysis unit 12 calculates the angle of the line of sight from the front for each frame image, and calculates the moving average or the moving variance between a plurality of frames of the angle as the line-of-sight change index value.
 なお、生体反応解析部12は、ユーザがどこを見ているかを解析するようにしてもよい。目線の変化はユーザの集中度にも関連する。例えば、表示中の話者の顔を見ているか、表示中の共有資料を見ているか、画面の外を見ているかなどを解析する。また、目線の動きが大きいか小さいか、動きの頻度が多いか少ないかなどを解析するようにしてもよい。そして、生体反応解析部12は、目線の変化の解析結果に応じた目線変化指標値を算出する。 The biological reaction analysis unit 12 may analyze where the user is looking. The change in the line of sight is also related to the degree of concentration of the user. For example, it analyzes whether the speaker's face being displayed, the shared material being displayed, or the outside of the screen is being viewed. In addition, it may be possible to analyze whether the movement of the line of sight is large or small, and whether the movement is frequent or infrequent. Then, the biological reaction analysis unit 12 calculates the line-of-sight change index value according to the analysis result of the line-of-sight change.
 例えば、生体反応解析部12は、見ている場所(話者の顔、共有資料、画面の外)と、目線の動きの大きさと、目線の動きの頻度とを説明変数とし、目線変化指標値を目的変数とする所定の関数を用いて、目線変化指標値を算出する。この関数は、例えば、見ている場所によって目線変化指標値の絶対値が変わり、目線の動きが大きいほど、また目線の動きの頻度が大きいほど目線変化指標値の絶対値が大きくなるような関数とすることが可能である。 For example, the biological reaction analysis unit 12 uses the place of viewing (speaker's face, shared material, outside the screen), the magnitude of eye movement, and the frequency of eye movement as explanatory variables, and the eye change index value. The line-of-sight change index value is calculated using a predetermined function with. This function is, for example, a function in which the absolute value of the line-of-sight change index value changes depending on the place of viewing, and the absolute value of the line-of-sight change index value increases as the movement of the line of sight increases and the frequency of the movement of the line of sight increases. It is possible.
 脈拍の変化の解析は、例えば以下のようにして行う。すなわち、フレーム画像ごとに、フレーム画像の中から顔の領域を特定する。そして、顔の色情報(RGBのG)の数値を捉える学習済みの画像解析モデルを用いて、顔表面のG色の変化を解析する。その結果を時間軸に合わせて並べることによって色情報の変化を表した波形を形成し、この波形から脈拍を特定する。人は緊張すると脈拍が速くなり、気持ちが落ち着くと脈拍が遅くなる。生体反応解析部12は、脈拍の変化の解析結果に応じた脈拍変化指標値を算出する。例えば、生体反応解析部12は、各フレームごとに特定した脈拍値の、複数フレーム間の移動平均または移動分散を脈拍変化指標値として算出する。 Analysis of pulse changes is performed, for example, as follows. That is, for each frame image, the face area is specified from the frame image. Then, using a trained image analysis model that captures the numerical value of the face color information (G in RGB), the change in the G color on the face surface is analyzed. By arranging the results along the time axis, a waveform showing the change in color information is formed, and the pulse is specified from this waveform. When a person is nervous, the pulse becomes faster, and when he / she feels calm, the pulse becomes slower. The biological reaction analysis unit 12 calculates a pulse change index value according to the analysis result of the pulse change. For example, the biological reaction analysis unit 12 calculates the moving average or the moving variance of the pulse values specified for each frame as the pulse change index value.
 顔の動きの変化の解析は、例えば以下のようにして行う。すなわち、生体反応解析部12は、フレーム画像ごとに、フレーム画像の中から顔の領域を特定し、顔の向きを解析する。そして、生体反応解析部12は、顔の向きの変化の解析結果に応じた顔向き変化指標値を算出する。例えば、生体反応解析部12は、フレーム画像ごとに正体時との顔の向きの差分をロール・ピッチ・ヨーで算出し、当該差分の複数フレーム間の移動平均または移動分散を顔向き変化指標値として算出する。 Analysis of changes in facial movement is performed, for example, as follows. That is, the biological reaction analysis unit 12 identifies a face region from the frame image for each frame image and analyzes the orientation of the face. Then, the biological reaction analysis unit 12 calculates a face orientation change index value according to the analysis result of the face orientation change. For example, the biological reaction analysis unit 12 calculates the difference in the orientation of the face from the true state for each frame image by roll pitch yaw, and determines the moving average or the moving variance of the difference between a plurality of frames as the face orientation change index value. Calculated as.
 なお、生体反応解析部12は、ユーザがどこを見ているかを解析するようにしてもよい。例えば、表示中の話者の顔を見ているか、表示中の共有資料を見ているか、画面の外を見ているかなどを解析する。また、顔の動きが大きいか小さいか、動きの頻度が多いか少ないかなどを解析するようにしてもよい。顔の動きと目線の動きとを合わせて解析するようにしてもよい。例えば、表示中の話者の顔をまっすぐ見ているか、上目遣いまたは下目使いに見ているか、斜めから見ているかなどを解析するようにしてもよい。生体反応解析部12は、顔の向きの変化の解析結果に応じた顔向き変化指標値を算出する。 The biological reaction analysis unit 12 may analyze where the user is looking. For example, it analyzes whether the speaker's face being displayed, the shared material being displayed, or the outside of the screen is being viewed. In addition, it may be possible to analyze whether the movement of the face is large or small, and whether the movement is frequent or infrequent. The movement of the face and the movement of the line of sight may be combined and analyzed. For example, it may be possible to analyze whether the speaker's face being displayed is viewed straight, whether the speaker is viewed with an upper eye or a lower eye, or whether the speaker is viewed from an angle. The biological reaction analysis unit 12 calculates a face orientation change index value according to the analysis result of the face orientation change.
 例えば、生体反応解析部12は、見ている場所(話者の顔、共有資料、画面の外)と、その場所を見ている向きと、顔の動きの大きさと、顔の動きの頻度とを説明変数とし、顔向き変化指標値を目的変数とする所定の関数を用いて、顔向き変化指標値を算出する。この関数は、例えば、見ている場所およびその場所を見ている向きによって顔向き変化指標値の絶対値が変わり、顔の動きが大きいほど、また顔の動きの頻度が大きいほど顔向き変化指標値の絶対値が大きくなるような関数とすることが可能である。 For example, the biological reaction analysis unit 12 determines the viewing location (speaker's face, shared materials, outside the screen), the direction in which the location is viewed, the magnitude of facial movement, and the frequency of facial movement. Is used as an explanatory variable, and a predetermined function with the face orientation change index value as the objective variable is used to calculate the face orientation change index value. This function, for example, changes the absolute value of the face orientation change index value depending on the place of viewing and the direction in which the person is looking. It is possible to make the function so that the absolute value of the value becomes large.
 発言内容の解析は、例えば以下のようにして行う。すなわち、生体反応解析部12は、指定した時間(例えば、30~150秒程度の時間)の音声について公知の音声認識処理を行うことによって音声を文字列に変換し、当該文字列を形態素解析することにより、助詞、冠詞などの会話を表す上で不要なワードを取り除く。そして、残ったワードをTF-IDF(Term Frequency - Inverse Document Frequency)法などによりベクトル化し、ベクトルの特徴に基づいて、ポジティブな感情変化が起きているか、ネガティブな感情変化が起きているかを解析し、その解析結果に応じた発言内容指標値を算出する。例えば、発言内容に応じて算出されるベクトルの特徴に基づいて、ベクトルの特徴量と発言内容の種類とを関連付ける情報を格納したデータベース等を利用して、どのような種類の発言内容であるかを推定する。そして、その推定結果を説明変数とし、発言内容指標値を目的変数とする所定の関数を用いて、発言内容指標値を算出するようにすることが可能である。 The content of the statement is analyzed as follows, for example. That is, the biological reaction analysis unit 12 converts the voice into a character string by performing a known voice recognition process on the voice for a specified time (for example, a time of about 30 to 150 seconds), and morphologically analyzes the character string. By doing so, words unnecessary for expressing conversation such as particles and acronyms are removed. Then, the remaining words are vectorized by the TF-IDF (Term Frequency-Inverse Document Frequency) method, etc., and based on the characteristics of the vector, it is analyzed whether a positive emotional change is occurring or a negative emotional change is occurring. , Calculate the remark content index value according to the analysis result. For example, what kind of remark content is used by using a database or the like that stores information relating the vector feature amount and the remark content type based on the vector feature calculated according to the remark content. To estimate. Then, it is possible to calculate the statement content index value by using a predetermined function using the estimation result as an explanatory variable and the statement content index value as the objective variable.
 別の例として、以下のようにしてもよい。すなわち、生体反応解析部12は、指定した時間内の発言内容から抽出したワードを辞書(各ワードがポジティブかネガティブかが定義されたもの)と突き合わせ、ポジティブなワードの出現回数とネガティブなワードの出現回数とをカウントする。そして、生体反応解析部12は、それぞれのカウント値を説明変数とし、発言内容指標値を目的変数とする所定の関数を用いて、発言内容指標値を算出する。 As another example, the following may be performed. That is, the biological reaction analysis unit 12 matches the words extracted from the content of remarks within the specified time with a dictionary (definition of whether each word is positive or negative), and the number of appearances of positive words and negative words. Count the number of appearances. Then, the biological reaction analysis unit 12 calculates the remark content index value by using a predetermined function using each count value as an explanatory variable and the remark content index value as the objective variable.
 声質の解析は、例えば以下のようにして行う。すなわち、生体反応解析部12は、指定した時間(例えば、30~150秒程度の時間)の音声について公知の音声解析処理を行うことによって音声の音響的特徴を特定する。そして、その音響的特徴を表す値に基づいて、声質変化指標値を算出する。例えば、生体反応解析部12は、音声の音響的特徴としてMFCC(メル周波数ケプストラム係数)を算出し、当該MFCCの所定の時間区間ごとの移動平均または移動分散を声質変化指標値として算出する。MFCCは一例であり、これに限定されるものではない。 Voice quality analysis is performed as follows, for example. That is, the biological reaction analysis unit 12 identifies the acoustic characteristics of the voice by performing a known voice analysis process on the voice for a specified time (for example, a time of about 30 to 150 seconds). Then, the voice quality change index value is calculated based on the value representing the acoustic feature. For example, the biological reaction analysis unit 12 calculates MFCC (mel frequency cepstrum coefficient) as an acoustic feature of speech, and calculates a moving average or a moving variance for each predetermined time interval of the MFCC as a voice quality change index value. MFCC is an example and is not limited to this.
 なお、生体反応解析部12は、音声の音響的特徴に基づいて、ポジティブな声質変化が起きているか、ネガティブな声質変化が起きているか、およびどの程度の大きさの声質変化が起きているかを解析し、その解析結果に応じた声質変化指標値を算出するようにしてもよい。例えば、顔の表情の解析と同様に、事前に機械学習させた音声解析モデルに従って、音声が中立/落ち着き/喜び/驚き/悲しみ/怒り/恐れ/嫌悪感のどの感情要素に該当するかを解析する。そして、その解析結果に基づいて、所定の時間区間ごとに感情変化が起きているか否か、感情変化が起きている場合はそれがポジティブな感情変化かネガティブな感情変化か、およびどの程度の大きさの感情変化が起きているかを解析し、その解析結果に応じた声質変化指標値を算出する。 The biological reaction analysis unit 12 determines whether a positive voice quality change is occurring, a negative voice quality change is occurring, and how loud the voice quality change is occurring, based on the acoustic characteristics of the voice. It may be analyzed and the voice quality change index value according to the analysis result may be calculated. For example, similar to the analysis of facial expressions, it analyzes which emotional element of neutrality / calmness / joy / surprise / sadness / anger / fear / disgust corresponds to the voice according to a voice analysis model machine-learned in advance. do. Then, based on the analysis result, whether or not the emotional change occurs at a predetermined time interval, and if the emotional change occurs, whether it is a positive emotional change or a negative emotional change, and how large it is. It analyzes whether or not the emotional change is occurring, and calculates the voice quality change index value according to the analysis result.
 生体反応解析部12は、以上のようにして算出した表情変化指標値、目線変化指標値、脈拍変化指標値、顔向き変化指標値、発言内容指標値、声質変化指標値の少なくとも1つを用いて生体反応指標値を算出する。例えば、表情変化指標値、目線変化指標値、脈拍変化指標値、顔向き変化指標値、発言内容指標値および声質変化指標値を重み付け計算することにより、生体反応指標値を算出する。 The biological reaction analysis unit 12 uses at least one of the facial expression change index value, the line-of-sight change index value, the pulse change index value, the face orientation change index value, the speech content index value, and the voice quality change index value calculated as described above. The biological reaction index value is calculated. For example, the biological reaction index value is calculated by weighting the facial expression change index value, the line-of-sight change index value, the pulse change index value, the face orientation change index value, the speech content index value, and the voice quality change index value.
 感情評価部13は、生体反応解析部12により対象者について解析された生体反応の変化に基づいて、複数の対象者間で平準化された評価基準に従って対象者の感情の度合いを評価する。例えば、感情評価部13は、生体反応解析部12により対象者について解析された生体反応の変化(生体反応指標値)に基づいて、複数の対象者間で平準化された評価基準に基づく感情反応絶対値を算出する。 The emotion evaluation unit 13 evaluates the degree of emotion of the subject according to the evaluation criteria leveled among the plurality of subjects based on the change in the biological reaction analyzed for the subject by the biological reaction analysis unit 12. For example, the emotion evaluation unit 13 has an emotional response based on an evaluation standard leveled among a plurality of subjects based on the change in the biological reaction (biological reaction index value) analyzed for the subject by the biological reaction analysis unit 12. Calculate the absolute value.
 感情評価部13が算出する感情反応絶対値は、例えば、生体反応解析部12により算出された生体反応指標値を、対象者による同じ感情の生起しやすさに応じて調整した値である。例えば、感情評価部13は、生体反応解析部12により算出された生体反応指標値に対し、同じ感情を生起する頻度に応じた重み値を乗算することによって感情反応絶対値を算出する。 The emotional response absolute value calculated by the emotional evaluation unit 13 is, for example, a value obtained by adjusting the biological reaction index value calculated by the biological reaction analysis unit 12 according to the likelihood of the same emotion occurring by the subject. For example, the emotion evaluation unit 13 calculates the absolute emotional response value by multiplying the biological reaction index value calculated by the biological reaction analysis unit 12 by a weight value according to the frequency of causing the same emotion.
 別の例として、感情評価部13は、生体反応解析部12により算出された表情変化指標値、目線変化指標値、脈拍変化指標値、顔向き変化指標値、発言内容指標値、声質変化指標値のそれぞれ(以下、各指標値と略すことがある)に対し、または、各指標値の中の少なくとも1つの指標値に対し、同じ感情を生起する頻度に応じた重み値を乗算することによって感情反応絶対値を算出するようにしてもよい。なお、以下では、感情反応絶対値の算出に用いる指標値(生体反応指標値、各指標値、少なくとも1つの指標値の何れか)を計算対象指標値という。 As another example, the emotion evaluation unit 13 has a facial expression change index value, a line-of-sight change index value, a pulse change index value, a face orientation change index value, a speech content index value, and a voice quality change index value calculated by the biological reaction analysis unit 12. Emotions by multiplying each of the above (hereinafter, may be abbreviated as each index value) or at least one index value in each index value by a weight value according to the frequency of causing the same emotion. The absolute reaction value may be calculated. In the following, the index value (any of the biological reaction index value, each index value, and at least one index value) used for calculating the absolute emotional response value is referred to as a calculation target index value.
 例えば、対象者Aについて算出された計算対象指標値と対象者Bについて算出された計算対象指標値とが同じ値であった場合としても、同じ感情の生起しやすさ(同じ感情を生起する頻度)が対象者Aと対象者Bとで異なる場合、感情評価部13により算出される感情反応絶対値は対象者Aと対象者Bとで異なる値となる。一例として、感情評価部13は、同じ感情を生起しやすいほど重み値が小さくなり、同じ感情を生起しにくいほど重み値が大きくなるような関数に従って感情反応絶対値を算出する。 For example, even if the calculated target index value calculated for the target person A and the calculated target index value calculated for the target person B are the same value, the susceptibility to generate the same emotion (frequency of generating the same emotion). ) Is different between the subject A and the subject B, the absolute emotional response value calculated by the emotion evaluation unit 13 is different between the subject A and the subject B. As an example, the emotion evaluation unit 13 calculates the absolute emotional response value according to a function such that the weight value becomes smaller as the same emotion is more likely to occur, and the weight value becomes larger as the same emotion is less likely to occur.
 同じ感情を生起する頻度は、例えば、オンラインセッション中に所定の時間区間ごと(例えば、500ミリ秒ごと)に算出される複数の計算対象指標値に基づいて、ほぼ同じ値が発生する回数をカウントすることによって求めることが可能である。ここでいう「ほぼ同じ値」とは、完全に同じ値だけでなく、所定の差分を許容して同じ値とみなすようにした、所定幅を持った値を意味する。なお、1回のオンラインセッションにおいて算出される計算対象指標値だけでなく、複数回のオンラインセッションにおいて算出される計算対象指標値を用いて、同じ感情を生起する頻度を求めるようにしてもよい。 The frequency of causing the same emotion counts, for example, the number of times that approximately the same value occurs based on a plurality of calculated index values calculated for each predetermined time interval (for example, every 500 milliseconds) during an online session. It can be obtained by doing. The term "almost the same value" as used herein means not only a value having exactly the same value but also a value having a predetermined width that allows a predetermined difference to be regarded as the same value. It should be noted that not only the calculated target index value calculated in one online session but also the calculated target index value calculated in a plurality of online sessions may be used to determine the frequency of causing the same emotion.
 この例の場合、感情評価部13は、例えば、各感情を表す情報(所定幅を持った計算対象指標値の区分)と、各感情について同じ感情を生起する頻度に応じて設定された重み値とを対応付けて成るテーブル情報をあらかじめ記憶しておく。そして、感情評価部13は、生体反応解析部12による解析結果とテーブル情報とを照合することによって、解析結果に対応する重み値を抽出し、当該重み値を計算対象指標値に乗算することによって感情反応絶対値を算出する。なお、テーブル情報に代えて、各感情を表す情報をもとに重み値が算出されるように設計された関数を用いてもよい。 In the case of this example, the emotion evaluation unit 13 has, for example, information representing each emotion (division of a calculation target index value having a predetermined width) and a weight value set according to the frequency of generating the same emotion for each emotion. The table information associated with and is stored in advance. Then, the emotion evaluation unit 13 extracts the weight value corresponding to the analysis result by collating the analysis result by the biological reaction analysis unit 12 with the table information, and multiplies the weight value by the calculation target index value. Calculate the absolute emotional response value. In addition, instead of the table information, a function designed to calculate the weight value based on the information representing each emotion may be used.
 このように、本実施形態では、同じ感情を生起する頻度をもとに算出した感情反応絶対値を用いて、対象者ごとに異なる感情の生起しやすさを考慮した感情の度合を評価しているので、対象者に関する真の意味での感情の度合いを評価することが可能となり、異なる対象者間で感情の度合いを客観的に対比することができる。 As described above, in the present embodiment, the degree of emotions considering the susceptibility to different emotions for each subject is evaluated by using the emotional response absolute value calculated based on the frequency of causing the same emotions. Therefore, it is possible to evaluate the degree of emotion in the true sense of the subject, and it is possible to objectively compare the degree of emotion between different subjects.
 以上説明した生体反応解析部12および感情評価部13の処理は、動画像取得部11が複数の対象者の動画像を取得したときにリアルタイムに行うようにしてもよいし、動画像記憶部101に記憶された動画像を用いて事後的に行うようにしてもよい。 The processing of the biological reaction analysis unit 12 and the emotion evaluation unit 13 described above may be performed in real time when the moving image acquisition unit 11 acquires the moving images of a plurality of subjects, or the moving image storage unit 101. It may be performed after the fact by using the moving image stored in.
 なお、上記実施形態では、対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価する例について説明したが、本発明はこれに限定されない。例えば、感情評価部13は、平常時の生体反応に対する現在の生体反応の違いの大きさに基づく感情の程度であって、対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価するようにしてもよい。平常時の生体反応とは、例えば、同一対象者の過去の生体反応分布に基づいて解析される主要な生体反応とすることが可能である。 In the above embodiment, an example of evaluating the degree of emotion adjusted according to the likelihood of the same emotion being generated by the subject has been described, but the present invention is not limited to this. For example, the emotion evaluation unit 13 is a degree of emotion based on the magnitude of the difference in the current biological reaction to the biological reaction in normal times, and the emotion is adjusted according to the likelihood of the same emotion being generated by the subject. The degree may be evaluated. The biological reaction in normal times can be, for example, a major biological reaction analyzed based on the past biological reaction distribution of the same subject.
 例えば、感情評価部13は、生体反応解析部12により算出された計算対象指標値を、平常時の生体反応に対する現在の生体反応の違いの大きさと、対象者による同じ感情の生起しやすさとに応じて調整することによって感情反応絶対値を算出する。このように算出される感情反応絶対値は、平常時の生体反応に対する現在の生体反応の違いの大きさに基づく感情の程度を表す値であって、対象者が同じ感情を生起しやすいまたは生起しにくい度合いに応じて調整された値である。 For example, the emotion evaluation unit 13 uses the calculated index value calculated by the biological reaction analysis unit 12 to determine the magnitude of the difference in the current biological reaction to the biological reaction in normal times and the susceptibility to the same emotion by the subject. The absolute value of emotional response is calculated by adjusting accordingly. The absolute emotional response value calculated in this way is a value representing the degree of emotion based on the magnitude of the difference in the current biological response to the biological response in normal times, and the subject is likely to generate the same emotion or occurs. It is a value adjusted according to the degree of difficulty.
 また、上記実施形態において説明した同じ感情を生起する頻度の求め方は一例であり、本発明はこれに限定されるものではない。例えば、オンラインセッション中に所定の時間区間ごとに生体反応解析部12により解析される感情要素(中立、落ち着き、喜び、驚き、悲しみ、怒り、恐れ、嫌悪感の何れか)に基づいて、同じ感情要素が発生する回数をカウントすることによって、同じ感情を生起する頻度を求めるようにしてもよい。あるいは、喜びおよび驚きの2つを「快」の感情と定義する一方、悲しみ、怒り、恐れおよび嫌悪感の4つを「不快」の感情と定義し、オンラインセッション中に所定の時間区間ごとに生体反応解析部12により解析される感情要素に基づいて、「快」の感情を生起する頻度と「不快」の感情を生起する頻度とを求めるようにしてもよい。 Further, the method of determining the frequency of causing the same emotion described in the above embodiment is an example, and the present invention is not limited to this. For example, the same emotion based on the emotional element (neutral, calm, joy, surprise, sadness, anger, fear, disgust) analyzed by the biological reaction analysis unit 12 at predetermined time intervals during the online session. By counting the number of times the element occurs, the frequency with which the same emotion is generated may be determined. Alternatively, joy and surprise are defined as "pleasant" emotions, while sadness, anger, fear and disgust are defined as "unpleasant" emotions, and every predetermined time interval during an online session. Based on the emotional elements analyzed by the biological reaction analysis unit 12, the frequency of generating "pleasant" emotions and the frequency of generating "unpleasant" emotions may be determined.
 生体反応解析部12により解析される感情要素に基づいて、同じ感情要素が発生する回数をカウントすることによって頻度を求める場合、感情評価部13は、例えば、各感情を表す情報(各感情要素)と、各感情について同じ感情を生起する頻度に応じて設定された重み値とを対応付けて成るテーブル情報をあらかじめ記憶しておく。そして、感情評価部13は、生体反応解析部12による解析結果とテーブル情報とを照合することによって、解析結果に対応する重み値を抽出し、当該重み値を当該感情要素に対応するスコアに乗算することによって計算対象指標値を算出し、さらにその計算対象指標値から感情反応絶対値を算出する。なお、テーブル情報に代えて、各感情を表す情報をもとに重み値が算出されるように設計された関数を用いてもよい。 When the frequency is obtained by counting the number of times the same emotion element occurs based on the emotion element analyzed by the biological reaction analysis unit 12, the emotion evaluation unit 13 is, for example, information representing each emotion (each emotion element). And the table information corresponding to the weight value set according to the frequency of causing the same emotion for each emotion is stored in advance. Then, the emotion evaluation unit 13 extracts the weight value corresponding to the analysis result by collating the analysis result by the biological reaction analysis unit 12 with the table information, and multiplies the weight value by the score corresponding to the emotion element. By doing so, the calculation target index value is calculated, and further, the emotional response absolute value is calculated from the calculation target index value. In addition, instead of the table information, a function designed to calculate the weight value based on the information representing each emotion may be used.
 同様に、生体反応解析部12により解析される感情要素に基づいて、快/不快の感情を生起する頻度を求める場合、感情評価部13は、例えば、各感情を表す情報(快/不快の区分)と、各感情について同じ感情を生起する頻度に応じて設定された重み値とを対応付けて成るテーブル情報をあらかじめ記憶しておく。そして、感情評価部13は、生体反応解析部12による解析結果とテーブル情報とを照合することによって、解析結果に対応する重み値を抽出し、当該重み値を快または不快と定義される感情要素に対応するスコアに乗算することによって計算対象指標値を算出し、さらにその計算対象指標値から感情反応絶対値を算出する。なお、テーブル情報に代えて、各感情を表す情報をもとに重み値が算出されるように設計された関数を用いてもよい。 Similarly, when determining the frequency of causing pleasant / unpleasant emotions based on the emotional elements analyzed by the biological reaction analysis unit 12, the emotional evaluation unit 13 may obtain, for example, information representing each emotion (comfort / discomfort classification). ) And the weight value set according to the frequency of causing the same emotion for each emotion are stored in advance. Then, the emotion evaluation unit 13 extracts the weight value corresponding to the analysis result by collating the analysis result by the biological reaction analysis unit 12 with the table information, and the emotion element whose weight value is defined as pleasant or unpleasant. The calculation target index value is calculated by multiplying the score corresponding to, and the emotional response absolute value is calculated from the calculation target index value. In addition, instead of the table information, a function designed to calculate the weight value based on the information representing each emotion may be used.
 また、上記実施形態では、同じ感情の生起しやすさを表す尺度として、同じ感情を生起する頻度を用いる例について説明したが、これに限定されない。例えば、同じ感情を生起する頻度に代えてまたは加えて、対象者の性質または性格を用いるようにしてもよい。 Further, in the above embodiment, an example in which the frequency of generating the same emotion is used as a measure for expressing the susceptibility to the same emotion has been described, but the present invention is not limited to this. For example, the nature or personality of the subject may be used in place of or in addition to the frequency with which the same emotions occur.
 また、上記実施形態では、複数人の参加者が使用する複数のユーザ端末200の間でセッション管理サーバ300を通じてオンラインセッションが行われる環境において、参加者を解析の対象者として感情を解析する例について説明したが、感情解析の対象者は、オンラインセッションの参加者に限定されるものではない。すなわち、動画像を取得できる環境にいる者であれば、本実施形態の対象者として感情の解析を行うことが可能である。 Further, in the above embodiment, an example of analyzing emotions with a participant as a target of analysis in an environment where an online session is performed through a session management server 300 between a plurality of user terminals 200 used by a plurality of participants. As explained, the target audience for emotion analysis is not limited to participants in online sessions. That is, any person who is in an environment where a moving image can be acquired can analyze emotions as a target person of the present embodiment.
 その他、上記実施形態は、何れも本発明を実施するにあたっての具体化の一例を示したものに過ぎず、これによって本発明の技術的範囲が限定的に解釈されてはならないものである。すなわち、本発明はその要旨、またはその主要な特徴から逸脱することなく、様々な形で実施することができる。 Other than that, all of the above embodiments are merely examples of the embodiment of the present invention, and the technical scope of the present invention should not be construed in a limited manner by this. That is, the present invention can be implemented in various forms without departing from its gist or its main features.
 11 動画像取得部
 12 生体反応解析部
 13 感情評価部
 100 感情解析装置
11 Moving image acquisition unit 12 Biological reaction analysis unit 13 Emotion evaluation unit 100 Emotion analysis device

Claims (10)

  1.  複数人の参加者でオンラインセッションが行われる環境において、上記参加者を解析の対象者として感情を解析する感情解析システムであって、
     上記オンラインセッション中に上記対象者のユーザ端末から送信される動画像を取得する動画像取得部と、
     上記動画像取得部により取得された動画像に基づいて、上記対象者について感情の変化に起因して起こる生体反応の変化を解析する生体反応解析部と、
     上記生体反応解析部により上記対象者について解析された上記生体反応の変化に基づいて、上記対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価する感情評価部とを備えた
    ことを特徴とする感情解析システム。
    It is an emotion analysis system that analyzes emotions with the above participants as the analysis target in an environment where an online session is held by multiple participants.
    A moving image acquisition unit that acquires a moving image transmitted from the user terminal of the target person during the online session, and a moving image acquisition unit.
    Based on the moving image acquired by the moving image acquisition unit, the biological reaction analysis unit that analyzes the change in the biological reaction caused by the emotional change of the subject, and the biological reaction analysis unit.
    An emotion evaluation unit that evaluates the degree of emotion adjusted according to the likelihood of the same emotion being generated by the subject based on the change in the biological reaction analyzed for the subject by the biological reaction analysis unit. An emotion analysis system characterized by being prepared.
  2.  上記感情評価部は、平常時の生体反応に対する現在の生体反応の違いの大きさに基づく感情の程度であって、上記対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価することを特徴とする請求項1に記載の感情解析システム。 The emotion evaluation unit is the degree of emotion based on the magnitude of the difference in the current biological reaction to the biological reaction in normal times, and the degree of emotion adjusted according to the likelihood of the same emotion being generated by the subject. The emotion analysis system according to claim 1, wherein the emotional analysis system is characterized in that.
  3.  上記生体反応解析部は、上記生体反応の変化を所定の基準に従って数値化することによって生体反応指標値を算出し、
     上記感情評価部は、上記生体反応解析部により算出された上記生体反応指標値を上記対象者による同じ感情の生起しやすさに応じて調整した値である感情反応絶対値を算出する
    ことを特徴とする請求項1に記載の感情解析システム。
    The biological reaction analysis unit calculates the biological reaction index value by quantifying the change in the biological reaction according to a predetermined standard.
    The emotion evaluation unit is characterized in that it calculates an absolute emotional response value, which is a value obtained by adjusting the biological reaction index value calculated by the biological reaction analysis unit according to the likelihood of the same emotion occurring by the subject. The emotion analysis system according to claim 1.
  4.  上記生体反応解析部は、上記生体反応の変化を所定の基準に従って数値化することによって生体反応指標値を算出し、
     上記感情評価部は、上記生体反応解析部により算出された上記生体反応指標値を、上記平常時の生体反応に対する上記現在の生体反応の違いの大きさと、上記対象者による同じ感情の生起しやすさとに応じて調整することによって感情反応絶対値を算出する
    ことを特徴とする請求項2に記載の感情解析システム。
    The biological reaction analysis unit calculates the biological reaction index value by quantifying the change in the biological reaction according to a predetermined standard.
    The emotion evaluation unit uses the biological reaction index value calculated by the biological reaction analysis unit as the magnitude of the difference in the current biological reaction with respect to the biological reaction in normal times, and the tendency of the subject to cause the same emotion. The emotion analysis system according to claim 2, wherein the absolute value of the emotional response is calculated by adjusting according to the situation.
  5.  上記感情評価部は、上記生体反応解析部により算出された上記生体反応指標値に対し、上記対象者が同じ感情を生起する頻度に応じた重み値を乗算することによって上記感情反応絶対値を算出することを特徴とする請求項3に記載の感情解析システム。 The emotion evaluation unit calculates the absolute emotional response value by multiplying the biological reaction index value calculated by the biological reaction analysis unit by a weight value according to the frequency at which the subject causes the same emotion. The emotion analysis system according to claim 3, wherein the emotion analysis system is characterized in that.
  6.  上記生体反応解析部は、上記動画像取得部により取得された動画像にける顔画像および音声の少なくとも一方を解析し、表情、目線、脈拍、顔の動き、発言内容、声質の少なくとも1つに関する生体反応の変化を所定の基準に従って数値化することにより、表情変化指標値、目線変化指標値、脈拍変化指標値、顔向き変化指標値、発言内容指標値および声質変化指標値の少なくとも1つを算出し、
     上記感情評価部は、上記生体反応解析部により算出された上記表情変化指標値、上記目線変化指標値、上記脈拍変化指標値、上記顔向き変化指標値、上記発言内容指標値および上記声質変化指標値の少なくとも1つに対し、上記対象者が同じ感情を生起する頻度に応じた重み値を乗算することによって上記感情反応絶対値を算出する
    ことを特徴とする請求項3に記載の感情解析システム。
     
    The biological reaction analysis unit analyzes at least one of the facial image and the voice in the moving image acquired by the moving image acquisition unit, and relates to at least one of facial expression, line of sight, pulse, facial movement, speech content, and voice quality. By quantifying the change in the biological reaction according to a predetermined standard, at least one of the facial expression change index value, the line-of-sight change index value, the pulse change index value, the face orientation change index value, the speech content index value, and the voice quality change index value can be obtained. Calculate and
    The emotion evaluation unit includes the facial expression change index value, the line-of-sight change index value, the pulse change index value, the face orientation change index value, the remark content index value, and the voice quality change index value calculated by the biological reaction analysis unit. The emotion analysis system according to claim 3, wherein the emotion response absolute value is calculated by multiplying at least one of the values by a weight value according to the frequency at which the subject causes the same emotion. ..
  7.  対象者について得られる動画像に基づいて、上記対象者について感情の変化に起因して起こる生体反応の変化を解析する生体反応解析部と、
     上記生体反応解析部により上記対象者について解析された上記生体反応の変化に基づいて、上記対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価する感情評価部とを備えた
    ことを特徴とする感情解析装置。
    Based on the moving image obtained for the subject, the biological reaction analysis unit that analyzes the change in the biological reaction caused by the change in emotions of the subject, and the biological reaction analysis unit.
    An emotion evaluation unit that evaluates the degree of emotion adjusted according to the likelihood of the same emotion being generated by the subject based on the change in the biological reaction analyzed for the subject by the biological reaction analysis unit. An emotion analysis device characterized by being equipped.
  8.  上記感情評価部は、平常時の生体反応に対する現在の生体反応の違いの大きさに基づく感情の程度であって、上記対象者による同じ感情の生起しやすさに応じて調整された感情の度合いを評価することを特徴とする請求項7に記載の感情解析装置。 The emotion evaluation unit is the degree of emotion based on the magnitude of the difference in the current biological reaction to the biological reaction in normal times, and the degree of emotion adjusted according to the likelihood of the same emotion being generated by the subject. The emotion analysis apparatus according to claim 7, wherein the emotional analysis device is characterized in that.
  9.  上記生体反応解析部は、上記生体反応の変化を所定の基準に従って数値化することによって生体反応指標値を算出し、
     上記感情評価部は、上記生体反応解析部により算出された上記生体反応指標値を上記対象者による同じ感情の生起しやすさに応じて調整した値である感情反応絶対値を算出する
    ことを特徴とする請求項7に記載の感情解析装置。
    The biological reaction analysis unit calculates the biological reaction index value by quantifying the change in the biological reaction according to a predetermined standard.
    The emotion evaluation unit is characterized in that it calculates an absolute emotional response value, which is a value obtained by adjusting the biological reaction index value calculated by the biological reaction analysis unit according to the likelihood of the same emotion occurring by the subject. The emotion analysis device according to claim 7.
  10.  上記生体反応解析部は、上記生体反応の変化を所定の基準に従って数値化することによって生体反応指標値を算出し、
     上記感情評価部は、上記生体反応解析部により算出された上記生体反応指標値を、上記平常時の生体反応に対する上記現在の生体反応の違いの大きさと、上記対象者による同じ感情の生起しやすさとに応じて調整することによって感情反応絶対値を算出する
    ことを特徴とする請求項8に記載の感情解析装置。
    The biological reaction analysis unit calculates the biological reaction index value by quantifying the change in the biological reaction according to a predetermined standard.
    The emotion evaluation unit uses the biological reaction index value calculated by the biological reaction analysis unit as the magnitude of the difference in the current biological reaction with respect to the biological reaction in normal times, and the tendency of the subject to cause the same emotion. The emotion analysis device according to claim 8, wherein the absolute value of the emotional response is calculated by adjusting according to the above.
PCT/JP2021/027638 2020-07-31 2021-07-27 Emotion analysis system and emotion analysis device WO2022025025A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2022539467A JPWO2022025025A1 (en) 2020-07-31 2021-07-27

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JPPCT/JP2020/029468 2020-07-31
PCT/JP2020/029468 WO2022024355A1 (en) 2020-07-31 2020-07-31 Emotion analysis system

Publications (1)

Publication Number Publication Date
WO2022025025A1 true WO2022025025A1 (en) 2022-02-03

Family

ID=80035318

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2020/029468 WO2022024355A1 (en) 2020-07-31 2020-07-31 Emotion analysis system
PCT/JP2021/027638 WO2022025025A1 (en) 2020-07-31 2021-07-27 Emotion analysis system and emotion analysis device

Family Applications Before (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/029468 WO2022024355A1 (en) 2020-07-31 2020-07-31 Emotion analysis system

Country Status (2)

Country Link
JP (1) JPWO2022025025A1 (en)
WO (2) WO2022024355A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130144937A1 (en) * 2011-12-02 2013-06-06 Samsung Electronics Co., Ltd. Apparatus and method for sharing user's emotion
JP2014186451A (en) * 2013-03-22 2014-10-02 Nippon Telegr & Teleph Corp <Ntt> Emotion estimation device, emotion estimation method, and program
US20170177928A1 (en) * 2014-08-08 2017-06-22 International Business Machines Corporation Sentiment analysis in a video conference

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001083984A (en) * 1999-09-09 2001-03-30 Alpine Electronics Inc Interface device
JP2011039934A (en) * 2009-08-17 2011-02-24 Tokai Univ Emotion estimation system and learning system using the same
JP2015141428A (en) * 2014-01-27 2015-08-03 株式会社日立システムズ Server device, feeling notification system, feeling notification method, and program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130144937A1 (en) * 2011-12-02 2013-06-06 Samsung Electronics Co., Ltd. Apparatus and method for sharing user's emotion
JP2014186451A (en) * 2013-03-22 2014-10-02 Nippon Telegr & Teleph Corp <Ntt> Emotion estimation device, emotion estimation method, and program
US20170177928A1 (en) * 2014-08-08 2017-06-22 International Business Machines Corporation Sentiment analysis in a video conference

Also Published As

Publication number Publication date
WO2022024355A1 (en) 2022-02-03
JPWO2022025025A1 (en) 2022-02-03

Similar Documents

Publication Publication Date Title
WO2022025200A1 (en) Reaction analysis system and reaction analysis device
JP6923827B2 (en) Communication skill evaluation systems, devices, methods, and programs
US11341619B2 (en) Method to provide a video with a computer-modified visual of a desired face of a person
WO2022024956A1 (en) Emotion analysis system and emotion analysis device
WO2022025024A1 (en) Reaction analysis system and reaction analysis device
WO2022025025A1 (en) Emotion analysis system and emotion analysis device
WO2022064622A1 (en) Emotion analysis system
CN114566187A (en) System, electronic device, and related method with post-session representation
WO2022024356A1 (en) Organization attribute analysis system
JP7465040B1 (en) Communication visualization system
WO2023007661A1 (en) Organizational attribute analysis system and organizational attribute analysis device
WO2022201272A1 (en) Video analysis program
WO2022269801A1 (en) Video analysis system
JP7145556B1 (en) Video image analysis system
JP7197955B1 (en) Video meeting evaluation terminal
WO2022254497A1 (en) Video analysis system
WO2023135939A1 (en) Information processing device, information processing method, and program
WO2022201267A1 (en) Video analysis program
JP7121433B1 (en) Video analysis program
JP7121436B1 (en) Video analysis program
JP7138997B1 (en) Video meeting evaluation terminal
JP7121439B1 (en) Video image analysis system
WO2022201265A1 (en) Video analysis program
WO2022264219A1 (en) Video analysis system
WO2022201268A1 (en) Video analysis programme

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21851213

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022539467

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21851213

Country of ref document: EP

Kind code of ref document: A1