WO2022180855A1 - Terminal, système et programme d'évaluation de session vidéo - Google Patents

Terminal, système et programme d'évaluation de session vidéo Download PDF

Info

Publication number
WO2022180855A1
WO2022180855A1 PCT/JP2021/007572 JP2021007572W WO2022180855A1 WO 2022180855 A1 WO2022180855 A1 WO 2022180855A1 JP 2021007572 W JP2021007572 W JP 2021007572W WO 2022180855 A1 WO2022180855 A1 WO 2022180855A1
Authority
WO
WIPO (PCT)
Prior art keywords
moving image
user
evaluation
analysis
unit
Prior art date
Application number
PCT/JP2021/007572
Other languages
English (en)
Japanese (ja)
Inventor
渉三 神谷
Original Assignee
株式会社I’mbesideyou
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社I’mbesideyou filed Critical 株式会社I’mbesideyou
Priority to PCT/JP2021/007572 priority Critical patent/WO2022180855A1/fr
Priority to JP2023502015A priority patent/JPWO2022180855A1/ja
Publication of WO2022180855A1 publication Critical patent/WO2022180855A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present disclosure relates to a video session evaluation terminal, a video session evaluation system, and a video session evaluation program.
  • Patent Document 1 Conventionally, there is known a technique for analyzing the emotions others receive in response to a speaker's remarks (see Patent Document 1, for example). There is also known a technique for analyzing changes in facial expressions of a subject in chronological order over a long period of time and estimating the emotions held during that period (see, for example, Patent Literature 2). Furthermore, there are known techniques for identifying factors that have the greatest influence on changes in emotions (see Patent Documents 3 to 5, for example). Furthermore, there is also known a technology that compares the subject's usual facial expression with the current facial expression and issues an alert when the facial expression is dark (see, for example, Patent Document 6).
  • Patent Documents 7 to 9 There is also known a technique for determining the degree of emotion of a subject by comparing the subject's normal (expressionless) facial expression with the current facial expression (for example, Patent Documents 7 to 9). reference). Furthermore, there is also known a technique for analyzing the feeling of an organization and the atmosphere within a group that an individual feels (see Patent Documents 10 and 11, for example).
  • the purpose of the present invention is to objectively evaluate exchanged communication in order to conduct more efficient communication in situations where online communication is the main focus.
  • Acquisition means for acquiring a moving image relating to an online session between the first user and the second user; face recognition means for recognizing at least face images of the first user and the second user included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for calculating an evaluation value from a plurality of viewpoints based on both the recognized face image and the recognized voice; determining means for determining the degree of matching of the second user to the first user based on the evaluation value; A moving image analysis system is obtained.
  • a moving image acquisition unit that acquires a plurality of moving images obtained by photographing a target person; a biological reaction analysis unit that analyzes changes in the biological reaction of the subject based on the moving image acquired by the moving image acquisition unit; Based on the change in the biological reaction of the subject analyzed by the biological reaction analysis unit, the emotional level of the subject is evaluated according to a standardized evaluation standard for the subject among the plurality of moving images.
  • a moving image analysis system comprising an emotion evaluation unit is obtained.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for classifying into predetermined emotional information based on both the recognized face image and the recognized voice; an annotation receiving means for receiving an annotation operation on the classified emotional information;
  • a moving image analysis system is obtained.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for calculating an evaluation value from a plurality of viewpoints based on both the recognized face image and the recognized voice; evaluation value providing means for providing the subject with an average value for each of the evaluation values from the plurality of viewpoints over a predetermined period; A moving image analysis system is obtained.
  • Acquisition means for acquiring a moving image relating to an online session between the first user and the second user; face recognition means for recognizing at least a face image of the first user included in the moving image for each predetermined frame; speech recognition means for recognizing at least the speech of the second user included in the moving image; evaluation means for calculating an evaluation value from a predetermined viewpoint based on at least the recognized face image; keyword detection means for detecting at least a predetermined keyword in the recognized speech; an alert transmission means for transmitting a predetermined alert based on the evaluation value when the keyword is detected; A moving image analysis system is obtained.
  • Acquisition means for acquiring a plurality of moving images showing at least a target person; voice recognition means for recognizing at least the voice of the target person included in the evaluation target moving image among the moving images; a unique word extracting means for extracting, from among the words included in the recognized speech, words not included in the moving image other than the evaluation target moving image; a text display means for converting the extracted word into a size according to the frequency of its utterance and displaying it as text;
  • a moving image analysis system is obtained.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the recognized voice; text conversion means for converting words included in the recognized speech into text and displaying the text; size setting means for setting the size of the converted text to a size corresponding to the frequency of speech; a color setting means for setting the color of the converted text to a color corresponding to the evaluation value; comprising A moving image analysis system is obtained.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for evaluating the degree of fatigue based on both the recognized face image and the voice, A moving image analysis system is obtained.
  • an acquisition means for acquiring at least a moving image; speech recognition means for recognizing utterances for each speaker included in the moving image; and an object generation unit that generates an object that associates the utterance with the target person and displays them in chronological order.
  • an acquisition means for acquiring at least a moving image; speech recognition means for recognizing utterances for each target person included in the moving image;
  • a moving image analysis system is obtained that includes a utterance object generation unit that plots a utterance object corresponding to the utterance in association with the target person.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; intonation acquisition means for extracting intonation information of the recognized speech; and evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the intonation information.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; context acquisition means for acquiring context information of the moving image; evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the recognized voice; a correction means for correcting the evaluation value using the context information; Video image analysis system.
  • exchanged communication can be objectively evaluated in order to conduct more efficient communication in situations where online communication is the main activity.
  • FIG. 1 is an example of a functional block diagram of an evaluation terminal according to an embodiment of the present invention
  • FIG. 3 is a diagram showing functional configuration example 1 of the evaluation terminal according to the embodiment of the present invention
  • FIG. 8 is a diagram showing functional configuration example 2 of the evaluation terminal according to the embodiment of the present invention
  • FIG. 10 is a diagram showing a functional configuration example 3 of the evaluation terminal according to the embodiment of the present invention
  • 7 is a screen display example according to the functional configuration example 3 of FIG. 6.
  • FIG. FIG. 7 is another screen display example according to the functional configuration example 3 of FIG. 6.
  • FIG. FIG. 12 is a diagram showing another configuration of functional configuration example 3 of the evaluation terminal according to the embodiment of the present invention
  • FIG. 12 is a diagram showing another configuration of functional configuration example 3 of the evaluation terminal according to the embodiment of the present invention
  • 1 shows a system according to a first embodiment of the invention
  • FIG. 1 shows a system according to a first embodiment of the invention
  • FIG. Fig. 3 shows a system according to a second embodiment of the invention
  • Fig. 3 shows a system according to a second embodiment of the invention
  • Fig. 3 shows a system according to a third embodiment of the invention
  • Fig. 4 shows a system according to a fourth embodiment of the invention
  • Fig. 5 shows a system according to a fifth embodiment of the invention
  • Fig. 6 shows a system according to a sixth embodiment of the invention
  • Fig. 6 shows a system according to a sixth embodiment of the invention
  • Fig. 6 shows a system according to a sixth embodiment of the invention
  • Fig. 12 shows a diagram showing another configuration of functional configuration example 3 of the evaluation terminal according to the embodiment of the present invention
  • 1 shows a system according to
  • FIG. 7 shows a system according to a seventh embodiment of the invention
  • Fig. 7 shows a system according to a seventh embodiment of the invention
  • FIG. 11 illustrates a system according to an eighth embodiment of the invention
  • Fig. 10 shows a system according to a ninth embodiment of the invention
  • Fig. 10 shows a system according to a tenth embodiment of the invention
  • Fig. 10 shows a system according to a tenth embodiment of the invention
  • FIG. 11 illustrates a system according to an eleventh embodiment of the invention
  • the present disclosure has the following configurations.
  • Acquisition means for acquiring a moving image relating to an online session between the first user and the second user; face recognition means for recognizing at least face images of the first user and the second user included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for calculating an evaluation value from a plurality of viewpoints based on both the recognized face image and the recognized voice; determining means for determining the degree of matching of the second user to the first user based on the evaluation value; Video image analysis system.
  • a moving image acquisition unit that acquires a plurality of moving images obtained by photographing a target person; a biological reaction analysis unit that analyzes changes in the biological reaction of the subject based on the moving image acquired by the moving image acquisition unit; Based on the change in the biological reaction of the subject analyzed by the biological reaction analysis unit, the emotional level of the subject is evaluated according to a standardized evaluation standard for the subject among the plurality of moving images.
  • a moving image analysis system comprising an emotion evaluation unit.
  • An acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for classifying into predetermined emotional information based on both the recognized face image and the recognized voice; an annotation receiving means for receiving an annotation operation on the classified emotional information; Video image analysis system.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for calculating an evaluation value from a plurality of viewpoints based on both the recognized face image and the recognized voice; evaluation value providing means for providing the subject with an average value for each of the evaluation values from the plurality of viewpoints over a predetermined period; Video image analysis system.
  • Acquisition means for acquiring a moving image relating to an online session between the first user and the second user; face recognition means for recognizing at least a face image of the first user included in the moving image for each predetermined frame; speech recognition means for recognizing at least the speech of the second user included in the moving image; evaluation means for calculating an evaluation value from a predetermined viewpoint based on at least the recognized face image; keyword detection means for detecting at least a predetermined keyword in the recognized speech; an alert transmission means for transmitting a predetermined alert based on the evaluation value when the keyword is detected; Video image analysis system.
  • Acquisition means for acquiring a plurality of moving images showing at least a target person; voice recognition means for recognizing at least the voice of the target person included in the evaluation target moving image among the moving images; a unique word extracting means for extracting, from among the words included in the recognized speech, words not included in the moving image other than the evaluation target moving image; a text display means for converting the extracted word into a size according to the frequency of its utterance and displaying it as text; Video image analysis system.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the recognized voice; text conversion means for converting words included in the recognized speech into text and displaying the text; size setting means for setting the size of the converted text to a size corresponding to the frequency of speech; a color setting means for setting the color of the converted text to a color corresponding to the evaluation value; comprising Video image analysis system.
  • an acquisition means for acquiring at least a moving image; face recognition means for recognizing at least a face image of a target person included in the moving image for each predetermined frame; voice recognition means for recognizing at least the voice of the subject included in the moving image; context acquisition means for acquiring context information of the moving image; evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the recognized voice; a correction means for correcting the evaluation value using the context information; Video image analysis system.
  • a video session in an environment where a video session (hereinafter referred to as an online session including one-way and two-way sessions) is held by a plurality of people, the person to be analyzed among the plurality of people is different from the others. It is a system that analyzes and evaluates specific emotions (feelings that occur in response to one's own or others' words and actions. pleasant/unpleasant, or their degree).
  • Online sessions are, for example, online meetings, online classes, online chats, etc. Terminals installed in multiple locations are connected to a server via a communication network such as the Internet, and moving images are transmitted between multiple terminals through the server. It's made to be interactable.
  • Moving images handled in online sessions include facial images and voices of users using terminals.
  • Moving images also include images such as materials that are shared and viewed by a plurality of users. It is possible to switch between the face image and the document image on the screen of each terminal to display only one of them, or to divide the display area and display the face image and the document image at the same time. In addition, it is possible to display the image of one user out of a plurality of users on the full screen, or divide the images of some or all of the users into small screens and display them. It is possible to designate one or a plurality of users among a plurality of users participating in an online session using terminals as analysis subjects.
  • an online session leader, moderator, or manager designates any user as an analysis subject.
  • Hosts of online sessions are, for example, instructors of online classes, chairpersons and facilitators of online meetings, coaches of sessions for coaching purposes, and the like.
  • An online session host is typically one of the users participating in the online session, but may be another person who does not participate in the online session. It should be noted that all participants may be subject to analysis without specifying the person to be analyzed.
  • an online session leader, moderator, or administrator hereinafter collectively referred to as the organizer to designate any user as an analysis subject.
  • Hosts of online sessions are, for example, instructors of online classes, chairpersons and facilitators of online meetings, coaches of sessions for coaching purposes, and the like.
  • An online session host is typically one of the users participating in the online session, but may be another person who does not participate in the online session.
  • the video session evaluation system displays at least moving images obtained from a video session established between a plurality of terminals.
  • the displayed moving image is acquired by the terminal, and at least a face image included in the moving image is identified for each predetermined frame unit. An evaluation value for the identified face image is then calculated.
  • the evaluation value is shared as necessary.
  • the acquired moving image is stored in the terminal, analyzed and evaluated on the terminal, and the result is provided to the user of the terminal. Therefore, for example, even a video session containing personal information or a video session containing confidential information can be analyzed and evaluated without providing the moving image itself to an external evaluation agency or the like.
  • the evaluation result evaluation value
  • the video session evaluation system includes user terminals 10 and 20 each having at least an input unit such as a camera unit and a microphone unit, a display unit such as a display, and an output unit such as a speaker. , a video session service terminal 30 for providing an interactive video session to the user terminals 10, 20, and an evaluation terminal 40 for performing part of the evaluation of the video session.
  • Each functional block, functional unit, and functional module described below can be configured by any of hardware, DSP (Digital Signal Processor), and software provided in a computer, for example.
  • DSP Digital Signal Processor
  • a computer CPU random access memory
  • RAM random access memory
  • ROM read-only memory
  • a series of processes by the systems and terminals described herein may be implemented using software, hardware, or a combination of software and hardware. It is possible to create a computer program for realizing each function of the information sharing support device 10 according to the present embodiment and implement it in a PC or the like. It is also possible to provide a computer-readable recording medium storing such a computer program.
  • the recording medium is, for example, a magnetic disk, an optical disk, a magneto-optical disk, a flash memory, or the like.
  • the above computer program may be distributed, for example, via a network without using a recording medium.
  • the evaluation terminal acquires a moving image from a video session service terminal, identifies at least a face image included in the moving image for each predetermined frame unit, and calculates an evaluation value for the face image ( will be described in detail later).
  • this service provides user terminals 10 and 20 with two-way images and voice. Communication is possible.
  • a moving image captured by the camera of the other user's terminal is displayed on the display of the user's terminal, and audio captured by the microphone of the other's user's terminal can be output from the speaker.
  • this service allows both or either of the user terminals to record moving images and sounds (collectively referred to as “moving images, etc.") in the storage unit of at least one of the user terminals. configured as possible.
  • the recorded moving image information Vs (hereinafter referred to as “recorded information”) is cached in the user terminal that started recording and is locally recorded only in one of the user terminals. If necessary, the user can view the recorded information by himself or share it with others within the scope of using this service.
  • FIG. 4 is a block diagram showing a configuration example according to this embodiment.
  • the video session evaluation system of this embodiment is implemented as a functional configuration of the user terminal 10.
  • the user terminal 10 has, as its functions, a moving image acquisition unit 11, a biological reaction analysis unit 12, a peculiar determination unit 13, a related event identification unit 14, a clustering unit 15, and an analysis result notification unit 16.
  • the moving image acquisition unit 11 acquires from each terminal a moving image obtained by photographing a plurality of people (a plurality of users) with a camera provided in each terminal during an online session. It does not matter whether the moving image acquired from each terminal is set to be displayed on the screen of each terminal. That is, the moving image acquisition unit 11 acquires moving images from each terminal, including moving images being displayed and moving images not being displayed on each terminal.
  • the biological reaction analysis unit 12 analyzes changes in the biological reaction of each of a plurality of people based on the moving images (whether or not they are being displayed on the screen) acquired by the moving image acquiring unit 11.
  • the biological reaction analysis unit 12 separates the moving image acquired by the moving image acquisition unit 11 into a set of images (collection of frame images) and voice, and analyzes changes in the biological reaction from each.
  • the biological reaction analysis unit 12 analyzes the user's facial image using a frame image separated from the moving image acquired by the moving image acquisition unit 11 to obtain at least one of facial expression, gaze, pulse, and facial movement. Analyze changes in biological reactions related to Further, the biological reaction analysis unit 12 analyzes the voice separated from the moving image acquired by the moving image acquisition unit 11 to analyze changes in the biological reaction related to at least one of the user's utterance content and voice quality.
  • the biological reaction analysis unit 12 calculates a biological reaction index value reflecting the change in biological reaction by quantifying the change in biological reaction according to a predetermined standard.
  • the analysis of changes in facial expressions is performed as follows. That is, for each frame image, a facial region is identified from the frame image, and the identified facial expressions are classified into a plurality of types according to an image analysis model machine-learned in advance. Then, based on the classification results, it analyzes whether positive facial expression changes occur between consecutive frame images, whether negative facial expression changes occur, and to what extent the facial expression changes occur, A facial expression change index value corresponding to the analysis result is output.
  • the analysis of changes in line of sight is performed as follows. That is, for each frame image, the eye region is specified in the frame image, and the orientation of both eyes is analyzed to analyze where the user is looking. For example, it analyzes whether the user is looking at the face of the speaker being displayed, whether the user is looking at the shared material being displayed, or whether the user is looking outside the screen. Also, it may be analyzed whether the eye movement is large or small, or whether the movement is frequent or infrequent. A change in line of sight is also related to the user's degree of concentration.
  • the biological reaction analysis unit 12 outputs a line-of-sight change index value according to the analysis result of the line-of-sight change.
  • the analysis of pulse changes is performed, for example, as follows. That is, for each frame image, the face area is specified in the frame image. Then, using a trained image analysis model that captures numerical values of face color information (G of RGB), changes in the G color of the face surface are analyzed. By arranging the results along the time axis, a waveform representing changes in color information is formed, and the pulse is identified from this waveform. When a person is tense, the pulse speeds up, and when the person is calm, the pulse slows down. The biological reaction analysis unit 12 outputs a pulse change index value according to the analysis result of the pulse change.
  • G of RGB face color information
  • analysis of changes in facial movement is performed as follows. That is, for each frame image, the face area is specified in the frame image, and the direction of the face is analyzed to analyze where the user is looking. For example, it analyzes whether the user is looking at the face of the speaker being displayed, whether the user is looking at the shared material being displayed, or whether the user is looking outside the screen. Further, it may be analyzed whether the movement of the face is large or small, or whether the movement is frequent or infrequent. The movement of the face and the movement of the line of sight may be analyzed together. For example, it may be analyzed whether the face of the speaker being displayed is viewed straight, whether the face is viewed with upward or downward gaze, or whether the face is viewed obliquely.
  • the biological reaction analysis unit 12 outputs a face orientation change index value according to the analysis result of the face orientation change.
  • Analysis of the contents of the statement is performed, for example, as follows. That is, the biological reaction analysis unit 12 converts the voice into a character string by performing known voice recognition processing on the voice for a specified time (for example, about 30 to 150 seconds), and morphologically analyzes the character string. By doing so, words such as particles and articles that are unnecessary for expressing conversation are removed. Then, vectorize the remaining words, analyze whether a positive emotional change has occurred, whether a negative emotional change has occurred, and to what extent the emotional change has occurred. Outputs the statement content index value.
  • Voice quality analysis is performed, for example, as follows. That is, the biological reaction analysis unit 12 identifies the acoustic features of the voice by performing known voice analysis processing on the voice for a specified time (for example, about 30 to 150 seconds). Then, based on the acoustic features, it analyzes whether a positive change in voice quality has occurred, whether a negative change in voice quality has occurred, and to what extent the change in voice quality has occurred, and according to the analysis results, output the voice quality change index value.
  • a specified time for example, about 30 to 150 seconds
  • the biological reaction analysis unit 12 uses at least one of the facial expression change index value, eye line change index value, pulse change index value, face direction change index value, statement content index value, and voice quality change index value calculated as described above. to calculate the biological reaction index value.
  • the biological reaction index value is calculated by weighting the facial expression change index value, eye line change index value, pulse change index value, face direction change index value, statement content index value, and voice quality change index value.
  • the peculiarity determination unit 13 determines whether or not the change in the analyzed biological reaction of the person to be analyzed is more specific than the change in the analyzed biological reaction of the person other than the person to be analyzed. In the present embodiment, the peculiarity determination unit 13 compares changes in the biological reaction of the person to be analyzed with those of others based on the biological reaction index values calculated for each of the plurality of users by the biological reaction analysis unit 12. is specific or not.
  • the peculiar determination unit 13 calculates the variance of the biological reaction index values calculated for each of the plurality of persons by the biological reaction analysis unit 12, and compares the biological reaction index values calculated for the analysis subject with the variance, It is determined whether or not the change in the analyzed biological reaction of the person to be analyzed is specific compared to others.
  • the following three patterns are conceivable as cases where the changes in biological reactions analyzed for the subject of analysis are more specific than those of others.
  • the first is a case where a relatively large change in biological reaction occurs in the subject of analysis, although no particularly large change in biological reaction has occurred in the other person.
  • the second is a case where a particularly large change in biological reaction has not occurred in the subject of analysis, but a relatively large change in biological reaction has occurred in the other person.
  • the third is a case where a relatively large change in biological reaction occurs in both the subject of analysis and the other person, but the content of the change differs between the subject of analysis and the other person.
  • the related event identification unit 14 identifies an event occurring in relation to at least one of the person to be analyzed, the other person, and the environment when the change in the biological reaction determined to be peculiar by the peculiarity determination unit 13 occurs. .
  • the related event identification unit 14 identifies from the moving image the speech and behavior of the person to be analyzed when a specific change in biological reaction occurs in the person to be analyzed.
  • the related event identifying unit 14 identifies, from the moving image, the speech and behavior of the other person when a specific change in the biological reaction of the person to be analyzed occurs.
  • the related event identification unit 14 identifies from the moving image the environment in which a specific change in the biological reaction of the person to be analyzed occurs.
  • the environment is, for example, the shared material being displayed on the screen, the background image of the person to be analyzed, and the like.
  • the clustering unit 15 clusters the change in the biological reaction determined to be specific by the peculiarity determination unit 13 (for example, one or a combination of eye gaze, pulse, facial movement, statement content, and voice quality), and the peculiarity Analyzing the degree of correlation with an event (event identified by the related event identification unit 14) that occurs when a change in biological reaction occurs, and if it is determined that the correlation is at a certain level or more , to cluster the subjects or events based on the correlation analysis results.
  • the peculiarity determination unit 13 for example, one or a combination of eye gaze, pulse, facial movement, statement content, and voice quality
  • the clustering unit 15 clusters the person to be analyzed or the event into one of a plurality of pre-segmented categories according to the content of the event, the degree of negativity, the magnitude of the correlation, and the like.
  • the clustering unit 15 clusters the person to be analyzed or the event into one of a plurality of pre-segmented classifications according to the content of the event, the degree of positivity, the degree of correlation, and the like.
  • the analysis result notification unit 16 reports at least one of the changes in the biological reaction determined to be specific by the peculiar determination unit 13, the event identified by the related event identification unit 14, and the classification clustered by the clustering unit 15. , to notify the designator of the subject of analysis (the subject of analysis or the organizer of the online session).
  • the analysis result notification unit 16 recognizes that when a change in a specific biological reaction that is different from that of the other person occurs in the person to be analyzed (one of the three patterns described above; the same applies hereinafter), the analysis target is Notifies the person to be analyzed of his/her own behavior. This allows the person to be analyzed to understand that he/she has a different feeling from others when he or she performs a certain behavior. At this time, the person to be analyzed may also be notified of the change in the specific biological reaction identified for the person to be analyzed. Furthermore, the person to be analyzed may be further notified of the change in the biological reaction of the other person to be compared.
  • the words and deeds of the person to be analyzed performed without being particularly conscious of their usual emotions, or the words and deeds of the person to be analyzed consciously accompanied by certain emotions, and the emotions and behaviors that others received
  • the emotion held by the person to be analyzed is different from the feeling held by the person to be analyzed at the time
  • the person to be analyzed is notified of the speech and behavior of the person to be analyzed at that time.
  • the analysis result notification unit 16 notifies the organizer of the online session of the event occurring when the person to be analyzed undergoes a specific change in biological reaction that is different from that of the other person, together with the change in the specific biological reaction. to notify.
  • the organizer of the online session can know what kind of event affects what kind of emotional change as a phenomenon specific to the specified analysis subject. Then, it becomes possible to perform appropriate treatment on the person to be analyzed according to the grasped contents.
  • the analysis result notification unit 16 notifies the organizer of the online session of the event occurring when a specific change in biological reaction occurs in the analysis subject, which is different from that of others, or the clustering result of the analysis subject. do.
  • online session organizers can grasp behavioral tendencies peculiar to analysis subjects and predict possible future behaviors and situations, depending on which classification the specified analysis subjects have been clustered into. be able to. Then, it becomes possible to take appropriate measures for the person to be analyzed.
  • the biological reaction index value is calculated by quantifying the change in biological reaction according to a predetermined standard, and the analysis subject is analyzed based on the biological reaction index value calculated for each of the plurality of people.
  • the biological reaction analysis unit 12 analyzes the movement of the line of sight for each of a plurality of people and generates a heat map indicating the direction of the line of sight.
  • the peculiar determination unit 13 compares the heat map generated for the person to be analyzed by the biological reaction analysis unit 12 with the heat map generated for the other person, so that the change in the biological reaction analyzed for the person to be analyzed It is determined whether it is specific compared with the change in biological response analyzed for.
  • moving images of a video session are stored in the local storage of the user terminal 10, and the above analysis is performed on the user terminal 10.
  • the machine specs of the user terminal 10 it is possible to analyze the moving image information without providing it to the outside.
  • the video session evaluation system of this embodiment may include a moving image acquisition unit 11, a biological reaction analysis unit 12, and a reaction information presentation unit 13a as functional configurations.
  • the reaction information presentation unit 13a presents information indicating changes in biological reactions analyzed by the biological reaction analysis unit 12a, including participants not displayed on the screen.
  • the reaction information presenting unit 13a presents information indicating changes in biological reactions to an online session leader, moderator, or administrator (hereinafter collectively referred to as the organizer).
  • Hosts of online sessions are, for example, instructors of online classes, chairpersons and facilitators of online meetings, coaches of sessions for coaching purposes, and the like.
  • An online session host is typically one of the users participating in the online session, but may be another person who does not participate in the online session.
  • the organizer of the online session can also grasp the state of the participants who are not displayed on the screen in an environment where the online session is held by multiple people.
  • FIG. 6 is a block diagram showing a configuration example according to this embodiment. As shown in FIG. 6, in the video session evaluation system of the present embodiment, functions similar to those of the above-described first embodiment are given the same reference numerals, and explanations thereof may be omitted.
  • the system includes a camera unit that acquires images of a video session, a microphone unit that acquires audio, an analysis unit that analyzes and evaluates moving images, and information obtained by evaluating the acquired moving images.
  • an object generator for generating a display object (described below) based on the display; and a display for displaying both the moving image of the video session and the display object during execution of the video session.
  • the analysis unit includes the moving image acquisition unit 11, the biological reaction analysis unit 12, the peculiar determination unit 13, the related event identification unit 14, the clustering unit 15, and the analysis result notification unit 16, as described above.
  • the function of each element is as described above.
  • the object generation unit generates an object 50 representing the recognized face part and the above-mentioned Information 100 indicating the content of the analysis/evaluation performed is superimposed on the moving image and displayed.
  • the object 50 may identify and display all faces of a plurality of persons when the faces of the plurality of persons are moved in the moving image.
  • the object 50 is, for example, when the camera function of the video session is stopped at the other party's terminal (that is, it is stopped by software within the application of the video session instead of physically covering the camera). If the other party's face is recognized by the other party's camera, the object 50 or the object 100 may be displayed in the part where the other party's face is located. This makes it possible for both parties to confirm that the other party is in front of the terminal even if the camera function is turned off. In this case, for example, in a video session application, the information obtained from the camera may be hidden while only the object 50 or object 100 corresponding to the face recognized by the analysis unit is displayed. Also, the video information acquired from the video session and the information recognized by the analysis unit may be divided into different display layers, and the layer relating to the former information may be hidden.
  • the objects 50 and 100 may be displayed in all areas or only in some areas. For example, as shown in FIG. 8, it may be displayed only on the moving image on the guest side.
  • the device described in this specification may be realized as a single device, or may be realized by a plurality of devices (for example, cloud servers) or the like, all or part of which are connected via a network.
  • the control unit 110 and the storage 130 of each terminal 10 may be realized by different servers connected to each other via a network.
  • the system includes user terminals 10, 20, a video session service terminal 30 for providing an interactive video session to the user terminals 10, 20, and an evaluation terminal 40 for evaluating the video session
  • Variation combinations of the following configurations are conceivable.
  • (1) Processing everything only on the user terminal As shown in FIG. 9, by performing the processing by the analysis unit on the terminal that is performing the video session (although a certain processing capacity is required), the video session can be performed. Analysis/evaluation results can be obtained at the same time (in real time) as you are.
  • an analysis unit may be provided in an evaluation terminal connected via a network or the like.
  • the moving images acquired by the user terminal are shared with the evaluation terminal at the same time as or after the video session, and are analyzed and evaluated by the analysis unit in the evaluation terminal. Together with or separately from the moving image data (that is, information including at least analysis data) is shared with the terminal and displayed on the display unit.
  • FIGS. 10 to 25 are realized by using the functional configuration examples 1 to 3 described above and their combinations.
  • FIG. 10 A first embodiment of the present invention will be described with reference to FIGS. 10 and 11.
  • FIG. The system according to this embodiment roughly evaluates the degree of matching between people. For example, analyze the reaction of the other party, evaluate the peculiarity of each other (expressions that do not usually appear), peculiarities with your past (expressions that did not appear in the past even with the same partner), neutral Evaluate by comparing with the normal state.
  • this matching is effective for online sessions conducted by lecturers and students. Matching various types of lecturers with various types of students is also important for continuing the course.
  • the system includes a type determination unit that determines each person's type based on the evaluation result by the analysis unit described above, and a matching determination unit that determines the degree of matching.
  • the type determination unit determines (estimates) the type of each user by referring to a type database (type DB) in which evaluation results and types are associated in advance.
  • the matching determination unit refers to a matching database (matching DB) in which the degree of matching for each type is defined in advance, and quantifies the degree of matching using the previously defined degree of compatibility between the above-determined types. .
  • the construction of the matching DB can be exemplified by defining in advance the degree of matching between a lecturer who is good at bringing out conversations and a student who is not good at speaking.
  • the degree of matching may be determined using a learning model that has learned teacher data including the conversation between the two parties and the evaluation of the conversation between the two parties. In this case, it is possible to give feedback (whether or not the lecturer is suitable for the person, etc.) on the results of the lectures between the persons who are actually matched.
  • an online session for type determination may be conducted before the course starts to determine the type of the student in advance.
  • the system acquires a moving image of an online session performed for type determination (step S1101), and determines the type (step S1102). Subsequently, the determined student type and instructor type are temporarily matched (step S1103).
  • the conditions required for the lecturer are obtained in advance through questionnaires, such as "I like a gentle teacher", "I like a teacher with a good tempo", etc., and the desired type is specified from the results of the questionnaire. It may be left as is.
  • the system acquires such information as condition information (step S1104).
  • the system considers the condition information, corrects the primary match degree (step S1105), and provides the corrected match degree. As a result, even if the system determines that a strict instructor is suitable as a matching partner, if the student has a condition of "preferring a gentle personality", the primary match degree will be calculated. The matching degree of each instructor is corrected, and the "strict but gentle instructor” is selected as the optimum instructor.
  • FIG. 12 A second embodiment of the present invention will be described with reference to FIGS. 12 and 13.
  • FIG. the system according to the present embodiment is based on the feeling (evaluation value) obtained from the user, and whether or not the person is likely to express that feeling in the first place (for example, a person who originally smiles a lot tends to have a high happy score).
  • Accurate evaluation by considering comparison with base emotions and evaluating the magnitude of displacement when emotions are expressed (the degree of emotional expression differs between people with small reactions and those with large reactions) It is.
  • Graphs (a) to (c) of FIG. 12 show (a) raw data (time-series emotion score) of a certain user, and (b) standard deviation (standard deviation processing), and (c) the standard deviation is standardized (zscore conversion with an average of 0 and a variance of 1) (standardization processing).
  • the system Based on the evaluation value of (b), the system performs an evaluation that takes into account the differences in facial expressions and normal facial expressions for each user. For example, it is possible to solve the problem that a user who often smiles during normal operation inevitably has a high smile score (happy score).
  • the system Based on the evaluation of (c), the system performs an evaluation that considers the richness of emotional expression for each user. For example, it is possible to improve the problem caused by the difference between a person who laughs softly and a person who laughs loudly.
  • FIG. (a) and (b) of FIG. 13 are graphs of Happy scores (expressing happiness levels) of user a and user b, respectively. Comparing the average scores of user a and user b, it can be seen that user A has a higher average score. In other words, user A smiles more than user B, and the Happy score inevitably tends to be high. In addition, comparing the range of each emotion (ST_A and ST_B), it can be seen that user A has a greater range of emotional expression (that is, the magnitude of reaction) than user B does.
  • the present system can exhibit its function as a system (apparatus) that generates training data by performing standard deviation processing and standardization processing on evaluation results obtained from various moving images.
  • a third embodiment of the present invention will be described with reference to FIG.
  • the system according to the present embodiment allows the subject to annotate (label) the outline and evaluation results.
  • An online session between a lecturer (first user) and a student (second user) will be described below as an example.
  • the system analyzes and evaluates the moving images of the online sessions described above, and visualizes them by outputting the evaluation results of the students (second users) as graphs.
  • the instructor can add the presence/absence of the emotion to the graph and supplementary information at that time (situation, speech and behavior, action, partner's action information, etc.).
  • the user provides the situation at that time (information about the situation such as "the class was lively” and "the other party responded well") to the point (Lab_1) where the Happy score was high. can be associated.
  • the location (Lab — 2) where the Happy score was low with the situation at that time information on the situation, such as “some assignment was given” or “severe things were told”).
  • annotations may be made for sections (Lab_3 and Lab_4). In this case, it is possible to make annotations such as "time spent teaching a difficult unit” and "time to summarize at the end of class”.
  • a teaching data set may be generated from a set of graphs (plotted evaluation values) and annotations, and machine learning may be performed by the analysis unit.
  • a fourth embodiment of the present invention will be described with reference to FIG.
  • the system according to the present embodiment in outline, generates chart information that can serve as a student's emotion chart in an online session of a lecture between a lecturer and a student, and shares the chart information with the lecturer.
  • the contents of the chart include, for example, the average value of each emotion of the student, characteristics, habits, frequent facial expressions, balance and degree of emotional expression by radar chart, ranking of words spoken when depressed, smile
  • the information necessary to face the psychological state of the student further, such as the ranking of the words used when the question appeared.
  • the system displays the evaluation results from multiple perspectives (neutral, happiness, surprise, discomfort, anger, sadness, fear, etc.) on the dashboard as a list by the analysis unit described above. Displaying together with facial expression icons that symbolize each emotion makes it easier to intuitively understand.
  • the display may be, for example, the evaluation results of online sessions for one day, or may be displayed on a weekly or monthly basis.
  • the dashboard may also display a digest of the video when each emotion was expressed most strongly, or textual information on the remarks made at that time. Also, the frequency of words used (preferred phrase) may be calculated and the ranking may be displayed.
  • the system according to the present embodiment determines whether there is inappropriate expression in a conversation between parties (for example, remarks using a position such as power harassment). It detects. Detection methods include rule-based (identification of prohibited keywords and expression of negative emotions on the other side) and machine learning approaches.
  • this system detects whether or not NG keywords (inappropriate words for a boss's remarks) are included in words uttered by the boss user in an online session between a boss and his subordinates, and Obtain the subordinate's evaluation value at that time, and if the expression of negative emotions such as fear, anxiety, sadness, and anger rises beyond a predetermined range compared to before the word was said records the boss's remarks as inappropriate remarks.
  • Inappropriate remarks may be notified, for example, to the company's human resources department, etc., together with the digest video and the text of the remarks.
  • FIG. 17 A sixth embodiment of the present invention will be described with reference to FIGS. 17 and 18.
  • FIG. The system according to the present embodiment is roughly a so-called word cloud using utterances of moving images.
  • the system recognizes the voice of the acquired moving image, converts the words included in the recognized voice into a size corresponding to the utterance frequency, and displays them as text.
  • words to be displayed it is possible to extract words that are frequently used in the evaluation target video, or words that are not included in videos other than the evaluation target video (words unique to this video). words) may be extracted.
  • the system may change the color of the text according to the user's evaluation result when the word is uttered. For example, a high HAPPY score may be given a red letter, and a word spoken when a SAD score is high may be given a blue score.
  • the word cloud shown in Fig. 17 displays the word "study" in the center.
  • the word is selected, as shown in FIG. 18, the text of the conversation when the word was uttered is displayed.
  • a play button P is displayed together with the conversation, and when the play button P is selected, a digest of the moving image corresponding to the text is played.
  • FIG. 19 A seventh embodiment of the present invention will be described with reference to FIGS. 19 and 20.
  • FIG. The system according to the present embodiment evaluates the degree of fatigue based on both facial images and voices included in the outline and moving images.
  • the system includes a fatigue evaluation condition reading unit and a fatigue evaluation unit.
  • evaluating the degree of fatigue according to the present embodiment, the steady state of the user is stored, and the degree of fatigue is evaluated based on the range of emotional fluctuations in the steady state and the current range of emotional fluctuations. Note that the evaluation of the degree of fatigue is not limited to this, and may be evaluated based on the amount of change in the pitch of conversation in the steady state and the amount of change in the pitch of the current conversation.
  • Step S2000 when a moving image is acquired (step S2000), a pre-learned fatigue level evaluation model is read (step S2001), the fatigue level is evaluated (step S2002), and the fatigue level is notified ( Step S2010) is performed.
  • the evaluation information for the normal time is read (S2101), each element is cross-analyzed by comparing with the evaluation result of each emotion, and the fatigue level is notified (step S2010).
  • the system acquires moving images and recognizes the utterances of each user included in the moving images.
  • the system includes an object generation unit that generates an object that associates written statements with users and displays them in chronological order.
  • FIG. 21 is a diagram showing the objects of the conversation rally of users A to C.
  • FIG. 21 When the utterance is up, the utterance object P is plotted, and the utterance objects P adjacent in time are connected by the connector C.
  • the system further comprises evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the utterance.
  • a color corresponding to the evaluation value may be given to the utterance object P of the conversation rally. For example, whether another user followed after a user spoke in a natural manner, or whether another user followed after the previous user's overbearing remarks would affect the improvement method.
  • the system according to this embodiment generally includes a statement object generation unit that recognizes a statement for each user included in a moving image and plots a statement object corresponding to the statement in association with the user.
  • system further comprises evaluation means for calculating an evaluation value from a predetermined point of view based on both the recognized face image and the utterance.
  • a color corresponding to the evaluation value may be assigned to the utterance object P.
  • FIG. 23 A tenth embodiment of the present invention will be described with reference to FIGS. 23 and 24.
  • FIG. The system according to the present embodiment generally includes intonation acquisition means for extracting intonation information of speech acquired from a moving image, and evaluation means for calculating an evaluation value from a predetermined viewpoint based on both the recognized face image and the intonation information. and
  • the intonation acquisition means may extract changes in pitch of speech per unit time.
  • FIG. 23 is a graph showing the standard deviation of intervals (pitch).
  • FIG. 24 is a graph of the standard deviation of volume. In this embodiment, the standard deviation is acquired for a predetermined number of frames from audio data of a predetermined sampling rate.
  • t2 and t3 in FIG. 23 are relatively monotonous, but t1 and t4 have large changes. Therefore, it can be estimated that the conversation is not lively at t1 and t4.
  • t2 and t3 which correspond to the same time axis as in FIG. 23, change little and are relatively monotonous.
  • t1 and t4 are relatively strong and weak. From this, it can be estimated that the conversation is not so lively during the time periods t2 and t3, while the conversation is lively during the time periods t1 and t4, both in terms of pitch and volume.
  • This system includes a context reading unit that reads context information and a correction unit that corrects evaluation results according to the context information.
  • the context information includes, for example, situation, number of conversations, acquaintance with the other party, one-way conversation style or two-way conversation style, etc. Category information classified by context, items to be corrected, and correction parameters. It may be prepared.
  • the system may accept contextual category information from the user in advance, or may automatically determine from the titles and metadata of video files and the like. Accordingly, by specifying the context information of the moving image and performing the correction associated with the corresponding category, it is possible to provide an appropriate evaluation result.

Abstract

Le problème décrit par la présente invention concerne l'évaluation d'une session en ligne par l'évaluation de la vidéo acquise au cours de la session en ligne. La solution selon la présente invention concerne un système comprenant : un moyen d'acquisition qui acquiert une vidéo se rapportant à une session en ligne entre un premier utilisateur et un deuxième utilisateur ; un moyen de reconnaissance de visage qui reconnaît au moins une image de visage du premier utilisateur incluse dans la vidéo pour chaque trame prescrite ; un moyen de reconnaissance de la parole qui reconnaît au moins la parole du deuxième utilisateur incluse dans la vidéo ; un moyen d'évaluation qui calcule une valeur d'évaluation à partir de perspectives multiples sur la base à la fois de l'image de visage reconnue ; un moyen de détection de mot-clé qui détecte un mot-clé prescrit dans au moins la parole reconnue ; et un moyen de transmission d'alerte qui envoie une alerte prescrite sur la base de la valeur d'évaluation lorsque le mot-clé est détecté.
PCT/JP2021/007572 2021-02-26 2021-02-26 Terminal, système et programme d'évaluation de session vidéo WO2022180855A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2021/007572 WO2022180855A1 (fr) 2021-02-26 2021-02-26 Terminal, système et programme d'évaluation de session vidéo
JP2023502015A JPWO2022180855A1 (fr) 2021-02-26 2021-02-26

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/007572 WO2022180855A1 (fr) 2021-02-26 2021-02-26 Terminal, système et programme d'évaluation de session vidéo

Publications (1)

Publication Number Publication Date
WO2022180855A1 true WO2022180855A1 (fr) 2022-09-01

Family

ID=83048747

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/007572 WO2022180855A1 (fr) 2021-02-26 2021-02-26 Terminal, système et programme d'évaluation de session vidéo

Country Status (2)

Country Link
JP (1) JPWO2022180855A1 (fr)
WO (1) WO2022180855A1 (fr)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018068618A (ja) * 2016-10-28 2018-05-10 株式会社東芝 感情推定装置、感情推定方法、感情推定プログラム、および感情カウントシステム
JP2020123204A (ja) * 2019-01-31 2020-08-13 株式会社日立システムズ 有害行為検出システムおよび方法

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018068618A (ja) * 2016-10-28 2018-05-10 株式会社東芝 感情推定装置、感情推定方法、感情推定プログラム、および感情カウントシステム
JP2020123204A (ja) * 2019-01-31 2020-08-13 株式会社日立システムズ 有害行為検出システムおよび方法

Also Published As

Publication number Publication date
JPWO2022180855A1 (fr) 2022-09-01

Similar Documents

Publication Publication Date Title
WO2022180860A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022168180A1 (fr) Terminal d'évaluation de session vidéo, système d'évaluation de session vidéo et programme d'évaluation de session vidéo
WO2022168185A1 (fr) Terminal d'évaluation de session vidéo, système d'évaluation de session vidéo et programme d'évaluation de session vidéo
JP7152825B1 (ja) ビデオセッション評価端末、ビデオセッション評価システム及びビデオセッション評価プログラム
WO2022180855A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180858A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180854A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180861A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180859A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180853A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180856A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180857A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180862A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022180852A1 (fr) Terminal, système et programme d'évaluation de session vidéo
WO2022249462A1 (fr) Système d'analyse vidéo
JP7197955B1 (ja) ビデオミーティング評価端末
WO2022201272A1 (fr) Programme d'analyse vidéo
JP7152819B1 (ja) 動画像分析プログラム
WO2022168174A1 (fr) Terminal d'évaluation de session vidéo, système d'évaluation de session vidéo et programme d'évaluation de session vidéo
WO2022230155A1 (fr) Système d'analyse vidéo
WO2022230136A1 (fr) Système d'analyse vidéo
WO2022168184A1 (fr) Terminal d'évaluation, système d'évaluation et programme d'évaluation de session vidéo
WO2022264220A1 (fr) Système d'analyse vidéo
WO2022269801A1 (fr) Système d'analyse vidéo
WO2022201267A1 (fr) Programme d'analyse vidéo

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21927954

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2023502015

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21927954

Country of ref document: EP

Kind code of ref document: A1