WO2022123688A1 - Scenario control method, scenario control device, and scenario control program - Google Patents

Scenario control method, scenario control device, and scenario control program Download PDF

Info

Publication number
WO2022123688A1
WO2022123688A1 PCT/JP2020/045856 JP2020045856W WO2022123688A1 WO 2022123688 A1 WO2022123688 A1 WO 2022123688A1 JP 2020045856 W JP2020045856 W JP 2020045856W WO 2022123688 A1 WO2022123688 A1 WO 2022123688A1
Authority
WO
WIPO (PCT)
Prior art keywords
scenario
audience
state
presentation
correction
Prior art date
Application number
PCT/JP2020/045856
Other languages
French (fr)
Japanese (ja)
Inventor
充裕 後藤
済央 野本
哲 小橋川
史朗 小澤
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Priority to PCT/JP2020/045856 priority Critical patent/WO2022123688A1/en
Publication of WO2022123688A1 publication Critical patent/WO2022123688A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/66Transforming electric information into light information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present invention relates to a scenario control method, a scenario control device, and a scenario control program.
  • a presenter system that automatically performs a presentation by combining a slide and a presenter agent is known (see Patent Documents 1 and 2).
  • the utterance content of each slide constituting the presentation is created as a scenario.
  • non-verbal movements such as face orientation and arm movement are also created as scenarios.
  • the presenter agent appeals the slide content to the audience by uttering the slide content with gestures and displaying the utterance script on the screen. ..
  • Patent Document 1 in the above presenter system, a scenario is created while assuming the audience state in advance. Therefore, if there is a mistake in the assumption of the audience status, even if the presenter agent is controlled according to the scenario, the appealing power of the presentation is low.
  • Patent Document 2 the content of the presentation is dynamically changed according to the state of the audience.
  • it is difficult to apply to an online conference environment because it is intended for an audience in a real space environment.
  • the present invention has been made in view of the above circumstances, and an object of the present invention is to provide a technique capable of improving the appealing power of a presentation in an online conference environment.
  • the scenario control method of one aspect of the present invention is a scenario control method for changing the scenario in a presenter system that controls the operation of the presenter agent based on the scenario to present the slide contents, and is acquired during the presentation at an online conference.
  • the first step of estimating the current audience state regarding the interest and attitude of the entire audience toward the presentation, and the current present.
  • the second step of determining the ideal audience status that raises the audience's interest in the presentation and concentrates the attitude of the entire audience, and the current present.
  • a third step of determining a correction scenario for transitioning the audience state to the ideal audience state, and a fourth step of changing the scenario to the correction scenario are performed.
  • the scenario control device of one aspect of the present invention is a scenario control device that changes the scenario in a presenter system that controls the operation of the presenter agent based on the scenario to present the slide contents, and is acquired during the presentation at an online conference.
  • an estimation unit that estimates the current audience state regarding the interest and attitude of the entire audience toward the presentation, and the current audience state.
  • the first decision-making part that determines the ideal audience state that raises the interest of the entire audience in the presentation and concentrates the attitude of the entire audience, and the current audience.
  • a second determination unit for determining a correction scenario for transitioning the state to the ideal audience state, and a change unit for changing the scenario to the correction scenario are provided.
  • the scenario control program of one aspect of the present invention is a scenario control program that causes a computer to execute the above scenario control method.
  • FIG. 1 is a diagram showing an overall configuration of a presenter system.
  • FIG. 2 is a diagram showing a scenario example of each slide.
  • FIG. 3 is a diagram showing an example of a state transition rule.
  • FIG. 4 is a diagram showing an example of a scenario correction rule.
  • FIG. 5 is a diagram showing an example of scenario correction history information.
  • FIG. 6 is a flow chart showing a characteristic processing operation of the scenario control device.
  • FIG. 7 is a diagram showing images of facial expression analysis and chat posting analysis.
  • FIG. 8 is a diagram showing an example of an audience state.
  • FIG. 9 is a diagram showing an example of determining the audience state of the transition destination.
  • FIG. 10 is a diagram showing an example of a correction scenario between each state.
  • FIG. 1 is a diagram showing an overall configuration of a presenter system.
  • FIG. 2 is a diagram showing a scenario example of each slide.
  • FIG. 3 is a diagram showing an example of a state transition rule.
  • FIG. 11 is a diagram showing an example of a correction scenario at the time of transition to the same state.
  • FIG. 12 is a flow chart showing a specific example of the processing operation of the scenario control device.
  • FIG. 13 is a flow chart showing a specific example of the selection process of the correction scenario.
  • FIG. 14 is a flow chart showing a specific example of the scenario change process.
  • FIG. 15 is a diagram showing a modified example of the scenario.
  • FIG. 16 is a diagram showing a modified example of the presenter system.
  • FIG. 17 is a diagram showing a hardware configuration example of the scenario control device.
  • the present invention applies a presenter system by combining slides and a presenter agent to an online conference system, and dynamically creates a scenario (speech content, non-verbal movement) for controlling the operation of the presenter agent acting on behalf of a person's presentation. It is a technology that improves the appeal of presentations in an online conference environment by changing it.
  • the behavior of the presenter agent is dynamically changed according to the current audience status of the audience participating in the online conference.
  • the scenario for controlling the behavior of the presenter agent is dynamically changed according to the audience state (audience interest / interest, audience attitude).
  • the present invention creates an ideal audience state in which the audience has a high level of interest and can concentrate on and understand the main points of the presentation according to the current audience state and the progress of the scenario (progress of the presentation). Change the scenario so that it transitions. Further, in the present invention, the current audience state is determined using the audience face information (face orientation, facial expression, etc.) and the chat posted content, and based on the determination result, the progress of the scenario and the utterance timing are determined. Determine the ideal audience condition.
  • FIG. 1 is a diagram showing an overall configuration of a presenter system according to the present embodiment.
  • the presenter system acquires and predefines the user terminal 6 used by the audience of the online conference, the presentation control device 5 that controls the display of slides and the speech / operation of the presenter agent 100 based on the scenario, and the audience state. It includes a scenario control device 1 that changes a running scenario based on a rule.
  • the user terminal 6, the presentation control device 5, and the scenario control device 1 are connected to a communication network capable of intercommunication.
  • the user terminal 6 is an information processing terminal used by an audience participating in an online conference.
  • the user terminal 6 includes a Web camera, a display, a data communication function, a data input / output function, and the like.
  • a slide for presentation, a presenter agent 100 for presenting the slide contents, and the like are displayed on the display of the user terminal 6.
  • the number of user terminals 6 is at least one.
  • the user terminal 6 may be a personal computer provided by a person concerned or a third party for an online conference, or may be a mobile terminal prepared by the audience.
  • the presentation control device 5 includes, for example, a scenario execution unit 50, a scenario storage unit 51, a scenario notification unit 52, and an instruction control unit 53.
  • the scenario execution unit 50 has a function of reading a scenario from the scenario storage unit 51 and executing the scenario. Further, the scenario execution unit 50 has a function of executing the changed scenario when the scenario is changed by the scenario control device 1.
  • the scenario storage unit 51 has a function of storing a scenario for controlling the operation of the presenter agent 100.
  • a scenario for controlling the operation of the presenter agent 100.
  • utterance contents for explaining each slide by voice and non-verbal movements such as facial expression, face orientation, and arm movement are set for each slide. ing.
  • the scenario notification unit 52 has a function of notifying the transition destination determination unit 40 of the scenario control device 1 of the scenario execution status (scenario progress, presentation progress).
  • the instruction control unit 53 has a function of controlling the operation (utterance content, non-verbal operation) of the presenter agent 100 displayed on the user terminal 6 based on the execution of the scenario by the scenario execution unit 50.
  • the scenario control device 1 is a device that changes the scenario being executed by the presentation control device 5 while estimating the current audience state of the audience watching the presentation in the online conference in real time from the audience face image and the posted chat content. Is.
  • the scenario control device 1 includes, for example, a face image acquisition unit 10, a face image determination unit 11, a facial expression analysis information storage unit 12, a chat acquisition unit 20, and a chat analysis unit 21. , Audience state estimation unit 30, transition destination determination unit 40, state transition rule storage unit 41, correction content selection unit 42, scenario correction rule storage unit 43, scenario correction unit 44, and correction history storage unit 45. , Equipped with.
  • the face image acquisition unit 10 has a function of receiving an audience face image taken during an online conference from the user terminal 6.
  • the face image determination unit 11 reads facial expression analysis information from the facial expression analysis information storage unit 12, analyzes the facial expression and face orientation of the audience in the audience facial image using the facial expression analysis information, and makes a presentation. It has a function to analyze the interests and trends of interests of the entire audience.
  • the facial expression analysis information storage unit 12 has a function of storing facial expression analysis information in which an image feature amount for determining whether the audience is in a positive state or a negative state is described.
  • the chat acquisition unit 20 has a function of receiving chat data posted by an audience during an online conference from a user terminal 6.
  • the chat analysis unit 21 has a function of analyzing the contents of chat data and analyzing the tendency of the attitude of the entire audience toward the presentation.
  • the audience state estimation unit 30 is based on the tendency of the entire audience's interest / interest by the face image determination unit 11 and the tendency of the entire audience's attitude by the chat analysis unit 21, and the audience's overall interest / interest and attitude toward the presentation. It has a function to estimate and calculate the current audience status of.
  • the transition destination determination unit 40 reads the state transition rule from the state transition rule storage unit 41, and uses the state transition rule to notify the current audience state and the scenario execution status notified from the scenario notification unit 52 of the presentation control device 5. Based on the above, it has a function to determine the ideal transition destination audience state that can improve the interest / interest and attitude of the entire audience.
  • the state transition rule storage unit 41 has a function of storing a state transition rule for determining the audience state of the transition destination. As illustrated in FIG. 3, the state transition rule is set with a scenario execution status and a transition destination state indicating the audience state of the transition destination.
  • the modification content selection unit 42 reads a scenario modification rule from the scenario modification rule storage unit 43, and selects a modification scenario used to transition the current audience state to the transition destination audience state from the scenario modification rule. Be prepared. Further, the correction content selection unit 42 reads the scenario correction history information from the correction history storage unit 45, refers to the scenario correction history information, and if the same state transition occurs a plurality of times, the same state transition is used in the past. It has a function to select a correction scenario different from the selected correction scenario.
  • the scenario correction rule storage unit 43 has a function of storing a scenario correction rule in which a correction scenario for transitioning the current audience state to the transition destination audience state is set.
  • the scenario modification rule includes the ID of the scenario for modification, the current state indicating the current audience state, the transition destination state indicating the audience state of the transition destination, the scenario modification content, and the modification target. And are set.
  • the modification target is set to one or both of the utterance content and the non-verbal behavior that make up the scenario.
  • the scenario correction unit 44 has a function of changing the scenario stored in the scenario storage unit 51 of the presentation control device 5 and the scenario read from the scenario storage unit 51 by the scenario execution unit 50 into a correction scenario.
  • the correction history storage unit 45 has a function of storing the number of times the correction scenario selected from the scenario correction rule is used as scenario correction history information. As illustrated in FIG. 5, the scenario correction history information stores the ID of the correction scenario and the number of times the correction scenario is used.
  • the present invention estimates the audience state of an audience who participates in an online conference system and is watching a presentation in real time, and dynamically changes the scenario of the presenter agent based on the estimated current audience state. Achieve a highly appealing presentation.
  • the structural features of the scenario control device 1 shown in FIG. 1 include the face image determination unit 11 and the chat analysis unit 21 that calculate the characteristics of the audience state, and the current audience state regarding the interest, interest, and attitude of the entire audience.
  • Audience state estimation unit (estimation unit) 30 that estimates the current audience state
  • a transition destination determination unit (first) that determines what state the current audience state should be transitioned to based on the current audience state and the scenario execution status.
  • FIG. 6 is a flow chart showing a characteristic processing operation of the scenario control device 1.
  • Step S1 the face image determination unit 11 analyzes the positive / negative state of each audience from the facial expressions and face orientations of each audience face image acquired from the plurality of user terminals 6, respectively, and positive / negative of the entire audience. The state is determined (see FIG. 7).
  • the chat analysis unit 21 analyzes the concentration / divergence state of each audience from the posted content and the amount of sentences of each chat posted by each audience during the online conference, and determines the concentration / divergence state of the entire audience. Judgment (see FIG. 7).
  • the chat analysis unit 21 acquires chats posted at regular intervals. Next, the chat analysis unit 21 calculates the amount of posted text Cl j (1 ⁇ j ⁇ M) for each acquired chat j and the degree of similarity Cs j between the chat j and the slide content.
  • the posted sentence amount Cl j is a value (0 ⁇ Cl ⁇ 1) obtained by normalizing the number of content words included in the chat j with the number of words in all chats acquired at regular intervals.
  • the similarity Cs j is a value (0 ⁇ Cs ⁇ 1) indicating the degree of similarity between the content word of the chat j and the content word of the question target slide.
  • the chat analysis unit 21 calculates the score S j for each chat j from the product of the calculated posted sentence amount Cl j and the similarity Cs j , and totals all the scores at regular intervals from the equation (4). Obtain the SC all . Finally, from the equation (5), the chat analysis unit 21 determines that the entire audience is in the concentrated state when SC all > threshold Th, and in other cases, the entire audience is determined to be in the divergent state. The larger the amount of posted text Cl j and the higher the degree of similarity Cs j with the slide, the more concentrated listening is represented.
  • Step S2 the audience state estimation unit 30 estimates and calculates the current audience state regarding the entire audience's interest / interest and attitude toward the presentation based on the positive / negative state determination result and the concentration / divergence state determination result. do.
  • the audience state estimation unit 30 classifies the interest and interest of the entire audience into high (positive state) and low (negative state), and the attitude of the entire audience is concentrated (concentrated state) and divergent (divergent state). ),
  • the current audience state is estimated from the four audience states (see FIG. 8). If it is a positive / concentrated state, it is a state 1 in which the person is interested in the presentation and wants to hear the main points and details of the presentation. If it is a positive / divergent state, it is assumed that the person is interested in the presentation but does not know the main points and details of the presentation. In the negative / concentrated state, the interest of the presentation is low, but the state 3 in which the presentation is obligatory to be listened to. If it is a negative / divergent state, it is assumed that the person is not interested in the presentation and is in a state of being distracted other than the presentation 4.
  • all of interest / interest and attitude are binary, but a multi-value other than the binary may be used.
  • a multi-value other than the binary may be used.
  • Step S3; Next, the transition destination determination unit 40 is an ideal transition destination that raises the interest and interest of the entire audience in the presentation and concentrates the attitude of the entire audience based on the current audience status and the scenario execution status. Determine the audience status.
  • the transition destination determination unit 40 determines any of the states 1 to 3 as the transition destination audience state.
  • state 1 as the transition destination audience state
  • the scenario execution status is the first half of the scenario (immediately after the start of the presentation)
  • the state to state 2 is to increase the interest and interest of the entire audience in the presentation.
  • Priority is given to the transition, and then the state transition to state 1 is performed.
  • the attitude shifts to a focused attitude (see FIGS. 9 (a) and 9 (b)).
  • the state 3 is given priority first in order to concentrate the attitude of the entire audience toward the presentation, and then the state transition to the state 1 is performed. In other words, after having a focused attitude, the interest / interest is increased (see FIGS. 9 (a) and 9 (c)).
  • Step S4 the modification content selection unit 42 determines a modification scenario for transitioning the current audience state to the transition destination audience state. For example, the modification content selection unit 42 acquires the modification scenario corresponding to the current audience state and the transition destination audience state from the scenario modification rule. At this time, the modification content selection unit 42 updates the number of times the modification scenario is used.
  • the correction scenario describes transition conditions such as “repeat the same explanation” and "eye contact operation” between the states of the state transition diagram shown in FIG. Transition conditions are set according to each state transition of "state transition source-state transition destination".
  • Nonverbal movements are physical movements such as facial expressions, face orientation, and arm positions / directions.
  • the utterance content is sound information / text information such as the utterance content uttered by the presenter agent, the utterance script, the intonation / interval of the voice, and the sound effect at the same time as the voice.
  • the correction content selection unit 42 refers to the number of times the correction scenario is used, and selects a correction scenario different from the past correction scenario so as not to make a uniform scenario correction. For example, as illustrated in FIG. 11, when the state transition from the state 4 to the state 3 is performed for the second time, a “pointing operation” different from the first “attention guidance to the slide” is selected.
  • Step S5 Finally, the scenario correction unit 44 changes the scenario being executed by the presentation control device 5 to the correction scenario.
  • FIG. 12 is a flow chart showing a specific example of the processing operation of the scenario control device 1.
  • Step S101 First, the face image acquisition unit 10 receives each audience face image from the plurality of user terminals 6. Then, the face image determination unit 11 determines the positive / negative state of each audience based on the facial expression analysis information of the facial expression analysis information storage unit 12, and calculates the positive / negative state of the entire audience for the presentation.
  • Step S102 Next, the chat acquisition unit 20 receives the chat data of the chat exchanged on the online tool from the user terminal 6. Then, the chat analysis unit 21 calculates the amount of posted text of the chat data and the degree of similarity between the posted content and the slide content, and calculates the concentration / divergence state of the entire audience with respect to the presentation.
  • Step S103 the audience state estimation unit 30 determines the current audience state regarding the audience's interest / interest and attitude toward the presentation based on the positive / negative state of the entire audience and the tendency of the concentration / divergence state of the entire audience. Estimate calculation. For example, one state is determined from the four states 1 to 4 illustrated in FIG.
  • Step S104 the transition destination determination unit 40 determines whether or not the state transition is necessary based on the estimation result of the current audience state.
  • the current audience state is state 1
  • the entire audience is watching the presentation with interest and interest, and it is determined that the state transition is unnecessary and the process is terminated.
  • the current audience state is the state 2 to the state 4, it is determined that the state transition is necessary, and the process proceeds to step S105.
  • Step S105 the transition destination determination unit 40 determines whether or not there is one candidate for the transition destination state to be transitioned from the current audience state. If the current audience state is state 4, the transition destination states are a plurality of states 1 to 3, and the process proceeds to step S106. If the current audience state is state 2 or state 3, the transition destination state is only state 1, and the process proceeds to step S107.
  • Step S106 When there are a plurality of candidates for the transition destination state, the transition destination determination unit 40 acquires the scenario execution status from the presentation control device 5 and either is the first half of the presentation (from the start to the policy change position) or the second half of the presentation (policy). Based on whether it ends from the conversion position), the corresponding transition destination state is read from the state transition rule (FIG. 3) of the state transition rule storage unit 41, and this is used as a candidate for the transition destination state.
  • the state transition rule FOG. 3
  • Step S107 the transition destination determination unit 40 determines the candidate for one transition destination state grasped in step S105 or the candidate for the transition destination state grasped in step S106 as the transition destination audience state.
  • Step S108 the correction content selection unit 42 selects all the correction scenarios that are candidates from the scenario correction rule (FIG. 4) of the scenario correction rule storage unit 43 based on the current audience state and the transition destination audience state.
  • the final correction scenario is selected so as not to overlap with the past correction results recorded in the scenario correction history information (FIG. 5) of the read and correction history storage unit 45 as much as possible.
  • a specific example of the correction scenario selection process will be described later.
  • Step S109 Finally, the scenario correction unit 44 changes the scenario being executed by the presentation control device 5 with the correction scenario, and updates the scenario correction history information of the correction history storage unit 45.
  • a specific example of the scenario change process will also be described later.
  • FIG. 13 is a flow chart showing a specific example of the selection process of the correction scenario.
  • Step S108-1 First, the correction content selection unit 42 corrects all the correction scenarios corresponding to the combination of the current audience state and the transition destination audience state from the scenario correction rule (FIG. 4) of the scenario correction rule storage unit 43. Get as a candidate for.
  • Step S108-2 Next, the modification content selection unit 42 determines whether or not the number of candidates for the modification scenario is one. If the number of candidates for the modification scenario is one, the process proceeds to step S108-6. If there are a plurality of candidates for the correction scenario, the process proceeds to step S108-3.
  • Step S108-3 When there are a plurality of candidates for the correction scenario, the correction content selection unit 42 uses the scenario correction history information (FIG. 5) of the correction history storage unit 45 to indicate that each candidate's correction scenario has been used so far. Get the number of uses respectively.
  • Step S108-4 After step S108-3, the modification content selection unit 42 discards the candidate whose number of uses is not the minimum among the candidates for the plurality of modification scenarios, and selects the candidate whose number of uses is the smallest.
  • Step S108-5 After step S108-4, the correction content selection unit 42 determines whether or not the number of remaining correction scenario candidates is one. If the number of remaining correction scenario candidates is one, the process proceeds to step S108-6. If the number of remaining correction scenario candidates is a plurality, the process proceeds to step S108-7.
  • Step S108-6 When it is determined in step S108-2 or step S108-5 that the number of candidates for the modification scenario is one, the modification content selection unit 42 selects the candidate for the modification scenario as the modification scenario to be changed.
  • Step S108-7 When it is determined in step S108-5 that the number of candidates for the correction scenario is a plurality, the correction content selection unit 42 randomly changes one candidate from the remaining candidates for the correction scenario. Select as.
  • FIG. 14 is a flow chart showing a specific example of the scenario change process.
  • Step S109-1 First, the scenario correction unit 44 acquires the correction target of the correction scenario selected in step S108 from the scenario correction rule (FIG. 4) of the scenario correction rule storage unit 43.
  • Step S109-2 Next, the scenario correction unit 44 determines the type of the acquired correction target. If the type of the modification target is non-verbal operation, the process proceeds to step S109-3. If the type of the correction target is the utterance content, the process proceeds to step S109-4. If the types of the modification targets are all (both non-verbal actions and utterance contents), the process proceeds to step S109-5.
  • Step S109-3 When the type of the modification target is non-verbal operation, the scenario modification unit 44 replaces the non-language operation set in the scenario before modification with the non-language operation of the scenario modification content, or the non-language operation thereof. Add the scenario correction contents to.
  • Step S109-4 When the type of the correction target is the utterance content, the scenario correction unit 44 adds the utterance content of the scenario correction content to the utterance content set in the scenario before the correction.
  • Step S109-5 When the types of modification targets are all, the scenario modification unit 44 changes the running scenario so as to repeat the explanation of the same slide.
  • An example of changing the scenario is shown in FIG. FIG. 15A is an operation example for raising interest / interest.
  • a sound effect is added to the utterance content.
  • the utterance content is emphasized by repeating the same utterance content. This makes it possible to increase the interest and interest of the audience in the slide contents.
  • FIG. 15 (b) is an operation example for promoting an attitude.
  • the important points of the slide are pointed with a hand or a finger.
  • making eye contact with the audience gives the audience the impression of speaking and encourages them to listen more intensively.
  • FIG. 15 (b-3) by paying attention to the important points of the slide and guiding the attention to the slide, the slide contents are carefully urged to listen to the explanation. This can enhance the audience's attitude towards the slide content.
  • the scenario correction unit 44 updates the scenario correction history information of the correction history storage unit 45 regarding the correction scenario that was the change target. For example, the scenario correction unit 44 increments the number of times the correction scenario is used.
  • the processing operation of the scenario control device 1 has been explained above.
  • the scenario control device 1 periodically executes the above processing operation, estimates the current audience state, raises the interest and interest of the entire audience in the presentation, and concentrates the attitude of the entire audience. Change the scenario.
  • the presentation control device 5 dynamically changes the scenario in the above-mentioned modified scenario while the scenario execution unit 50 instructs the instruction control unit 53 to control the scenario based on the scenario. Further, the presentation control device 5 notifies the scenario control device 1 of the scenario execution status required for scenario correction one by one via the scenario notification unit 52.
  • the scenario control device 1 estimates the audience state during the remote presentation on the online conference system, and changes the scenario based on the estimation result. This makes it possible to provide a more appealing presentation in the presenter system in the remote presentation.
  • the presenter system in which the presenter agent is operated in the virtual space in the computer has been described as a premise (see FIG. 16A).
  • the entity of the presenter agent may be either real space or virtual space.
  • a tangible presenter agent 100 may be used. In this case, the presenter agent 100 is present for each audience.
  • the present invention is not limited to the above embodiment.
  • the present invention can be modified in a number of ways within the scope of the gist of the present invention.
  • the scenario control device 1 of the present embodiment is, for example, as shown in FIG. 17, a CPU (Central Processing Unit, processor) 901, a memory 902, a storage (Hard Disk Drive, Solid State Drive) 903, and a communication device 904. And, it can be realized by using a general-purpose computer system including an input device 905 and an output device 906.
  • the memory 902 and the storage 903 are storage devices.
  • each function of the scenario control device 1 is realized by the CPU 901 executing a predetermined program loaded on the memory 902.
  • the scenario control device 1 may be mounted on one computer.
  • the scenario control device 1 may be implemented by a plurality of computers.
  • the scenario control device 1 may be a virtual machine mounted on a computer.
  • the program for the scenario control device 1 can be stored in a computer-readable recording medium such as an HDD, SSD, USB (Universal Serial Bus) memory, CD (Compact Disc), or DVD (Digital Versatile Disc).
  • the scenario control program for the scenario control device 1 can also be distributed via the communication network.
  • Scenario control device 10 Face image acquisition unit 11: Face image determination unit 12: Face expression analysis information storage unit 20: Chat acquisition unit 21: Chat analysis unit 30: Audience state estimation unit 40: Transition destination determination unit 41: Status Transition rule storage unit 42: Correction content selection unit 43: Scenario correction rule storage unit 44: Scenario correction unit 45: Correction history storage unit 5: Presentation control device 50: Scenario execution unit 51: Scenario storage unit 52: Scenario notification unit 53: Instruction control unit 6: User terminal 7: Display panel 8: Camera 100: Presenter agent 901: CPU 902: Memory 903: Storage 904: Communication device 905: Input device 906: Output device

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The present invention provides a scenario control method for changing a scenario in a presenter system that presents slide contents by controlling the operation of a presenter agent on the basis of the scenario. This scenario control method comprises: a first step of estimating a current audience state regarding the interest and attitude of the entire audience toward a presentation in an online meeting on the basis of face information of the audience acquired during the presentation and the contents of posts posted by the audience during the presentation; a second step of determining an ideal audience state for increasing the interest of the entire audience toward the presentation and concentrating the attitude of the entire audience on the basis of the current audience state and the degree of progress of the scenario; a third step of determining a correction scenario for transitioning the current audience state to the ideal audience state; and a fourth step of changing the scenario to the correction scenario.

Description

シナリオ制御方法、シナリオ制御装置、および、シナリオ制御プログラムScenario control method, scenario control device, and scenario control program
 本発明は、シナリオ制御方法、シナリオ制御装置、および、シナリオ制御プログラムに関する。 The present invention relates to a scenario control method, a scenario control device, and a scenario control program.
 スライドとプレゼンタエージェントとを組み合わせ、自動でプレゼンを実施するプレゼンタシステムが知られている(特許文献1,2参照)。プレゼンタシステムでは、プレゼンタエージェントにプレゼン内容を聴衆へ伝達させるため、プレゼンを構成する各スライドの発話内容をシナリオとして作成する。また、発話内容や発話内容を文字化した発話スクリプトを強調・補足するため、顔の向きや腕の動きなどの非言語動作もシナリオとして作成する。発話内容と非言語動作とを含むシナリオに基づき、プレゼンタエージェントは、身振り手振りの動作を交えながら、スライド内容を発話し、さらには発話スクリプトを画面に表示することで、スライド内容を聴衆へ訴求する。 A presenter system that automatically performs a presentation by combining a slide and a presenter agent is known (see Patent Documents 1 and 2). In the presenter system, in order to let the presenter agent convey the content of the presentation to the audience, the utterance content of each slide constituting the presentation is created as a scenario. In addition, in order to emphasize and supplement the utterance content and the utterance script in which the utterance content is transcribed, non-verbal movements such as face orientation and arm movement are also created as scenarios. Based on a scenario that includes utterance content and non-verbal movement, the presenter agent appeals the slide content to the audience by uttering the slide content with gestures and displaying the utterance script on the screen. ..
特開2019-144732号公報JP-A-2019-144732 特開2020-86774号公報Japanese Unexamined Patent Publication No. 2020-86774
 特許文献1では、上記プレゼンタシステムにおいて、事前に聴衆状態を想定しながらシナリオを作成している。そのため、聴衆状態の想定に間違いがある場合、シナリオ通りにプレゼンタエージェントを制御しても、プレゼンの訴求力は低い。特許文献2では、聴衆状態に応じて動的にプレゼン内容を変化させている。しかし、実空間環境の聴衆を想定しており、オンライン会議環境への適用は難しい。 In Patent Document 1, in the above presenter system, a scenario is created while assuming the audience state in advance. Therefore, if there is a mistake in the assumption of the audience status, even if the presenter agent is controlled according to the scenario, the appealing power of the presentation is low. In Patent Document 2, the content of the presentation is dynamically changed according to the state of the audience. However, it is difficult to apply to an online conference environment because it is intended for an audience in a real space environment.
 本発明は、上記事情に鑑みてなされたものであり、本発明の目的は、オンライン会議環境でプレゼンの訴求力を改善可能な技術を提供することである。 The present invention has been made in view of the above circumstances, and an object of the present invention is to provide a technique capable of improving the appealing power of a presentation in an online conference environment.
 本発明の一態様のシナリオ制御方法は、シナリオに基づきプレゼンタエージェントの動作を制御してスライド内容をプレゼンさせるプレゼンタシステムにおいて、前記シナリオを変更するシナリオ制御方法であって、オンライン会議でプレゼン中に取得した聴衆の顔情報と、前記プレゼン中に前記聴衆が投稿した投稿内容と、を基に、前記プレゼンに対する聴衆全体の興味および態度に関する現在の聴衆状態を推定する第1のステップと、前記現在の聴衆状態と、前記シナリオの進行度と、に基づき、前記プレゼンに対する聴衆全体の興味を高め、かつ、聴衆全体の態度を集中させる理想的な聴衆状態を決定する第2のステップと、前記現在の聴衆状態を前記理想的な聴衆状態に遷移させる修正用シナリオを決定する第3のステップと、前記シナリオを前記修正用シナリオに変更する第4のステップと、を行う。 The scenario control method of one aspect of the present invention is a scenario control method for changing the scenario in a presenter system that controls the operation of the presenter agent based on the scenario to present the slide contents, and is acquired during the presentation at an online conference. Based on the facial information of the audience and the content posted by the audience during the presentation, the first step of estimating the current audience state regarding the interest and attitude of the entire audience toward the presentation, and the current present. Based on the audience status and the progress of the scenario, the second step of determining the ideal audience status that raises the audience's interest in the presentation and concentrates the attitude of the entire audience, and the current present. A third step of determining a correction scenario for transitioning the audience state to the ideal audience state, and a fourth step of changing the scenario to the correction scenario are performed.
 本発明の一態様のシナリオ制御装置は、シナリオに基づきプレゼンタエージェントの動作を制御してスライド内容をプレゼンさせるプレゼンタシステムにおいて、前記シナリオを変更するシナリオ制御装置であって、オンライン会議でプレゼン中に取得した聴衆の顔情報と、前記プレゼン中に前記聴衆が投稿した投稿内容と、を基に、前記プレゼンに対する聴衆全体の興味および態度に関する現在の聴衆状態を推定する推定部と、前記現在の聴衆状態と、前記シナリオの進行度と、に基づき、前記プレゼンに対する聴衆全体の興味を高め、かつ、聴衆全体の態度を集中させる理想的な聴衆状態を決定する第1の決定部と、前記現在の聴衆状態を前記理想的な聴衆状態に遷移させる修正用シナリオを決定する第2の決定部と、前記シナリオを前記修正用シナリオに変更する変更部と、を備える。 The scenario control device of one aspect of the present invention is a scenario control device that changes the scenario in a presenter system that controls the operation of the presenter agent based on the scenario to present the slide contents, and is acquired during the presentation at an online conference. Based on the facial information of the audience and the content posted by the audience during the presentation, an estimation unit that estimates the current audience state regarding the interest and attitude of the entire audience toward the presentation, and the current audience state. And, based on the progress of the scenario, the first decision-making part that determines the ideal audience state that raises the interest of the entire audience in the presentation and concentrates the attitude of the entire audience, and the current audience. A second determination unit for determining a correction scenario for transitioning the state to the ideal audience state, and a change unit for changing the scenario to the correction scenario are provided.
 本発明の一態様のシナリオ制御プログラムは、上記シナリオ制御方法をコンピュータに実行させるシナリオ制御プログラムである。 The scenario control program of one aspect of the present invention is a scenario control program that causes a computer to execute the above scenario control method.
 本発明によれば、オンライン会議環境でプレゼンの訴求力を向上可能な技術を提供できる。 According to the present invention, it is possible to provide a technique capable of improving the appeal of a presentation in an online conference environment.
図1は、プレゼンタシステムの全体構成を示す図である。FIG. 1 is a diagram showing an overall configuration of a presenter system. 図2は、各スライドのシナリオ例を示す図である。FIG. 2 is a diagram showing a scenario example of each slide. 図3は、状態遷移ルールの例を示す図である。FIG. 3 is a diagram showing an example of a state transition rule. 図4は、シナリオ修正ルールの例を示す図である。FIG. 4 is a diagram showing an example of a scenario correction rule. 図5は、シナリオ修正履歴情報の例を示す図である。FIG. 5 is a diagram showing an example of scenario correction history information. 図6は、シナリオ制御装置の特徴的な処理動作を示すフロー図である。FIG. 6 is a flow chart showing a characteristic processing operation of the scenario control device. 図7は、顔表情分析およびチャット投稿分析のイメージを示す図である。FIG. 7 is a diagram showing images of facial expression analysis and chat posting analysis. 図8は、聴衆状態の例を示す図である。FIG. 8 is a diagram showing an example of an audience state. 図9は、遷移先の聴衆状態の決定例を示す図である。FIG. 9 is a diagram showing an example of determining the audience state of the transition destination. 図10は、各状態間の修正シナリオの例を示す図である。FIG. 10 is a diagram showing an example of a correction scenario between each state. 図11は、同一状態遷移時での修正シナリオの例を示す図である。FIG. 11 is a diagram showing an example of a correction scenario at the time of transition to the same state. 図12は、シナリオ制御装置の処理動作の具体例を示すフロー図である。FIG. 12 is a flow chart showing a specific example of the processing operation of the scenario control device. 図13は、修正用シナリオの選択処理の具体例を示すフロー図である。FIG. 13 is a flow chart showing a specific example of the selection process of the correction scenario. 図14は、シナリオ変更処理の具体例を示すフロー図である。FIG. 14 is a flow chart showing a specific example of the scenario change process. 図15は、シナリオの変更例を示す図である。FIG. 15 is a diagram showing a modified example of the scenario. 図16は、プレゼンタシステムの変形例を示す図である。FIG. 16 is a diagram showing a modified example of the presenter system. 図17は、シナリオ制御装置のハードウェア構成例を示す図である。FIG. 17 is a diagram showing a hardware configuration example of the scenario control device.
 以下、図面を参照して、本発明の実施形態を説明する。図面の記載において同一部分には同一符号を付し説明を省略する。 Hereinafter, embodiments of the present invention will be described with reference to the drawings. In the description of the drawings, the same parts are designated by the same reference numerals and the description thereof will be omitted.
 [発明の概要]
 本発明は、スライドとプレゼンタエージェントとの組み合せによるプレゼンタシステムをオンライン会議システムに適用し、人のプレゼンを代行するプレゼンタエージェントの動作を制御するためのシナリオ(発話内容、非言語動作)を動的に変更することで、オンライン会議環境でプレゼンの訴求力を向上させる技術である。
[Outline of the invention]
The present invention applies a presenter system by combining slides and a presenter agent to an online conference system, and dynamically creates a scenario (speech content, non-verbal movement) for controlling the operation of the presenter agent acting on behalf of a person's presentation. It is a technology that improves the appeal of presentations in an online conference environment by changing it.
 具体的には、プレゼン内容を聴衆へ効果的に伝達するため、オンライン会議に参加している聴衆の現在の聴衆状態に応じて、プレゼンタエージェントの動作を動的に変更する。つまり、プレゼンタエージェントの動作を制御するためのシナリオを、聴衆状態(聴衆の興味・関心、聴衆の態度)に合わせて動的に変更する。 Specifically, in order to effectively convey the content of the presentation to the audience, the behavior of the presenter agent is dynamically changed according to the current audience status of the audience participating in the online conference. In other words, the scenario for controlling the behavior of the presenter agent is dynamically changed according to the audience state (audience interest / interest, audience attitude).
 特に、本発明は、現在の聴衆状態とシナリオの進行度(プレゼンの進行度)とに応じて、聴衆が高い興味・関心を持ち、プレゼンの要点に集中して理解できる理想的な聴衆状態に遷移するように、シナリオを変更する。また、本発明は、聴衆顔情報(顔の向き、顔の表情など)とチャット投稿内容とを用いて現在の聴衆状態を判定し、その判定結果を基に、シナリオの進行度や発話タイミングによって理想的な聴衆状態を決定する。 In particular, the present invention creates an ideal audience state in which the audience has a high level of interest and can concentrate on and understand the main points of the presentation according to the current audience state and the progress of the scenario (progress of the presentation). Change the scenario so that it transitions. Further, in the present invention, the current audience state is determined using the audience face information (face orientation, facial expression, etc.) and the chat posted content, and based on the determination result, the progress of the scenario and the utterance timing are determined. Determine the ideal audience condition.
 このような特徴を備えることで、複雑なシナリオを予め設定することなく、オンライン会議の聴衆にとって分かりやすいプレゼンを実現可能となり、オンライン会議環境でプレゼンの訴求力を向上可能となる。 By providing such features, it is possible to realize a presentation that is easy for the audience of the online conference to understand without setting a complicated scenario in advance, and it is possible to improve the appealing power of the presentation in the online conference environment.
 [プレゼンタシステムの構成]
 図1は、本実施形態に係るプレゼンタシステムの全体構成を示す図である。プレゼンタシステムは、オンライン会議の聴衆が使用するユーザ端末6と、シナリオに基づいて、スライドの表示およびプレゼンタエージェント100の発話・動作を制御する提示制御装置5と、聴衆状態を取得し、事前定義したルールに基づいて実行中のシナリオを変更するシナリオ制御装置1と、を備える。ユーザ端末6、提示制御装置5、および、シナリオ制御装置1は、相互通信可能な通信ネットワークに接続されている。
[Presenter system configuration]
FIG. 1 is a diagram showing an overall configuration of a presenter system according to the present embodiment. The presenter system acquires and predefines the user terminal 6 used by the audience of the online conference, the presentation control device 5 that controls the display of slides and the speech / operation of the presenter agent 100 based on the scenario, and the audience state. It includes a scenario control device 1 that changes a running scenario based on a rule. The user terminal 6, the presentation control device 5, and the scenario control device 1 are connected to a communication network capable of intercommunication.
 [ユーザ端末の構成]
 ユーザ端末6は、オンライン会議に参加中の聴衆が使用する情報処理端末である。ユーザ端末6は、Webカメラ、ディスプレイ、データ通信機能、データ入出力機能など、を備える。オンライン会議中、ユーザ端末6のディスプレイには、プレゼン用のスライド、スライド内容をプレゼンするプレゼンタエージェント100など、が表示される。ユーザ端末6の数は、少なくとも1つ以上である。ユーザ端末6は、オンライン会議用に関係者や第三者から提供されたパソコンでもよいし、聴衆が自前で用意した携帯端末でもよい。
[User terminal configuration]
The user terminal 6 is an information processing terminal used by an audience participating in an online conference. The user terminal 6 includes a Web camera, a display, a data communication function, a data input / output function, and the like. During the online conference, a slide for presentation, a presenter agent 100 for presenting the slide contents, and the like are displayed on the display of the user terminal 6. The number of user terminals 6 is at least one. The user terminal 6 may be a personal computer provided by a person concerned or a third party for an online conference, or may be a mobile terminal prepared by the audience.
 [提示制御装置の構成]
 提示制御装置5は、図1に例示したように、例えば、シナリオ実行部50と、シナリオ記憶部51と、シナリオ通知部52と、指示制御部53と、を備える。
[Configuration of presentation control device]
As illustrated in FIG. 1, the presentation control device 5 includes, for example, a scenario execution unit 50, a scenario storage unit 51, a scenario notification unit 52, and an instruction control unit 53.
 シナリオ実行部50は、シナリオ記憶部51からシナリオを読み出し、そのシナリオを実行する機能を備える。また、シナリオ実行部50は、シナリオ制御装置1によりシナリオが変更された場合、変更後のシナリオを実行する機能を備える。 The scenario execution unit 50 has a function of reading a scenario from the scenario storage unit 51 and executing the scenario. Further, the scenario execution unit 50 has a function of executing the changed scenario when the scenario is changed by the scenario control device 1.
 シナリオ記憶部51は、プレゼンタエージェント100の動作を制御するためのシナリオを記憶しておく機能を備える。シナリオには、図2に例示するように、スライド毎に、各スライドを音声で説明するための発話内容と、顔の表情、顔の向き、腕の動きなどの非言語動作と、が設定されている。 The scenario storage unit 51 has a function of storing a scenario for controlling the operation of the presenter agent 100. In the scenario, as illustrated in FIG. 2, utterance contents for explaining each slide by voice and non-verbal movements such as facial expression, face orientation, and arm movement are set for each slide. ing.
 シナリオ通知部52は、シナリオ実行状況(シナリオの進行度、プレゼンの進行度)をシナリオ制御装置1の遷移先決定部40へ通知する機能を備える。 The scenario notification unit 52 has a function of notifying the transition destination determination unit 40 of the scenario control device 1 of the scenario execution status (scenario progress, presentation progress).
 指示制御部53は、シナリオ実行部50によるシナリオの実行に基づき、ユーザ端末6に表示されているプレゼンタエージェント100の動作(発話内容、非言語動作)を制御する機能を備える。 The instruction control unit 53 has a function of controlling the operation (utterance content, non-verbal operation) of the presenter agent 100 displayed on the user terminal 6 based on the execution of the scenario by the scenario execution unit 50.
 [シナリオ制御装置の構成]
 シナリオ制御装置1は、オンライン会議でプレゼンを視聴している聴衆の現在の聴衆状態を、聴衆顔画像と投稿チャット内容からリアルタイムに推定しながら、提示制御装置5で実行中のシナリオを変更する装置である。
[Scenario control device configuration]
The scenario control device 1 is a device that changes the scenario being executed by the presentation control device 5 while estimating the current audience state of the audience watching the presentation in the online conference in real time from the audience face image and the posted chat content. Is.
 シナリオ制御装置1は、図1に例示したように、例えば、顔画像取得部10と、顔画像判定部11と、顔表情分析情報記憶部12と、チャット取得部20と、チャット分析部21と、聴衆状態推定部30と、遷移先決定部40と、状態遷移ルール記憶部41と、修正内容選択部42と、シナリオ修正ルール記憶部43と、シナリオ修正部44と、修正履歴記憶部45と、を備える。 As illustrated in FIG. 1, the scenario control device 1 includes, for example, a face image acquisition unit 10, a face image determination unit 11, a facial expression analysis information storage unit 12, a chat acquisition unit 20, and a chat analysis unit 21. , Audience state estimation unit 30, transition destination determination unit 40, state transition rule storage unit 41, correction content selection unit 42, scenario correction rule storage unit 43, scenario correction unit 44, and correction history storage unit 45. , Equipped with.
 顔画像取得部10は、ユーザ端末6からオンライン会議中に撮影された聴衆顔画像を受信する機能を備える。 The face image acquisition unit 10 has a function of receiving an audience face image taken during an online conference from the user terminal 6.
 顔画像判定部11は、顔表情分析情報記憶部12から顔表情分析情報を読み出し、その顔表情分析情報を用いて聴衆顔画像内の聴衆の顔の表情や顔の向きなどを分析し、プレゼンに対する聴衆全体の興味・関心の傾向を分析する機能を備える。 The face image determination unit 11 reads facial expression analysis information from the facial expression analysis information storage unit 12, analyzes the facial expression and face orientation of the audience in the audience facial image using the facial expression analysis information, and makes a presentation. It has a function to analyze the interests and trends of interests of the entire audience.
 顔表情分析情報記憶部12は、聴衆がポジティブ状態であるかネガティブ状態であるかを判定するための画像特徴量が記載された顔表情分析情報を記憶しておく機能を備える。 The facial expression analysis information storage unit 12 has a function of storing facial expression analysis information in which an image feature amount for determining whether the audience is in a positive state or a negative state is described.
 チャット取得部20は、ユーザ端末6からオンライン会議中に聴衆が投稿したチャットデータを受信する機能を備える。 The chat acquisition unit 20 has a function of receiving chat data posted by an audience during an online conference from a user terminal 6.
 チャット分析部21は、チャットデータの内容を分析し、プレゼンに対する聴衆全体の態度の傾向を分析する機能を備える。 The chat analysis unit 21 has a function of analyzing the contents of chat data and analyzing the tendency of the attitude of the entire audience toward the presentation.
 聴衆状態推定部30は、顔画像判定部11による聴衆全体の興味・関心の傾向と、チャット分析部21による聴衆全体の態度の傾向と、を基に、プレゼンに対する聴衆全体の興味・関心および態度に関する現在の聴衆状態を推定計算する機能を備える。 The audience state estimation unit 30 is based on the tendency of the entire audience's interest / interest by the face image determination unit 11 and the tendency of the entire audience's attitude by the chat analysis unit 21, and the audience's overall interest / interest and attitude toward the presentation. It has a function to estimate and calculate the current audience status of.
 遷移先決定部40は、状態遷移ルール記憶部41から状態遷移ルールを読み出し、その状態遷移ルールを用いて、現在の聴衆状態と、提示制御装置5のシナリオ通知部52から通知されたシナリオ実行状況と、に基づき、聴衆全体の興味・関心および態度を向上させることが可能な理想的な遷移先の聴衆状態を決定する機能を備える。 The transition destination determination unit 40 reads the state transition rule from the state transition rule storage unit 41, and uses the state transition rule to notify the current audience state and the scenario execution status notified from the scenario notification unit 52 of the presentation control device 5. Based on the above, it has a function to determine the ideal transition destination audience state that can improve the interest / interest and attitude of the entire audience.
 状態遷移ルール記憶部41は、遷移先の聴衆状態を決定するための状態遷移ルールを記憶しておく機能を備える。状態遷移ルールには、図3に例示するように、シナリオ実行状況と、遷移先の聴衆状態を示す遷移先状態と、が設定されている。 The state transition rule storage unit 41 has a function of storing a state transition rule for determining the audience state of the transition destination. As illustrated in FIG. 3, the state transition rule is set with a scenario execution status and a transition destination state indicating the audience state of the transition destination.
 修正内容選択部42は、シナリオ修正ルール記憶部43からシナリオ修正ルールを読み出し、そのシナリオ修正ルールから、現在の聴衆状態を遷移先の聴衆状態に遷移させるために用いる修正用シナリオを選択する機能を備える。また、修正内容選択部42は、修正履歴記憶部45からシナリオ修正履歴情報を読み出し、そのシナリオ修正履歴情報を参照して、同一の状態遷移が複数回ある場合、当該同一の状態遷移で過去に選択した修正用シナリオとは別の修正用シナリオを選択する機能を備える。 The modification content selection unit 42 reads a scenario modification rule from the scenario modification rule storage unit 43, and selects a modification scenario used to transition the current audience state to the transition destination audience state from the scenario modification rule. Be prepared. Further, the correction content selection unit 42 reads the scenario correction history information from the correction history storage unit 45, refers to the scenario correction history information, and if the same state transition occurs a plurality of times, the same state transition is used in the past. It has a function to select a correction scenario different from the selected correction scenario.
 シナリオ修正ルール記憶部43は、現在の聴衆状態を遷移先の聴衆状態に遷移させるための修正用シナリオが設定されたシナリオ修正ルールを記憶しておく機能を備える。シナリオ修正ルールには、図4に例示するように、修正用シナリオのIDと、現在の聴衆状態を示す現在状態と、遷移先の聴衆状態を示す遷移先状態と、シナリオ修正内容と、修正ターゲットと、が設定されている。修正ターゲットには、シナリオを構成する発話内容と非言語動作とのうち一方または両方が設定されている。 The scenario correction rule storage unit 43 has a function of storing a scenario correction rule in which a correction scenario for transitioning the current audience state to the transition destination audience state is set. As illustrated in FIG. 4, the scenario modification rule includes the ID of the scenario for modification, the current state indicating the current audience state, the transition destination state indicating the audience state of the transition destination, the scenario modification content, and the modification target. And are set. The modification target is set to one or both of the utterance content and the non-verbal behavior that make up the scenario.
 シナリオ修正部44は、提示制御装置5のシナリオ記憶部51に記憶されているシナリオ、シナリオ実行部50がシナリオ記憶部51から読み出したシナリオを、修正用シナリオに変更する機能を備える。 The scenario correction unit 44 has a function of changing the scenario stored in the scenario storage unit 51 of the presentation control device 5 and the scenario read from the scenario storage unit 51 by the scenario execution unit 50 into a correction scenario.
 修正履歴記憶部45は、シナリオ修正ルールから選択した修正用シナリオの利用回数をシナリオ修正履歴情報として記憶する機能を備える。シナリオ修正履歴情報には、図5に例示するように、修正用シナリオのIDと、修正用シナリオの利用回数と、が記憶される。 The correction history storage unit 45 has a function of storing the number of times the correction scenario selected from the scenario correction rule is used as scenario correction history information. As illustrated in FIG. 5, the scenario correction history information stores the ID of the correction scenario and the number of times the correction scenario is used.
 [シナリオ制御装置の構成上の特徴および特徴的な処理動作]
 本発明は、オンライン会議システムに参加してプレゼンを視聴している聴衆の聴衆状態をリアルタイムに推定し、推定した現在の聴衆状態を踏まえて、プレゼンタエージェントのシナリオを動的に変更することで、訴求力の高いプレゼンを実現する。
[Structural features and characteristic processing operations of the scenario control device]
The present invention estimates the audience state of an audience who participates in an online conference system and is watching a presentation in real time, and dynamically changes the scenario of the presenter agent based on the estimated current audience state. Achieve a highly appealing presentation.
 そのため、図1に示したシナリオ制御装置1の構成上の特徴は、聴衆状態の特徴を算出する顔画像判定部11およびチャット分析部21と、聴衆全体の興味・関心および態度に関する現在の聴衆状態を推定する聴衆状態推定部(推定部)30と、現在の聴衆状態とシナリオ実行状況とを基に、現在の聴衆状態をどのような状態へ遷移させるかを決定する遷移先決定部(第1の決定部)40と、事前に定義されたシナリオ修正ルールと過去に行われたシナリオの修正履歴とを基に、修正用シナリオを決定する修正内容選択部(第2の決定部)42と、実行中のシナリオを修正用シナリオに変更するシナリオ修正部(変更部)44と、を備えた点にある。 Therefore, the structural features of the scenario control device 1 shown in FIG. 1 include the face image determination unit 11 and the chat analysis unit 21 that calculate the characteristics of the audience state, and the current audience state regarding the interest, interest, and attitude of the entire audience. Audience state estimation unit (estimation unit) 30 that estimates the current audience state, and a transition destination determination unit (first) that determines what state the current audience state should be transitioned to based on the current audience state and the scenario execution status. (Decision unit) 40, a modification content selection unit (second determination unit) 42 that determines a scenario for modification based on a predefined scenario modification rule and a history of modification of a scenario performed in the past, and It is provided with a scenario correction unit (change unit) 44 for changing a running scenario to a correction scenario.
 続いて、上記構成上の特徴に基づくシナリオ制御装置1の特徴的な処理動作を説明する。図6は、シナリオ制御装置1の特徴的な処理動作を示すフロー図である。 Subsequently, the characteristic processing operation of the scenario control device 1 based on the above-mentioned structural characteristics will be described. FIG. 6 is a flow chart showing a characteristic processing operation of the scenario control device 1.
 ステップS1;
 まず、顔画像判定部11が、複数のユーザ端末6からそれぞれ取得した各聴衆顔画像の顔の表情や顔の向きから、各聴衆のポジティブ/ネガティブ状態をそれぞれ分析し、聴衆全体のポジティブ/ネガティブ状態を判定する(図7参照)。
Step S1;
First, the face image determination unit 11 analyzes the positive / negative state of each audience from the facial expressions and face orientations of each audience face image acquired from the plurality of user terminals 6, respectively, and positive / negative of the entire audience. The state is determined (see FIG. 7).
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
 例えば、顔画像判定部11は、式(1)より、i番目の聴衆fst(1≦i≦N)の顔表情がポジティブ状態であればS=「1」とし、ネガティブ状態であればS=「0」とし、ポジティブ/ネガティブ状態の算定が困難であればS=「0」とする。そして、顔画像判定部11は、式(2)より、すべての聴衆のポジティブ/ネガティブ状態を合計したSallを求める。その後、顔画像判定部11は、式(3)より、Sall>(N/2)である場合、聴衆全体をポジティブ状態と決定し、それ以外の場合(ポジティブ状態とネガティブ状態とが同数の場合を含む)、聴衆全体をネガティブ状態と決定する。ポジティブ状態の場合、プレゼンに対する聴衆全体の興味・関心は高い。ネガティブ状態の場合、プレゼンに対する聴衆全体の興味・関心は低い。 For example, from the equation (1), the face image determination unit 11 sets Si = “1” if the facial expression of the i-th audience fst i (1 ≦ i ≦ N) is in a positive state, and if it is in a negative state, it sets Si = “1”. S i = "0", and if it is difficult to calculate the positive / negative state, S i = "0". Then, the face image determination unit 11 obtains All from the equation (2), which is the sum of the positive / negative states of all the audiences. After that, the face image determination unit 11 determines from the equation (3) that if All > (N / 2), the entire audience is in a positive state, and in other cases (the number of positive states and negative states is the same). (Including cases), determine the entire audience as negative. In the positive state, the audience as a whole is highly interested in the presentation. In the negative state, the audience as a whole has low interest in the presentation.
 次に、チャット分析部21が、オンライン会議中に各聴衆により投稿された各チャットの投稿内容や投稿文量から、各聴衆の集中/発散状態をそれぞれ分析し、聴衆全体の集中/発散状態を判定する(図7参照)。 Next, the chat analysis unit 21 analyzes the concentration / divergence state of each audience from the posted content and the amount of sentences of each chat posted by each audience during the online conference, and determines the concentration / divergence state of the entire audience. Judgment (see FIG. 7).
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000005
Figure JPOXMLDOC01-appb-M000005
 例えば、チャット分析部21は、一定間隔で投稿されたチャットを取得する。次に、チャット分析部21は、取得したチャットj毎の投稿文量Cl(1≦j≦M)と、チャットjとスライド内容との類似度Csと、を算出する。投稿文量Clは、チャットjに含まれる内容語数を一定間隔で取得した全チャットの単語数で正規化した値(0≦Cl≦1)である。類似度Csは、チャットjの内容語と質問対象スライドの内容語との類似度を示す値(0≦Cs≦1)である。次に、チャット分析部21は、算出した投稿文量Clと類似度Csとの乗算値からチャットj毎のスコアSを算出し、式(4)より、一定間隔の全スコアを合算したSCallを求める。最後に、チャット分析部21は、式(5)より、SCall>閾値Thである場合、聴衆全体を集中状態と決定し、それ以外の場合、聴衆全体を発散状態と決定する。投稿文量Clが多く、スライドとの類似度Csが高い程、集中して聞いている状態を表すことになる。 For example, the chat analysis unit 21 acquires chats posted at regular intervals. Next, the chat analysis unit 21 calculates the amount of posted text Cl j (1 ≦ j ≦ M) for each acquired chat j and the degree of similarity Cs j between the chat j and the slide content. The posted sentence amount Cl j is a value (0 ≦ Cl ≦ 1) obtained by normalizing the number of content words included in the chat j with the number of words in all chats acquired at regular intervals. The similarity Cs j is a value (0 ≦ Cs ≦ 1) indicating the degree of similarity between the content word of the chat j and the content word of the question target slide. Next, the chat analysis unit 21 calculates the score S j for each chat j from the product of the calculated posted sentence amount Cl j and the similarity Cs j , and totals all the scores at regular intervals from the equation (4). Obtain the SC all . Finally, from the equation (5), the chat analysis unit 21 determines that the entire audience is in the concentrated state when SC all > threshold Th, and in other cases, the entire audience is determined to be in the divergent state. The larger the amount of posted text Cl j and the higher the degree of similarity Cs j with the slide, the more concentrated listening is represented.
 ステップS2;
 次に、聴衆状態推定部30が、ポジティブ/ネガティブ状態の判定結果と、集中/発散状態の判定結果と、を基に、プレゼンに対する聴衆全体の興味・関心および態度に関する現在の聴衆状態を推定計算する。
Step S2;
Next, the audience state estimation unit 30 estimates and calculates the current audience state regarding the entire audience's interest / interest and attitude toward the presentation based on the positive / negative state determination result and the concentration / divergence state determination result. do.
 例えば、聴衆状態推定部30は、プレゼンに対する聴衆全体の興味・関心を高(ポジティブ状態)・低(ネガティブ状態)で区分けし、聴衆全体の態度を集中的(集中状態)・発散的(発散状態)でさらに区分けした4つの聴衆状態の中から、現在の聴衆状態を推定する(図8参照)。ポジティブ/集中状態であれば、プレゼンに興味があり、プレゼンの要点や詳細を聞こうとする状態1とする。ポジティブ/発散状態であれば、プレゼンに興味はあるが、プレゼンの要点や詳細が分からない状態2とする。ネガティブ/集中状態であれば、プレゼンの興味は低いが、プレゼンを義務的に聞こうとする状態3とする。ネガティブ/発散状態であれば、プレゼンに興味はなく、プレゼン以外に気が移っている状態4とする。 For example, the audience state estimation unit 30 classifies the interest and interest of the entire audience into high (positive state) and low (negative state), and the attitude of the entire audience is concentrated (concentrated state) and divergent (divergent state). ), The current audience state is estimated from the four audience states (see FIG. 8). If it is a positive / concentrated state, it is a state 1 in which the person is interested in the presentation and wants to hear the main points and details of the presentation. If it is a positive / divergent state, it is assumed that the person is interested in the presentation but does not know the main points and details of the presentation. In the negative / concentrated state, the interest of the presentation is low, but the state 3 in which the presentation is obligatory to be listened to. If it is a negative / divergent state, it is assumed that the person is not interested in the presentation and is in a state of being distracted other than the presentation 4.
 なお、本実施形態では、興味・関心および態度のいずれも二値としたが、二値以外の多値でもよい。例えば、二値以外の多値の状態判定として聴衆状態モデルを拡張することで、より細かい聴衆状態の変化にも対応可能となる。 In this embodiment, all of interest / interest and attitude are binary, but a multi-value other than the binary may be used. For example, by extending the audience state model as a multi-valued state determination other than binary, it is possible to deal with finer changes in the audience state.
 ステップS3;
 次に、遷移先決定部40が、現在の聴衆状態と、シナリオ実行状況と、に基づき、プレゼンに対する聴衆全体の興味・関心を高め、かつ、聴衆全体の態度を集中させる理想的な遷移先の聴衆状態を決定する。
Step S3;
Next, the transition destination determination unit 40 is an ideal transition destination that raises the interest and interest of the entire audience in the presentation and concentrates the attitude of the entire audience based on the current audience status and the scenario execution status. Determine the audience status.
 例えば、遷移先決定部40は、現在の聴衆状態が状態4である場合、状態1~状態3のうちいずれかの状態を遷移先の聴衆状態として決定する。特に、状態1を遷移先の聴衆状態として決定する場合、シナリオ実行状況がシナリオの前半(プレゼン開始直後)であれば、先ずはプレゼンに対する聴衆全体の興味・関心を高めるために状態2への状態遷移を優先し、その後、状態1への状態遷移とする。つまり、興味・関心を高めてから集中的態度へ遷移させる(図9(a),(b)参照)。一方、シナリオ実行状況がシナリオの後半(一定時間経過後)であれば、先ずはプレゼンに対する聴衆全体の態度を集中させるために状態3を優先し、その後、状態1への状態遷移とする。つまり、集中的態度としてから興味・関心を高める(図9(a),(c)参照)。 For example, when the current audience state is the state 4, the transition destination determination unit 40 determines any of the states 1 to 3 as the transition destination audience state. In particular, when determining state 1 as the transition destination audience state, if the scenario execution status is the first half of the scenario (immediately after the start of the presentation), first the state to state 2 is to increase the interest and interest of the entire audience in the presentation. Priority is given to the transition, and then the state transition to state 1 is performed. In other words, after raising interest / interest, the attitude shifts to a focused attitude (see FIGS. 9 (a) and 9 (b)). On the other hand, if the scenario execution status is the latter half of the scenario (after a certain period of time has elapsed), the state 3 is given priority first in order to concentrate the attitude of the entire audience toward the presentation, and then the state transition to the state 1 is performed. In other words, after having a focused attitude, the interest / interest is increased (see FIGS. 9 (a) and 9 (c)).
 ステップS4;
 次に、修正内容選択部42が、現在の聴衆状態を遷移先の聴衆状態に遷移させるための修正用シナリオを決定する。例えば、修正内容選択部42は、シナリオ修正ルールから現在の聴衆状態および遷移先の聴衆状態に対応する修正用シナリオを取得する。このとき、修正内容選択部42は、その修正用シナリオの利用回数を更新する。
Step S4;
Next, the modification content selection unit 42 determines a modification scenario for transitioning the current audience state to the transition destination audience state. For example, the modification content selection unit 42 acquires the modification scenario corresponding to the current audience state and the transition destination audience state from the scenario modification rule. At this time, the modification content selection unit 42 updates the number of times the modification scenario is used.
 修正用シナリオとは、図10に示す状態遷移図の状態間に、「同じ説明を繰り返す」や「アイコンタクト動作」などの遷移条件を記述したものである。「状態遷移元-状態遷移先」の各状態遷移に応じた遷移条件が設定される。非言語動作とは、顔の表情、顔の向き、腕の位置・方向など、身体的な動きである。発話内容とは、プレゼンタエージェントが発声する発話内容、発話スクリプト、音声の抑揚・間、音声と同時に効果音を鳴らすという、音情報・文字情報である。 The correction scenario describes transition conditions such as "repeat the same explanation" and "eye contact operation" between the states of the state transition diagram shown in FIG. Transition conditions are set according to each state transition of "state transition source-state transition destination". Nonverbal movements are physical movements such as facial expressions, face orientation, and arm positions / directions. The utterance content is sound information / text information such as the utterance content uttered by the presenter agent, the utterance script, the intonation / interval of the voice, and the sound effect at the same time as the voice.
 なお、現在の聴衆状態を遷移先の聴衆状態に遷移させる状態遷移と同じ状態遷移が過去にも行われていた場合、当該過去で使用した修正用シナリオと同一の修正用シナリオを使用しても、プレゼンの訴求力は向上しない可能性がある。この場合、修正内容選択部42は、修正用シナリオの利用回数を参照し、画一的なシナリオ修正にならないように、過去の修正用シナリオとは別の修正用シナリオを選択する。例えば、図11に例示するように、状態4から状態3への状態遷移を2回目に行う場合、1回目の「スライドへの注意誘導」とは異なる「ポインティング動作」を選択する。 If the same state transition as the state transition that transitions the current audience state to the transition destination audience state has been performed in the past, even if the same correction scenario as the correction scenario used in the past is used. , The appeal of the presentation may not improve. In this case, the correction content selection unit 42 refers to the number of times the correction scenario is used, and selects a correction scenario different from the past correction scenario so as not to make a uniform scenario correction. For example, as illustrated in FIG. 11, when the state transition from the state 4 to the state 3 is performed for the second time, a “pointing operation” different from the first “attention guidance to the slide” is selected.
 ステップS5;
 最後に、シナリオ修正部44が、提示制御装置5で実行中のシナリオを修正用シナリオに変更する。
Step S5;
Finally, the scenario correction unit 44 changes the scenario being executed by the presentation control device 5 to the correction scenario.
 [シナリオ制御装置の処理動作の具体例]
 図12は、シナリオ制御装置1の処理動作の具体例を示すフロー図である。
[Specific example of processing operation of scenario control device]
FIG. 12 is a flow chart showing a specific example of the processing operation of the scenario control device 1.
 ステップS101;
 まず、顔画像取得部10が、複数のユーザ端末6から各聴衆顔画像をそれぞれ受信する。そして、顔画像判定部11が、顔表情分析情報記憶部12の顔表情分析情報に基づき、各聴衆のポジティブ/ネガティブ状態をそれぞれ判定し、プレゼンに対する聴衆全体のポジティブ/ネガティブ状態を算出する。
Step S101;
First, the face image acquisition unit 10 receives each audience face image from the plurality of user terminals 6. Then, the face image determination unit 11 determines the positive / negative state of each audience based on the facial expression analysis information of the facial expression analysis information storage unit 12, and calculates the positive / negative state of the entire audience for the presentation.
 ステップS102;
 次に、チャット取得部20が、ユーザ端末6からオンラインツール上で交されるチャットのチャットデータを受信する。そして、チャット分析部21が、チャットデータの投稿文量と、投稿内容とスライド内容との類似度と、を計算し、プレゼンに対する聴衆全体の集中/発散状態を算出する。
Step S102;
Next, the chat acquisition unit 20 receives the chat data of the chat exchanged on the online tool from the user terminal 6. Then, the chat analysis unit 21 calculates the amount of posted text of the chat data and the degree of similarity between the posted content and the slide content, and calculates the concentration / divergence state of the entire audience with respect to the presentation.
 ステップS103;
 次に、聴衆状態推定部30が、聴衆全体のポジティブ/ネガティブ状態と、聴衆全体の集中/発散状態の傾向と、を基に、プレゼンに対する聴衆全体の興味・関心および態度に関する現在の聴衆状態を推定計算する。例えば、図8に例示した4つの状態1~状態4の中から1つの状態が決定される。
Step S103;
Next, the audience state estimation unit 30 determines the current audience state regarding the audience's interest / interest and attitude toward the presentation based on the positive / negative state of the entire audience and the tendency of the concentration / divergence state of the entire audience. Estimate calculation. For example, one state is determined from the four states 1 to 4 illustrated in FIG.
 ステップS104;
 次に、遷移先決定部40が、現在の聴衆状態の推定結果を基に、状態遷移が必要か否かを判定する。現在の聴衆状態が状態1である場合、聴衆全体は興味・関心を持って集中してプレゼンを視聴しており、状態遷移は不要と判定して処理を終了する。現在の聴衆状態が状態2~状態4である場合、状態遷移は必要と判定してステップS105へ進む。
Step S104;
Next, the transition destination determination unit 40 determines whether or not the state transition is necessary based on the estimation result of the current audience state. When the current audience state is state 1, the entire audience is watching the presentation with interest and interest, and it is determined that the state transition is unnecessary and the process is terminated. When the current audience state is the state 2 to the state 4, it is determined that the state transition is necessary, and the process proceeds to step S105.
 ステップS105;
 次に、遷移先決定部40は、現在の聴衆状態から遷移させる遷移先状態の候補が1つか否かを判定する。現在の聴衆状態が状態4であれば、遷移先状態は状態1~状態3の複数であり、ステップS106へ進む。現在の聴衆状態が状態2または状態3であれば、遷移先状態は状態1のみであり、ステップS107へ進む。
Step S105;
Next, the transition destination determination unit 40 determines whether or not there is one candidate for the transition destination state to be transitioned from the current audience state. If the current audience state is state 4, the transition destination states are a plurality of states 1 to 3, and the process proceeds to step S106. If the current audience state is state 2 or state 3, the transition destination state is only state 1, and the process proceeds to step S107.
 ステップS106;
 遷移先状態の候補が複数である場合、遷移先決定部40は、提示制御装置5からシナリオ実行状況を取得し、プレゼンの前半(開始から方針転換位置まで)であるか、プレゼンの後半(方針転換位置から終了)であるかに基づき、状態遷移ルール記憶部41の状態遷移ルール(図3)から対応する遷移先状態を読み出し、それを遷移先状態の候補とする。
Step S106;
When there are a plurality of candidates for the transition destination state, the transition destination determination unit 40 acquires the scenario execution status from the presentation control device 5 and either is the first half of the presentation (from the start to the policy change position) or the second half of the presentation (policy). Based on whether it ends from the conversion position), the corresponding transition destination state is read from the state transition rule (FIG. 3) of the state transition rule storage unit 41, and this is used as a candidate for the transition destination state.
 ステップS107;
 次に、遷移先決定部40は、ステップS105で把握した1つ遷移先状態の候補、または、ステップS106で把握した遷移先状態の候補を、遷移先の聴衆状態として決定する。
Step S107;
Next, the transition destination determination unit 40 determines the candidate for one transition destination state grasped in step S105 or the candidate for the transition destination state grasped in step S106 as the transition destination audience state.
 ステップS108;
 次に、修正内容選択部42が、現在の聴衆状態と、遷移先の聴衆状態と、に基づき、シナリオ修正ルール記憶部43のシナリオ修正ルール(図4)から候補となるすべての修正用シナリオを読み出し、修正履歴記憶部45のシナリオ修正履歴情報(図5)に記録された過去の修正結果と可能な限り重複しないように、最終的な修正用シナリオを選択する。修正用シナリオの選択処理の具体例は、後述する。
Step S108;
Next, the correction content selection unit 42 selects all the correction scenarios that are candidates from the scenario correction rule (FIG. 4) of the scenario correction rule storage unit 43 based on the current audience state and the transition destination audience state. The final correction scenario is selected so as not to overlap with the past correction results recorded in the scenario correction history information (FIG. 5) of the read and correction history storage unit 45 as much as possible. A specific example of the correction scenario selection process will be described later.
 ステップS109;
 最後に、シナリオ修正部44が、提示制御装置5で実行中のシナリオを修正用シナリオで変更し、修正履歴記憶部45のシナリオ修正履歴情報を更新する。シナリオ変更処理の具体例も後述する。
Step S109;
Finally, the scenario correction unit 44 changes the scenario being executed by the presentation control device 5 with the correction scenario, and updates the scenario correction history information of the correction history storage unit 45. A specific example of the scenario change process will also be described later.
 [修正用シナリオ選択処理]
 図13は、修正用シナリオの選択処理の具体例を示すフロー図である。
[Correction scenario selection process]
FIG. 13 is a flow chart showing a specific example of the selection process of the correction scenario.
 ステップS108-1;
 まず、修正内容選択部42は、シナリオ修正ルール記憶部43のシナリオ修正ルール(図4)から、現在の聴衆状態と遷移先の聴衆状態との組み合わせに対応するすべての修正用シナリオを修正用シナリオの候補として取得する。
Step S108-1;
First, the correction content selection unit 42 corrects all the correction scenarios corresponding to the combination of the current audience state and the transition destination audience state from the scenario correction rule (FIG. 4) of the scenario correction rule storage unit 43. Get as a candidate for.
 ステップS108-2;
 次に、修正内容選択部42は、修正用シナリオの候補数が1つであるか否かを判定する。修正用シナリオの候補数が1つである場合、ステップS108-6へ進む。修正用シナリオの候補数が複数である場合、ステップS108-3へ進む。
Step S108-2;
Next, the modification content selection unit 42 determines whether or not the number of candidates for the modification scenario is one. If the number of candidates for the modification scenario is one, the process proceeds to step S108-6. If there are a plurality of candidates for the correction scenario, the process proceeds to step S108-3.
 ステップS108-3;
 修正用シナリオの候補数が複数である場合、修正内容選択部42は、修正履歴記憶部45のシナリオ修正履歴情報(図5)から、各候補の修正用シナリオがこれまでに利用された過去の利用回数をそれぞれ取得する。
Step S108-3;
When there are a plurality of candidates for the correction scenario, the correction content selection unit 42 uses the scenario correction history information (FIG. 5) of the correction history storage unit 45 to indicate that each candidate's correction scenario has been used so far. Get the number of uses respectively.
 ステップS108-4;
 ステップS108-3の後、修正内容選択部42は、複数の修正用シナリオの候補のうち、利用回数が最小でない候補を破棄し、利用回数が最小の候補を選択する。
Step S108-4;
After step S108-3, the modification content selection unit 42 discards the candidate whose number of uses is not the minimum among the candidates for the plurality of modification scenarios, and selects the candidate whose number of uses is the smallest.
 ステップS108-5;
 ステップS108-4の後、修正内容選択部42は、残存する修正用シナリオの候補数が1つであるか否かを判定する。残存する修正用シナリオの候補数が1つである場合、ステップS108-6へ進む。残存する修正用シナリオの候補数が複数である場合、ステップS108-7へ進む。
Step S108-5;
After step S108-4, the correction content selection unit 42 determines whether or not the number of remaining correction scenario candidates is one. If the number of remaining correction scenario candidates is one, the process proceeds to step S108-6. If the number of remaining correction scenario candidates is a plurality, the process proceeds to step S108-7.
 ステップS108-6;
 ステップS108-2またはステップS108-5で修正用シナリオの候補数が1つと判定された場合、修正内容選択部42は、その1つの修正用シナリオの候補を変更対象の修正用シナリオとして選択する。
Step S108-6;
When it is determined in step S108-2 or step S108-5 that the number of candidates for the modification scenario is one, the modification content selection unit 42 selects the candidate for the modification scenario as the modification scenario to be changed.
 ステップS108-7;
 ステップS108-5で修正用シナリオの候補数が複数と判定された場合、修正内容選択部42は、残存する複数の修正用シナリオの候補の中からランダムに1つの候補を変更対象の修正用シナリオとして選択する。
Step S108-7;
When it is determined in step S108-5 that the number of candidates for the correction scenario is a plurality, the correction content selection unit 42 randomly changes one candidate from the remaining candidates for the correction scenario. Select as.
 [シナリオ変更処理]
 図14は、シナリオ変更処理の具体例を示すフロー図である。
[Scenario change processing]
FIG. 14 is a flow chart showing a specific example of the scenario change process.
 ステップS109-1;
 まず、シナリオ修正部44は、シナリオ修正ルール記憶部43のシナリオ修正ルール(図4)から、ステップS108で選択された修正用シナリオの修正ターゲットを取得する。
Step S109-1;
First, the scenario correction unit 44 acquires the correction target of the correction scenario selected in step S108 from the scenario correction rule (FIG. 4) of the scenario correction rule storage unit 43.
 ステップS109-2;
 次に、シナリオ修正部44は、取得した修正ターゲットの種別を判定する。修正ターゲットの種別が非言語動作である場合、ステップS109-3へ進む。修正ターゲットの種別が発話内容である場合、ステップS109-4へ進む。修正ターゲットの種別が全て(非言語動作と発話内容との両方)である場合、ステップS109-5へ進む。
Step S109-2;
Next, the scenario correction unit 44 determines the type of the acquired correction target. If the type of the modification target is non-verbal operation, the process proceeds to step S109-3. If the type of the correction target is the utterance content, the process proceeds to step S109-4. If the types of the modification targets are all (both non-verbal actions and utterance contents), the process proceeds to step S109-5.
 ステップS109-3;
 修正ターゲットの種別が非言語動作である場合、シナリオ修正部44は、修正前のシナリオに設定されている非言語動作に対してシナリオ修正内容の非言語動作で置換し、または、その非言語動作にシナリオ修正内容を追加する。
Step S109-3;
When the type of the modification target is non-verbal operation, the scenario modification unit 44 replaces the non-language operation set in the scenario before modification with the non-language operation of the scenario modification content, or the non-language operation thereof. Add the scenario correction contents to.
 ステップS109-4;
 修正ターゲットの種別が発話内容である場合、シナリオ修正部44は、修正前のシナリオに設定されている発話内容に対して、シナリオ修正内容の発話内容を追加する。
Step S109-4;
When the type of the correction target is the utterance content, the scenario correction unit 44 adds the utterance content of the scenario correction content to the utterance content set in the scenario before the correction.
 ステップS109-5;
 修正ターゲットの種別が全てである場合、シナリオ修正部44は、同じスライドの説明を繰り返すように、実行中のシナリオを変更する。シナリオの変更例を図15に示す。図15(a)は、興味・関心を高めるための動作例である。図15(a-1)のように、発話内容に対して効果音を追加する。図15(a-2)のように、同じ発話内容を繰り返すことで、発話内容を強調する。これにより、スライド内容に対する聴衆の興味・関心を高めることができる。
Step S109-5;
When the types of modification targets are all, the scenario modification unit 44 changes the running scenario so as to repeat the explanation of the same slide. An example of changing the scenario is shown in FIG. FIG. 15A is an operation example for raising interest / interest. As shown in FIG. 15 (a-1), a sound effect is added to the utterance content. As shown in FIG. 15 (a-2), the utterance content is emphasized by repeating the same utterance content. This makes it possible to increase the interest and interest of the audience in the slide contents.
 図15(b)は、態度を促すための動作例である。図15(b-1)のように、スライドの重要点を手や指でポインティングする。図15(b-2)のように、聴衆に対してアイコンタクトすることで、聴衆に語りかける印象を与え、より集中して話を聞くように促す。図15(b-3)のように、スライドの重要点を注視してスライドへの注意を誘導することで、スライド内容を注意して説明を聞くように促す。これにより、スライド内容に対する聴衆の態度を高めることができる。 FIG. 15 (b) is an operation example for promoting an attitude. As shown in FIG. 15 (b-1), the important points of the slide are pointed with a hand or a finger. As shown in FIG. 15 (b-2), making eye contact with the audience gives the audience the impression of speaking and encourages them to listen more intensively. As shown in FIG. 15 (b-3), by paying attention to the important points of the slide and guiding the attention to the slide, the slide contents are carefully urged to listen to the explanation. This can enhance the audience's attitude towards the slide content.
 ステップS109-6;
 最後に、シナリオ修正部44は、変更対象であった修正用シナリオに関する修正履歴記憶部45のシナリオ修正履歴情報を更新する。例えば、シナリオ修正部44は、当該修正用シナリオの利用回数をインクリメントする。
Step S109-6;
Finally, the scenario correction unit 44 updates the scenario correction history information of the correction history storage unit 45 regarding the correction scenario that was the change target. For example, the scenario correction unit 44 increments the number of times the correction scenario is used.
 以上、シナリオ制御装置1の処理動作を説明した。シナリオ制御装置1は、上記処理動作を定期的に実行し、現在の聴衆状態を推定しながら、プレゼンに対する聴衆全体の興味・関心を高め、かつ、聴衆全体の態度を集中させるように、プレゼンタエージェントのシナリオを変更する。提示制御装置5は、シナリオ記憶部51のシナリオに基づき、シナリオ実行部50が指示制御部53に当該シナリオに基づいた制御を指示しながら、動的に当該シナリオを上述した修正シナリオで変更する。また、提示制御装置5は、シナリオ通知部52を介して、シナリオ修正に必要なシナリオ実行状況をシナリオ制御装置1へ逐一通知する。 The processing operation of the scenario control device 1 has been explained above. The scenario control device 1 periodically executes the above processing operation, estimates the current audience state, raises the interest and interest of the entire audience in the presentation, and concentrates the attitude of the entire audience. Change the scenario. Based on the scenario of the scenario storage unit 51, the presentation control device 5 dynamically changes the scenario in the above-mentioned modified scenario while the scenario execution unit 50 instructs the instruction control unit 53 to control the scenario based on the scenario. Further, the presentation control device 5 notifies the scenario control device 1 of the scenario execution status required for scenario correction one by one via the scenario notification unit 52.
 [本実施形態の効果]
 本実施形態では、シナリオ制御装置1が、オンライン会議システム上のリモートプレゼン中における聴衆状態を推定し、その推定結果に基づきシナリオを変更する。これにより、リモートプレゼンにおけるプレゼンタシステムにおいて、より訴求力の高いプレゼンテーションを提供可能となる。
[Effect of this embodiment]
In the present embodiment, the scenario control device 1 estimates the audience state during the remote presentation on the online conference system, and changes the scenario based on the estimation result. This makes it possible to provide a more appealing presentation in the presenter system in the remote presentation.
 [本実施形態の変形例]
 本実施形態では、コンピュータ内のバーチャル空間でプレゼンタエージェントを動作させるプレゼンタシステムを前提に説明した(図16(a)参照)。その他、スライドコンテンツ、オーラルコンテンツ(発話内容)、ジェスチャ動作(非言語動作)の3要素が表現・実行可能であれば、プレゼンタエージェントの実体は実空間とバーチャル空間とのどちらでもよい。例えば、図16(b)に示すように、実体のあるプレゼンタエージェント100を用いてもよい。この場合、プレゼンタエージェント100は、聴衆毎に存在させる。
[Modified example of this embodiment]
In the present embodiment, the presenter system in which the presenter agent is operated in the virtual space in the computer has been described as a premise (see FIG. 16A). In addition, as long as the three elements of slide content, oral content (utterance content), and gesture motion (non-verbal motion) can be expressed and executed, the entity of the presenter agent may be either real space or virtual space. For example, as shown in FIG. 16 (b), a tangible presenter agent 100 may be used. In this case, the presenter agent 100 is present for each audience.
 [その他]
 本発明は、上記実施形態に限定されない。本発明は、本発明の要旨の範囲内で数々の変形が可能である。
[others]
The present invention is not limited to the above embodiment. The present invention can be modified in a number of ways within the scope of the gist of the present invention.
 本実施形態のシナリオ制御装置1は、例えば、図17に示すように、CPU(Central Processing Unit、プロセッサ)901と、メモリ902と、ストレージ(Hard Disk Drive、Solid State Drive)903と、通信装置904と、入力装置905と、出力装置906と、を備えた汎用的なコンピュータシステムを用いて実現できる。メモリ902及びストレージ903は、記憶装置である。当該コンピュータシステムにおいて、CPU901がメモリ902上にロードされた所定のプログラムを実行することにより、シナリオ制御装置1の各機能が実現される。 The scenario control device 1 of the present embodiment is, for example, as shown in FIG. 17, a CPU (Central Processing Unit, processor) 901, a memory 902, a storage (Hard Disk Drive, Solid State Drive) 903, and a communication device 904. And, it can be realized by using a general-purpose computer system including an input device 905 and an output device 906. The memory 902 and the storage 903 are storage devices. In the computer system, each function of the scenario control device 1 is realized by the CPU 901 executing a predetermined program loaded on the memory 902.
 シナリオ制御装置1は、1つのコンピュータで実装されてもよい。シナリオ制御装置1は、複数のコンピュータで実装されてもよい。シナリオ制御装置1は、コンピュータに実装される仮想マシンであってもよい。シナリオ制御装置1用のプログラムは、HDD、SSD、USB(Universal Serial Bus)メモリ、CD(Compact Disc)、DVD(Digital Versatile Disc)などのコンピュータ読取り可能な記録媒体に記憶できる。シナリオ制御装置1用のシナリオ制御プログラムは、通信ネットワークを介して配信することもできる。 The scenario control device 1 may be mounted on one computer. The scenario control device 1 may be implemented by a plurality of computers. The scenario control device 1 may be a virtual machine mounted on a computer. The program for the scenario control device 1 can be stored in a computer-readable recording medium such as an HDD, SSD, USB (Universal Serial Bus) memory, CD (Compact Disc), or DVD (Digital Versatile Disc). The scenario control program for the scenario control device 1 can also be distributed via the communication network.
 1:シナリオ制御装置
 10:顔画像取得部
 11:顔画像判定部
 12:顔表情分析情報記憶部
 20:チャット取得部
 21:チャット分析部
 30:聴衆状態推定部
 40:遷移先決定部
 41:状態遷移ルール記憶部
 42:修正内容選択部
 43:シナリオ修正ルール記憶部
 44:シナリオ修正部
 45:修正履歴記憶部
 5:提示制御装置
 50:シナリオ実行部
 51:シナリオ記憶部
 52:シナリオ通知部
 53:指示制御部
 6:ユーザ端末
 7:表示パネル
 8:カメラ
 100:プレゼンタエージェント
 901:CPU
 902:メモリ
 903:ストレージ
 904:通信装置
 905:入力装置
 906:出力装置
1: Scenario control device 10: Face image acquisition unit 11: Face image determination unit 12: Face expression analysis information storage unit 20: Chat acquisition unit 21: Chat analysis unit 30: Audience state estimation unit 40: Transition destination determination unit 41: Status Transition rule storage unit 42: Correction content selection unit 43: Scenario correction rule storage unit 44: Scenario correction unit 45: Correction history storage unit 5: Presentation control device 50: Scenario execution unit 51: Scenario storage unit 52: Scenario notification unit 53: Instruction control unit 6: User terminal 7: Display panel 8: Camera 100: Presenter agent 901: CPU
902: Memory 903: Storage 904: Communication device 905: Input device 906: Output device

Claims (7)

  1.  シナリオに基づきプレゼンタエージェントの動作を制御してスライド内容をプレゼンさせるプレゼンタシステムにおいて、前記シナリオを変更するシナリオ制御方法であって、
     オンライン会議でプレゼン中に取得した聴衆の顔情報と、前記プレゼン中に前記聴衆が投稿した投稿内容と、を基に、前記プレゼンに対する聴衆全体の興味および態度に関する現在の聴衆状態を推定する第1のステップと、
     前記現在の聴衆状態と、前記シナリオの進行度と、に基づき、前記プレゼンに対する聴衆全体の興味を高め、かつ、聴衆全体の態度を集中させる理想的な聴衆状態を決定する第2のステップと、
     前記現在の聴衆状態を前記理想的な聴衆状態に遷移させる修正用シナリオを決定する第3のステップと、
     前記シナリオを前記修正用シナリオに変更する第4のステップと、
     を行うシナリオ制御方法。
    It is a scenario control method that changes the scenario in the presenter system that controls the operation of the presenter agent based on the scenario and presents the slide contents.
    The first estimation of the current audience status regarding the audience's overall interest and attitude toward the presentation based on the facial information of the audience acquired during the presentation at the online conference and the content posted by the audience during the presentation. Steps and
    Based on the current audience status and the progress of the scenario, a second step of determining an ideal audience status that raises the audience's interest in the presentation and concentrates the attitude of the audience as a whole.
    A third step of determining a correction scenario for transitioning the current audience state to the ideal audience state, and
    The fourth step of changing the scenario to the correction scenario, and
    Scenario control method to do.
  2.  前記第1のステップでは、
     前記プレゼンに対する聴衆全体の興味を高・低で区分けし、前記プレゼンに対する聴衆全体の態度を集中的・発散的でさらに区分けした4つの聴衆状態の中から、前記現在の聴衆状態を推定する請求項1に記載のシナリオ制御方法。
    In the first step,
    A claim that estimates the current audience state from four audience states in which the interest of the entire audience for the presentation is divided into high and low, and the attitude of the entire audience toward the presentation is further divided into intensive and divergent. The scenario control method according to 1.
  3.  前記第2のステップでは、
     前記シナリオの進行度がシナリオの前半である場合、前記プレゼンに対する聴衆全体の興味を高める状態を優先的に決定し、前記シナリオの進行度がシナリオの後半である場合、前記プレゼンに対する聴衆全体の態度を集中させる状態を優先的に決定する請求項1または2に記載のシナリオ制御方法。
    In the second step,
    If the progress of the scenario is the first half of the scenario, the state that raises the interest of the whole audience to the presentation is preferentially determined, and if the progress of the scenario is the second half of the scenario, the attitude of the whole audience to the presentation. The scenario control method according to claim 1 or 2, wherein the state of concentrating is preferentially determined.
  4.  前記第3のステップでは、
     前記現在の聴衆状態を前記理想的な聴衆状態に遷移させる状態遷移と同じ状態遷移が過去に行われていた場合、前記過去に決定した修正用シナリオとは別の修正用シナリオを決定する請求項1ないし3のいずれかに記載のシナリオ制御方法。
    In the third step,
    A claim for determining a modification scenario different from the modification scenario determined in the past when the same state transition as the state transition for transitioning the current audience state to the ideal audience state has been performed in the past. The scenario control method according to any one of 1 to 3.
  5.  前記修正用シナリオは、
     前記プレゼンタエージェントの発話音声と非言語動作とのうち一方または両方を修正するシナリオである請求項1ないし4のいずれかに記載のシナリオ制御方法。
    The correction scenario is
    The scenario control method according to any one of claims 1 to 4, which is a scenario for modifying one or both of the spoken voice and the non-verbal operation of the presenter agent.
  6.  シナリオに基づきプレゼンタエージェントの動作を制御してスライド内容をプレゼンさせるプレゼンタシステムにおいて、前記シナリオを変更するシナリオ制御装置であって、
     オンライン会議でプレゼン中に取得した聴衆の顔情報と、前記プレゼン中に前記聴衆が投稿した投稿内容と、を基に、前記プレゼンに対する聴衆全体の興味および態度に関する現在の聴衆状態を推定する推定部と、
     前記現在の聴衆状態と、前記シナリオの進行度と、に基づき、前記プレゼンに対する聴衆全体の興味を高め、かつ、聴衆全体の態度を集中させる理想的な聴衆状態を決定する第1の決定部と、
     前記現在の聴衆状態を前記理想的な聴衆状態に遷移させる修正用シナリオを決定する第2の決定部と、
     前記シナリオを前記修正用シナリオに変更する変更部と、
     を備えるシナリオ制御装置。
    A scenario control device that changes the scenario in a presenter system that controls the operation of the presenter agent based on the scenario to present the slide contents.
    An estimation unit that estimates the current audience status regarding the audience's overall interest and attitude toward the presentation based on the facial information of the audience acquired during the presentation at the online conference and the content posted by the audience during the presentation. When,
    Based on the current audience status and the progress of the scenario, the first decision-making part that determines the ideal audience status that raises the audience's interest in the presentation and concentrates the attitude of the entire audience. ,
    A second determinant that determines a correction scenario that transitions the current audience state to the ideal audience state.
    A change part that changes the scenario to the correction scenario,
    A scenario control device.
  7.  請求項1ないし5のいずれかに記載のシナリオ制御方法をコンピュータに実行させるシナリオ制御プログラム。 A scenario control program that causes a computer to execute the scenario control method according to any one of claims 1 to 5.
PCT/JP2020/045856 2020-12-09 2020-12-09 Scenario control method, scenario control device, and scenario control program WO2022123688A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2020/045856 WO2022123688A1 (en) 2020-12-09 2020-12-09 Scenario control method, scenario control device, and scenario control program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2020/045856 WO2022123688A1 (en) 2020-12-09 2020-12-09 Scenario control method, scenario control device, and scenario control program

Publications (1)

Publication Number Publication Date
WO2022123688A1 true WO2022123688A1 (en) 2022-06-16

Family

ID=81973376

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/045856 WO2022123688A1 (en) 2020-12-09 2020-12-09 Scenario control method, scenario control device, and scenario control program

Country Status (1)

Country Link
WO (1) WO2022123688A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017073107A (en) * 2015-10-08 2017-04-13 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America Control method for information presentation device, and information presentation device
JP2019186780A (en) * 2018-04-12 2019-10-24 富士通株式会社 User support program, user support apparatus, and user support method
JP6605174B1 (en) * 2019-06-26 2019-11-13 グリー株式会社 Computer program, information processing method, and moving image distribution system
JP2020086774A (en) * 2018-11-21 2020-06-04 日本電信電話株式会社 Apparatus, method and program for controlling scenario

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017073107A (en) * 2015-10-08 2017-04-13 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America Control method for information presentation device, and information presentation device
JP2019186780A (en) * 2018-04-12 2019-10-24 富士通株式会社 User support program, user support apparatus, and user support method
JP2020086774A (en) * 2018-11-21 2020-06-04 日本電信電話株式会社 Apparatus, method and program for controlling scenario
JP6605174B1 (en) * 2019-06-26 2019-11-13 グリー株式会社 Computer program, information processing method, and moving image distribution system

Similar Documents

Publication Publication Date Title
Berdasco et al. User experience comparison of intelligent personal assistants: Alexa, Google Assistant, Siri and Cortana
JP6776462B2 (en) Automatic assistant with meeting ability
US11417330B2 (en) Determining conversation analysis indicators for a multiparty conversation
US7607097B2 (en) Translating emotion to braille, emoticons and other special symbols
US11074928B2 (en) Conversational analytics
US20140036023A1 (en) Conversational video experience
US11521620B2 (en) Synthesizing higher order conversation features for a multiparty conversation
US20160004299A1 (en) Systems and methods for assessing, verifying and adjusting the affective state of a user
Liu et al. An investigation into back-end advancements for speaker recognition in multi-session and noisy enrollment scenarios
US11417318B2 (en) Computationally reacting to a multiparty conversation
US11810357B2 (en) Segmenting and generating conversation features for a multiparty conversation
US20190348063A1 (en) Real-time conversation analysis system
Pugh et al. Say What? Automatic Modeling of Collaborative Problem Solving Skills from Student Speech in the Wild.
CN114097025A (en) Two-pass end-to-end speech recognition
US20230222932A1 (en) Methods, systems, and media for context-aware estimation of student attention in online learning
Lasecki et al. Scribe: deep integration of human and machine intelligence to caption speech in real time
JP2022074024A (en) Method, computer program product, and computer system (personal speech proposal using audience feedback)
WO2022123688A1 (en) Scenario control method, scenario control device, and scenario control program
US10120932B2 (en) Social capture rules
US11514924B2 (en) Dynamic creation and insertion of content
CN117076635A (en) Information processing method, apparatus, device and storage medium
JP7185179B2 (en) Facilitation support system and facilitation support program
Yu et al. Inferring human interactions in meetings: a multimodal approach
JP7490538B2 (en) Information processing terminal, user support method and program
US11929095B1 (en) Speed adjustment of recorded audio and video to maximize desirable cognitive effects for the audience

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20965071

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20965071

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP