WO2021090750A1 - Dispositif de traitement d'informations, procédé de traitement d'informations et programme - Google Patents

Dispositif de traitement d'informations, procédé de traitement d'informations et programme Download PDF

Info

Publication number
WO2021090750A1
WO2021090750A1 PCT/JP2020/040558 JP2020040558W WO2021090750A1 WO 2021090750 A1 WO2021090750 A1 WO 2021090750A1 JP 2020040558 W JP2020040558 W JP 2020040558W WO 2021090750 A1 WO2021090750 A1 WO 2021090750A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
message
face image
information processing
voice
Prior art date
Application number
PCT/JP2020/040558
Other languages
English (en)
Japanese (ja)
Inventor
貴司 丸山
Original Assignee
株式会社 資生堂
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社 資生堂 filed Critical 株式会社 資生堂
Publication of WO2021090750A1 publication Critical patent/WO2021090750A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/802D [Two Dimensional] animation, e.g. using sprites
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/63Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present invention relates to an information processing device, an information processing method, and a program.
  • Rev. 2012-017786 discloses a technique for promoting conversation based on factual information (preference information, similarity, and action history).
  • the actual conversation is developed interactively according to the combination of the reactions of each participant in the conversation during the conversation (for example, the facial expression of the conversation partner and the meaning of the words spoken by the conversation partner).
  • Retable 2012-017786 does not affect the development of the conversation according to the combination of reactions during the conversation because it refers only to the factual information that has been confirmed at the time before the conversation. .. Therefore, it does not contribute to the promotion of conversation among a plurality of users.
  • An object of the present invention is to promote conversation among a plurality of users.
  • One aspect of the present invention is An information processing device that supports conversations between multiple users.
  • a means for acquiring the first face image of the first user is provided.
  • a means for acquiring the first voice of the first user is provided.
  • a means for acquiring the second face image of the second user is provided.
  • a means for generating the second voice of the second user is provided.
  • a means for generating a first predicted face image of the first user at a first designated age aged or decreased from the age of the first user based on the first face image is provided.
  • a means for generating a second predicted face image of the second user at a second designated age aged or decreased from the age of the second user based on the second face image is provided.
  • a means for presenting the first predicted face image to the second user is provided.
  • a means for presenting the second predicted face image to the first user is provided.
  • a means for generating a message prompting the first user and at least one of the second users to speak is provided.
  • a means for presenting the message to at least one of the first user and the second user is provided.
  • a means for presenting a second voice corresponding to the message to the first user is provided.
  • a means for presenting the first voice corresponding to the message to the second user is an information processing device.
  • FIG. 7 It is a block diagram which shows the structure of the information processing system of this embodiment. It is a functional block diagram of the 1st client apparatus of FIG. It is a functional block diagram of the 2nd client apparatus of FIG. It is a functional block diagram of the host device of FIG. It is a figure which shows an example of the form of the information processing system of FIG. It is explanatory drawing of the outline of this embodiment. It is a sequence diagram of information processing of this embodiment. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which shows the screen example displayed in the information processing of FIG. 7. It is a figure which
  • FIG. 1 is a block diagram showing a configuration of an information processing system of the present embodiment.
  • the information processing system 1 includes a first client device 10, a second client device 20, and a host device 30.
  • the host device 30 is communicably connected to the first client device 10 and the second client device 20.
  • the first client device 10 is a computer (an example of an "information processing device") that transmits a request to the host device 30.
  • the first client device 10 is, for example, a smartphone, a tablet terminal, or a personal computer.
  • the host device 30 is a computer (an example of an "information processing device") that provides the first client device 10 with a response in response to a request transmitted from the first client device 10.
  • the host device 30 is, for example, a web host device.
  • FIG. 2 is a functional block diagram of the first client device of FIG.
  • the first client device 10 includes a storage device 11, a processor 12, an input / output interface 13, a communication interface 14, a camera 15, a display 16, a microphone 17, a speaker 18, and the like. To be equipped with.
  • the storage device 11 is configured to store programs and data.
  • the storage device 11 is, for example, a combination of a ROM (Read Only Memory), a RAM (Random Access Memory), and a storage (for example, a flash memory or a hard disk).
  • the program includes, for example, the following program. ⁇ OS (Operating System) program ⁇ Application (for example, web browser) program that executes information processing
  • the data includes, for example, the following data.
  • -Database referenced in information processing-Data obtained by executing information processing that is, the execution result of information processing
  • the processor 12 is configured to realize the function of the first client device 10 by activating the program stored in the storage device 11.
  • the processor 12 is, for example, a CPU (Central Processing Unit), an ASIC (Application Specific Integrated Circuit), an FPGA (Field Programmable Gate Array), or a combination thereof.
  • the input / output interface 13 is configured to acquire a user's instruction from an input device connected to the first client device 10 and output information to an output device connected to the first client device 10.
  • the input device is, for example, a hard key, a keyboard, a pointing device, a touch panel, or a combination thereof.
  • the input device also includes a camera 15 and a display 16.
  • Output devices include a display 16 and a speaker 18.
  • the communication interface 14 is configured to control communication between the first client device 10 and the host device 30.
  • the camera 15 is configured to acquire an image.
  • the display 16 is configured to display an image generated by the processor 12.
  • the microphone 17 is configured to acquire voice.
  • the speaker 18 is configured to output the sound generated by the processor 12.
  • FIG. 3 is a functional block diagram of the second client device of FIG.
  • the second client device 20 includes a storage device 21, a processor 22, an input / output interface 23, a communication interface 24, a camera 25, a display 26, a microphone 27, a speaker 28, and the like. To be equipped with.
  • the storage device 21 is configured to store programs and data.
  • the storage device 21 is, for example, a combination of a ROM, a RAM, and a storage.
  • the program includes, for example, the following program. -OS program-Program of an application (for example, a web browser) that executes information processing
  • the data includes, for example, the following data.
  • -Database referenced in information processing-Data obtained by executing information processing that is, the execution result of information processing
  • the processor 22 is configured to realize the function of the second client device 20 by activating the program stored in the storage device 21.
  • the processor 22 is, for example, a CPU, an ASIC, an FPGA, or a combination thereof.
  • the input / output interface 23 is configured to acquire a user's instruction from an input device connected to the second client device 20 and output information to an output device connected to the second client device 20.
  • the input device is, for example, a hard key, a keyboard, a pointing device, a touch panel, or a combination thereof.
  • the input device also includes a camera 25 and a microphone 27.
  • Output devices include a display 26 and a speaker 28.
  • the communication interface 24 is configured to control communication between the second client device 20 and the host device 30.
  • the camera 25 is configured to acquire an image.
  • the display 26 is configured to display an image generated by the processor 22.
  • the microphone 27 is configured to acquire voice.
  • the speaker 28 is configured to output the voice generated by the processor 22.
  • FIG. 4 is a functional block diagram of the host device of FIG.
  • the host device 30 includes a storage device 31, a processor 32, an input / output interface 33, and a communication interface 34.
  • the storage device 31 is configured to store programs and data.
  • the storage device 31 is, for example, a combination of a ROM, a RAM, and a storage (for example, a flash memory or a hard disk).
  • the program includes, for example, the following program. ⁇ OS program ⁇ Application program that executes information processing
  • the data includes, for example, the following data. ⁇ Database referenced in information processing ⁇ Execution result of information processing
  • the processor 32 is configured to realize the function of the host device 30 by activating the program stored in the storage device 31.
  • the processor 32 is, for example, a CPU, an ASIC, an FPGA, or a combination thereof.
  • the input / output interface 33 is configured to acquire a user's instruction from an input device connected to the host device 30 and output information to an output device connected to the host device 30.
  • the input device is, for example, a keyboard, a pointing device, a touch panel, or a combination thereof.
  • the output device is, for example, a display.
  • the output device also includes a printer.
  • the communication interface 34 is configured to control communication between the host device 30 and the first client device 10 and the second client device 20.
  • FIG. 5 is a diagram showing an example of the form of the information processing system of FIG.
  • the information processing system 1 includes the first space SP1 to the third space SP3.
  • the third space SP is arranged between the first space SP1 and the second space SP2.
  • the first space SP1 is a space in which the first user U1 enters.
  • the first user U1 can enter the first space SP1 from the entrance ENT1.
  • a camera 15, a display 16, a microphone 17, and a speaker 18 are arranged in the first space SP1.
  • a shielding portion (for example, a wall) W1 is arranged at the boundary between the first space SP1 and the third space SP.
  • the display 16 is arranged on the shielding portion W1 so as to face the X- direction.
  • the first user U1 can visually recognize the image displayed on the display 16 by directing the line of sight in the X + direction.
  • the second space SP2 is a space in which the second user U2 enters.
  • the second user U2 can enter the second space SP2 from the entrance ENT2.
  • a camera 25, a display 26, a microphone 27, and a speaker 28 are arranged in the second space SP2.
  • a shielding portion W2 is arranged at the boundary between the second space SP2 and the third space SP.
  • the display 26 is arranged on the shielding portion W2 so as to face the X + direction.
  • the second user U2 can visually recognize the image displayed on the display 26 by directing the line of sight in the X- direction.
  • the shielding portion W1 is configured to visually separate the first space SP1 and the second space SP2. Therefore, the first user U1 and the second user U2 cannot directly recognize each other's voice and facial expression.
  • a host device 30 is arranged in the third space SP.
  • FIG. 6 is an explanatory diagram of an outline of the present embodiment.
  • the host device 30 in FIG. 1 is an information processing device that supports conversations between a plurality of users.
  • the host device 30 includes the following means. -Means for acquiring the first face image (plural still images or moving images) of the first user-Means for acquiring the first sound of the first user-Second face image (plural still images or moving images) of the second user Means for acquiring an image) -Means for acquiring a second voice of a second user-A first user in a first designated age who has been aged or diminished from the age of the first user based on a first face image.
  • FIG. 7 is a sequence diagram of information processing of the present embodiment.
  • FIG. 8 is a diagram showing an example of a screen displayed in the information processing of FIG. 7.
  • FIG. 9 is a diagram showing an example of a screen displayed in the information processing of FIG. 7.
  • FIG. 10 is a diagram showing an example of a screen displayed in the information processing of FIG. 7.
  • FIG. 11 is a diagram showing an example of a screen displayed in the information processing of FIG. 7.
  • FIG. 12 is a diagram showing an example of a screen displayed in the information processing of FIG. 7.
  • the first client device 10 executes the reception of user information (S110). Specifically, the processor 12 displays the screen P10 (FIG. 8) on the display 16.
  • the screen P10 includes field objects F10a to F10e and operation objects B10.
  • the field object F10a is an object that accepts user input regarding the name of the first user.
  • the field object F10b is an object that accepts user input regarding the relationship between the first user and the second user.
  • the relationship includes, for example, at least one of the following: ⁇ Family (for example, at least one of parent and child, couple, sibling, and relative) ⁇ Dating (that is, a couple) ⁇ Friends ⁇ Colleagues at work
  • the field object F10c is an object that accepts user input regarding the duration of the relationship between the first user and the second user (hereinafter referred to as "relationship period").
  • the relationship period is, for example, one of the following. ⁇ Less than 5 years ⁇ 5 years or more
  • the field object F10d is an object that accepts user input regarding the gender of the first user.
  • the field object F10e is an object that accepts user input regarding the date of birth of the first user.
  • the operation object B10 is an object that receives a user instruction for confirming the user input for the field objects F10a to F10e.
  • the camera 15 acquires an image of the face of the first user (hereinafter referred to as “first user image”).
  • the processor 12 transmits the combination of the information input to the field objects F10a to F10e and the first user image (hereinafter referred to as “first user information”) to the host device 30.
  • the second client device 20 executes the reception of user information (S120). Specifically, the processor 22 displays the screen P20 (FIG. 8) on the display 26.
  • the screen P20 includes field objects F20a to F20e and operation objects B20.
  • the field object F20a is an object that accepts user input regarding the name of the second user.
  • the field object F20b is an object that accepts user input regarding the relationship between the first user and the second user.
  • the field object F20c is an object that accepts user input regarding the period of relationship between the first user and the second user.
  • the field object F20d is an object that accepts user input regarding the gender of the second user.
  • the field object F20e is an object that accepts user input regarding the date of birth of the second user.
  • the operation object B20 is an object that receives a user instruction for confirming a user input for the field objects F20a to F20e.
  • the camera 25 acquires an image of the face of the second user (hereinafter referred to as “second user image”).
  • the processor 22 transmits the combination of the information input to the field objects F20a to F20e and the second user image (hereinafter referred to as “second user information”) to the host device 30.
  • the host device 30 executes the transfer of user information (S130). Specifically, the processor 32 acquires the first user information from the first client device 10. The processor 32 acquires the second user information from the second client device 20. The processor 32 calculates the value of the current age of the first user (hereinafter referred to as "first current age”) based on the date of birth included in the first user information. The processor 32 calculates the value of the current age of the second user (hereinafter referred to as "second current age”) based on the date of birth included in the second user information. The processor 32 transmits the first user information to the second client device 20. The processor 32 transmits the second user information to the first client device 10.
  • the first client device 10 executes the reception (S111) of the designated age. Specifically, the processor 12 displays the screen P11 (FIG. 9) on the display 16.
  • the screen P11 includes a display object A11, an image object IMG11, field objects F11a to F11b, and an operation object B11.
  • a message prompting the first user to have a conversation (for example, a message "Please select the age of the other party") is displayed on the display object A11. This message is a predetermined message.
  • the image object IMG 11 is a second user image included in the second user information transmitted from the host device 30.
  • the field object F11a is an object that accepts user input for designating a fictitious age of the first user (hereinafter referred to as "first designated age").
  • the initial value of the field object F11a is the value of the first current age.
  • the field object F11b is an object that accepts user input for designating a fictitious age of the second user (hereinafter referred to as "second designated age").
  • the initial value of the field object F11b is the value of the second current age.
  • the operation object B11a is an object that receives a user instruction for confirming the user input of the field objects F11a to F11b.
  • the processor 12 transmits the value given to the field object F11b to the host device 30.
  • the second client device 20 executes the reception (S121) of the designated age. Specifically, the processor 22 displays the screen P21 (FIG. 9) on the display 26.
  • the screen P21 includes a display object A21, an image object IMG21, field objects F21a to F21b, and operation objects B21a to B21b.
  • a message prompting the second user to have a conversation (for example, a message "Please select the age of the other party") is displayed on the display object A21.
  • This message is a predetermined message.
  • the image object IMG 21 is a first user image included in the first user information transmitted from the host device 30.
  • the field object F21a is an object that accepts user input for designating the first designated age.
  • the initial value of the field object F21a is the value of the first current age.
  • the value of the field object F21a is synchronized with the value of the field object F11a.
  • the field object F21b is an object that accepts user input for designating the second designated age.
  • the initial value of the field object F21b is a value calculated from the date of birth input to the field object F20e (FIG. 8) (that is, the current age of the first user).
  • the value of the field object F21b is synchronized with the value of the field object F11b.
  • the operation object B21a is an object that receives a user instruction for confirming the user input of the field objects F21a to F21b.
  • the operation object B21b is an object that receives a user instruction for ending the conversation.
  • the processor 12 When the second user inputs an arbitrary value (for example, a fictitious age value of the first user) into the field object F21b and operates the operation object B21a, the processor 12 causes the value given to the field object F21b (for example). That is, the fictitious age of the first user (hereinafter referred to as "first designated age”) is transmitted to the host device 30.
  • an arbitrary value for example, a fictitious age value of the first user
  • first designated age the fictitious age of the first user
  • the host device 30 performs morphing (S131). Specifically, the conversion model is stored in the storage device 31.
  • the conversion model defines the correlation between the combination of the current age of the person and the fictitious age of the person and the conversion filter for converting the face image of the person.
  • the processor 32 acquires the value of the second designated age given to the field object F11b from the first client device 10.
  • the processor 32 acquires the value of the first designated age given to the field object F21a from the second client device 20.
  • the processor 32 gives a combination of the first current age and the first designated age by giving the value of the first current age obtained in step S130 and the value of the first designated age obtained in step S111 to the conversion model. Select the conversion filter that corresponds to.
  • the processor 32 applies the selected conversion filter to the first user image to generate an image of the face of the first user corresponding to the first designated age (hereinafter referred to as “first predicted face image”).
  • the processor 32 gives a combination of the second current age and the second designated age by giving the value of the second current age obtained in step S130 and the value of the second designated age obtained in step S121 to the conversion model. Select the conversion filter that corresponds to.
  • the processor 32 applies the selected conversion filter to the second user image to generate an image of the face of the second user corresponding to the second designated age (hereinafter referred to as “second predicted face image”).
  • step S131 the host device 30 executes message transmission (S132).
  • the processor 32 transmits the first message data to the first client device 10.
  • the first message data includes the following information. ⁇ Second predicted face image ⁇ Predetermined initial message
  • the processor 32 transmits the second message data to the second client device 20.
  • the second message data includes the following information. ⁇ First predicted face image ⁇ Predetermined initial message
  • step S132 the first client device 10 executes the reception of conversation information (S112).
  • the processor 12 displays the screen P12a (FIG. 10) on the display 16.
  • the screen P12a corresponds to the first message data.
  • the screen P12a includes a display object A12a, an image object IMG12, field objects F11a to F11b, and an operation object B11.
  • the display object A12a displays the initial message included in the first message data.
  • the image object IMG12 is a second predicted face image included in the first message data.
  • the second user outputs from the speaker 18 while the first user sees the message displayed on the display object A12a and the image object IMG12 (that is, the image of the face of the second user corresponding to the second designated age).
  • the microphone 17 acquires the voice of the first user (hereinafter referred to as "first spoken voice”).
  • the camera 15 acquires a face image of the first user in conversation (hereinafter referred to as "first face image”).
  • the processor 12 transmits the conversation information to the host device 30.
  • the conversation information includes the following information. -First face image acquired by camera 15-First utterance voice acquired by microphone 17
  • step S132 the second client device 20 executes the reception of conversation information (S122).
  • the processor 22 displays the screen P22a (FIG. 10) on the display 26.
  • the screen P22a corresponds to the second message data.
  • the screen P22a includes display objects A22a, image objects IMG22, field objects F21a to F21b, and operation objects B21a to B21b.
  • the display object A22a displays the initial message included in the second message data.
  • the image object IMG22 is a first predicted face image included in the second message data.
  • the first user outputs from the speaker 28 while the second user sees the message displayed on the display object A22a and the image object IMG22 (that is, the image of the face of the first user corresponding to the first designated age).
  • the microphone 27 acquires the voice of the second user (hereinafter referred to as "second spoken voice").
  • the camera 25 acquires a face image of the second user in conversation (hereinafter referred to as “second face image”).
  • the processor 22 transmits the conversation information to the host device 30.
  • the conversation information includes the following information. -Second face image acquired by camera 25-Second utterance voice acquired by microphone 27
  • the host device 30 executes emotion analysis (S133).
  • the storage device 31 stores an emotion analysis model in which the correlation between the facial expression of the person and the emotion of the person is defined.
  • the processor 32 identifies the emotion of the first user by inputting the first utterance voice transmitted in step S112 into the emotion analysis model.
  • the processor 32 identifies the emotion of the second user by inputting the second utterance voice transmitted in step S122 into the emotion analysis model.
  • the storage device 31 stores an emotion analysis model in which the correlation between the spoken voice of the person and the emotion of the person is defined.
  • the processor 32 identifies the emotion of the first user by inputting the first utterance voice transmitted in step S112 into the emotion analysis model.
  • the processor 32 identifies the emotion of the second user by inputting the second utterance voice transmitted in step S122 into the emotion analysis model.
  • the host device 30 executes message generation (S134). Specifically, the storage device 31 stores a message generation model.
  • the message generation model defines the correlation between a person's emotions and a message that encourages the next conversation.
  • step S133 the processor 32 generates a message to be presented to the first user by inputting the emotion of the first user obtained in step S133 into the message generation model, and generates the message.
  • the first message data including the first message data is transmitted to the first client device 10.
  • the processor 32 generates a message to be presented to the second user by inputting the emotion of the second user obtained in step S133 into the message generation model, and secondly generates the second message data including the message. It is transmitted to the client device 20.
  • the processor 32 should present the emotions of the first user and the emotions of the second user obtained in step S133 to the first user and the second user by inputting them into the message generation model. A message is generated, and message data including the message is transmitted to the first client device 10 and the second client device 20.
  • the first client device 10 re-executes the reception of conversation information (S112). Specifically, the processor 12 displays the screen P12b (FIG. 11) on the display 16. The screen P12b corresponds to the message data transmitted from the host device 30 in step S132.
  • the screen P12b includes a display object A12b, an image object IMG12, field objects F11a to F11b, and an operation object B11.
  • the message transmitted from the host device 30 in step S133 is displayed on the display object A12b.
  • the second user outputs from the speaker 18 while the first user sees the message displayed on the display object A12b and the image object IMG12 (that is, the image of the face of the second user corresponding to the second designated age).
  • the microphone 17 acquires the first spoken voice.
  • the processor 12 transmits the conversation information to the host device 30.
  • the conversation information includes the following information. -First face image acquired by camera 15-First utterance voice acquired by microphone 17
  • the second client device 20 re-executes the reception of conversation information (S122). Specifically, the processor 22 displays the screen P22b (FIG. 11) on the display 26.
  • the screen P22b corresponds to the message data transmitted from the host device 30 in step S134.
  • the screen P22b includes display objects A22b, image objects IMG22, field objects F21a to F21b, and operation objects B21a to B21b.
  • the message included in the second message data is displayed on the display object A22b.
  • the first user outputs from the speaker 28 while the second user sees the message displayed on the display object A22b and the image object IMG22 (that is, the image of the face of the first user corresponding to the first designated age).
  • the microphone 27 acquires the second spoken voice.
  • the processor 22 transmits the conversation information to the host device 30.
  • the conversation information includes the following information. -Second face image acquired by camera 25-Second utterance voice acquired by microphone 27
  • the first client device 10 executes the acquisition of the commemorative image (S113). Specifically, when the end condition of the conversation is satisfied, the processor 12 displays the screen P13 (FIG. 12) on the display.
  • the end condition of the conversation is, for example, at least one of the following. -A predetermined time (for example, 10 minutes) has elapsed from the start of the conversation (for example, the end of steps S111 and S121) -A predetermined number of messages have been presented (that is, step 134 has been executed) -First A user or second user has given a user instruction (eg, a voice instruction and at least one operation on an object displayed on the screen) to end the conversation.
  • the image object IMG13 including the image object IMG13 and the operation object B13 on the screen P13 is a preview image acquired by the camera 25 at the time of execution of step S113.
  • the operation object B13 is an object that receives a user instruction for acquiring a preview image corresponding to the image object IMG13 as a commemorative image of the second user (hereinafter referred to as “second commemorative image”).
  • the processor 12 transmits the second commemorative image to the host device 30 and sends the end notification of step S113 to the second client device 20.
  • the second client device 20 executes the acquisition of the commemorative image (S123). Specifically, when the first user operates the operation object B13, the processor 12 displays the screen P23 (FIG. 12) on the display.
  • the image object IMG23 including the image object IMG23 and the operation object B23 on the screen P23 is a preview image acquired by the camera 15 at the time of executing step S123.
  • the operation object B23 is an object that receives a user instruction for acquiring a preview image corresponding to the image object IMG23 as a commemorative image of the first user (hereinafter referred to as “first commemorative image”).
  • the processor 22 transmits the first commemorative image and the end notification of step S123 to the host device 30.
  • the host device 30 executes saving of log data (S135).
  • the processor 32 stores the following information in association with the storage device 31. -First user information transmitted from the first client device 10 in step S110-Second user information transmitted from the second client device 20 in step S120-Output request data transmitted from the first client device 10 in step S113 First predicted face image and first voice included in-Second predicted face image and second voice included in output request data transmitted from the second client device 20 in step S123-From execution of steps S111 and S121 to step S113 And the time to at least one execution of S123 (ie, elapsed time of conversation) -Second commemorative image transmitted from the first client device 10 in step S113-First commemorative image transmitted from the second client device 20 in step S123.
  • S123 ie, elapsed time of conversation
  • the processor 32 generates a set commemorative image by synthesizing the first commemorative image and the second commemorative image.
  • the processor 32 transmits the first commemorative image, the second commemorative image, and the collective commemorative image to the first client device 10 and the second client device 20.
  • step S135 the first client device 10 executes an output request (S114). Specifically, the processor 12 displays the screen P14 (FIG. 12) on the display 16.
  • the screen P14 includes an image object IMG14 and operation objects B14a to B14b.
  • the image object IMG14 is a group commemorative image.
  • the operation object B14a is an object that receives a user instruction for sharing the group commemorative image on the network service.
  • the network service is, for example, at least one of a social network service, a chat service, and an image sharing service.
  • the operation object B14b is an object that receives a user instruction for printing a set commemorative image.
  • the processor 12 transmits a user instruction for sharing the group commemorative image to the host device.
  • the second client device 20 executes the output request (S124). Specifically, when the second user operates the operation object B21b (FIGS. 10 to 11), the processor 22 displays the screen P24 (FIG. 12) on the display.
  • the screen P24 includes an image object IMG24 and operation objects B24a to B24b.
  • the image object IMG24 is a group commemorative image.
  • the operation object B24a is an object that receives a user instruction for sharing the group commemorative image on the network service.
  • the network service is, for example, at least one of a social network service, a chat service, and an image sharing service.
  • the operation object B24b is an object that receives a user instruction for printing a group commemorative image.
  • the processor 22 transmits a user instruction for printing the group commemorative image to the host device 30.
  • the host device 30 executes the output of the commemorative image (S136). Specifically, the processor 32 transmits the group commemorative image to the web server that provides the network service in response to the user instruction transmitted from the first client device 10 in step S113. As a result, the group commemorative image is shared with users other than the first user (for example, at least one of the second user and a third party) via the network service. The processor 32 transmits the group commemorative image to the printer in response to the user instruction transmitted from the second client device 20 in step S123. As a result, the group commemorative image is printed via the printer.
  • each user is presented with a predicted face image of the conversation partner and a message for promoting the conversation during the conversation.
  • a predicted face image of the conversation partner and a message for promoting the conversation during the conversation.
  • the first modification is an example in which a message corresponding to at least one of the context of the first voice and the context of the second voice is presented to at least one of the first user and the second user.
  • step S134 the host device 30 of the first modification generates a message corresponding to at least one of the contexts of the first voice and the second voice.
  • the message generation model defines the correlation between the voice context and the message that prompts the next conversation.
  • the processor 32 of the first example of the modification 1 generates a message to be presented to the first user by inputting the first voice obtained in step S133 into the message generation model, and includes the message. 1 Message data is transmitted to the first client device 10. The processor 32 generates a message to be presented to the second user by inputting the second voice obtained in step S133 into the message generation model, and generates the second message data including the message in the second client device. Send to 20.
  • the processor 32 of the second example of the modification 1 inputs the combination of the first voice and the second voice obtained in step S133 into the message generation model, so that the message to be presented to the first user and the second user can be presented.
  • the message data that is generated and includes the message is transmitted to the first client device 10 and the second client device 20.
  • a message prompting a conversation is presented according to the context of the voice spoken by at least one of the first user and the second user. Thereby, the conversation between the first user and the second user can be further promoted.
  • the second modification is an example in which a message corresponding to at least one of the environment of the first user and the environment of the second user is presented to at least one of the first user and the second user.
  • the host device 30 of the second modification includes information on the environment of the first user (hereinafter referred to as “first environment information”) and information on the environment of the second user (hereinafter referred to as “second environment information””. Generate a message corresponding to at least one of). Specifically, the message generation model defines the correlation between the environment and the message that encourages the next conversation.
  • the information processing system 1 includes a sensor that acquires environmental information.
  • Environmental information includes at least one of the following: -Temperature information regarding the temperature of the first space SP1-Temperature information regarding the temperature of the second space SP2-Humidity information regarding the humidity of the first space SP1-Humidity information regarding the temperature of the second space SP2
  • the processor 32 of the first example of the modification 2 generates a message to be presented to the first user by inputting the first environment information obtained by the sensor into the message generation model, and includes the message. 1 Message data is transmitted to the first client device 10. The processor 32 generates a message to be presented to the second user by inputting the second environment information obtained by the sensor into the message generation model, and generates the second message data including the message as the second client device. Send to 20.
  • the processor 32 of the second example of the second modification is a message to be presented to the first user and the second user by inputting the combination of the first environment information and the second environment information obtained by the sensor into the message generation model. Is generated, and message data including the message is transmitted to the first client device 10 and the second client device 20.
  • At least one of the first space SP1 and the second space SP2 is outdoors.
  • Environmental information includes at least one of the following: -Temperature information regarding the temperature of the first space SP1-Temperature information regarding the temperature of the second space SP2
  • the host device 30 may acquire at least one of the first environmental information and the second environmental information from a device other than the sensor (for example, a server that provides the environmental information) instead of the sensor.
  • a device other than the sensor for example, a server that provides the environmental information
  • a message prompting a conversation is presented according to the environment of at least one of the first user and the second user. Thereby, the conversation between the first user and the second user can be further promoted.
  • the third modification is an example in which a message corresponding to at least one of the attributes of the first user and the attribute of the second user is presented to at least one of the first user and the second user.
  • step S134 the host device 30 of the third modification has information on the attributes of the first user (hereinafter referred to as “first attribute information”) and information on the attributes of the second user (hereinafter referred to as “second attribute information””. Generate a message corresponding to at least one of). Specifically, the message generation model defines the correlation between the attributes and the message for prompting the next conversation.
  • the attribute includes at least one of the following: -Current age of the first user-Current age of the second user-Gender of the first user-Gender of the second user
  • the processor 32 of the first example of the modification 3 inputs the attribute of the first user included in the first user information transmitted from the first client device 10 to the message generation model in step S110 to the first user.
  • a message to be presented is generated, and the first message data including the message is transmitted to the first client device 10.
  • the processor 32 generates a message to be presented to the second user by inputting the attributes of the second user included in the second user information transmitted from the second client device 20 in step S120 into the message generation model.
  • the second message data including the message is transmitted to the second client device 20.
  • the processor 32 of the second example of the modification 3 has the attributes of the first user included in the first user information transmitted from the first client device 10 in step S110, and is transmitted from the second client device 20 in step S120.
  • the processor 32 of the second example of the modification 3 has the attributes of the first user included in the first user information transmitted from the first client device 10 in step S110, and is transmitted from the second client device 20 in step S120.
  • the message generation model By inputting the combination of the attributes of the second user included in the second user information into the message generation model, a message to be presented to the first user and the second user is generated, and the message data including the message is generated. It transmits to the first client device 10 and the second client device 20.
  • a message prompting a conversation is presented according to the attributes of at least one of the first user and the second user. Thereby, the conversation between the first user and the second user can be further promoted.
  • the fourth modification is an example in which a message corresponding to the relationship between the first user and the second user is presented to at least one of the first user and the second user.
  • step S134 the host device 30 of the modification 4 generates a message according to the relationship between the first user and the second user.
  • the message generation model defines the correlation between the relationship between a plurality of users and the message for prompting the next conversation.
  • the processor 32 uses the relationship included in at least one of the first user information transmitted from the first client device 10 in step S110 and the second user information transmitted from the second client device 20 in step S120 as the message generation model. By inputting, a message to be presented to the first user and the second user is generated, and message data including the message is transmitted to the first client device 10 and the second client device 20.
  • a message prompting a conversation is presented according to the relationship between the first user and the second user. Thereby, the conversation between the first user and the second user can be further promoted.
  • the fifth modification is an example in which a message corresponding to at least one of the first designated age and the second designated age is presented to at least one of the first user and the second user.
  • step S134 the host device 30 of the modification 5 generates a message corresponding to at least one of the first designated age and the second designated age.
  • the message generation model defines the correlation between age and a message to encourage the next conversation.
  • the processor 32 of the first example of the modification 5 has the value of the first designated age transmitted from the first client device 10 in step S111 and the value of the second designated age transmitted from the second client device 20 in step S121. Judge the magnitude relationship of values.
  • the processor 32 generates a message to be presented to the first user and the second user by inputting the magnitude relation into the message generation model, and the message data including the message is transmitted to the first client device 10 and the second user. It is transmitted to the client device 20.
  • the processor 32 of the second example of the modification 5 has the value of the first designated age transmitted from the first client device 10 in step S111 and the value of the second designated age transmitted from the second client device 20 in step S121. Calculate the difference in values (hereinafter referred to as "specified age difference").
  • the processor 32 generates a message to be presented to the first user and the second user by inputting the designated age difference into the message generation model, and generates message data including the message in the first client device 10 and the first client device 10. 2 Transmit to the client device 20.
  • a message prompting a conversation is presented according to the designated age of at least one of the first user and the second user. Thereby, the conversation between the first user and the second user can be further promoted.
  • Modification 6 A modification 6 will be described. Modification 6 is an example of playing back the replay moving images of the first user and the second user.
  • the processor 12 or 22 of the modification 6 transmits the replay request data for requesting the reproduction of the replay moving image to the host device 30 according to the instruction of the user.
  • the processor 32 generates a replay moving image by synthesizing the following information stored in the storage device 31 according to the replay request data transmitted from the first client device 10 or the second client device 20.
  • ⁇ 1st face image ⁇ 1st predicted face image ⁇ 1st voice
  • 2nd face image ⁇ 2nd predicted face image ⁇ 2nd voice
  • the processor 32 transmits the replay moving image to the first client device 10 or the second client device 20.
  • the processor 12 or 22 reproduces the replay moving image transmitted from the first client device 10 or the second client device 20. As a result, the image of the replay moving image is displayed on the display 16 or 26. The sound of the replay moving image is output from the speaker 18 or 28.
  • a replay video that is a record of conversations between the first user and the second user is provided. This can give the user motivation for conversation.
  • the processor 12 or 22 may save the replay moving image file in the host device 30.
  • the host device 30 is a web server
  • the host device 30 assigns a URL (Uniform Resource Locator) to the replay video file.
  • the host device 30 is used by the first user, the second user, and the user who made the browsing request in response to the browsing request of the third user (that is, a third party) different from the first user and the second user.
  • the replay video is presented via the client device.
  • the browsing request is, for example, an access to the address of the web server (for example, a URL (Uniform Resource Locator)) in which the replay video file is stored.
  • the replay moving image can be provided to at least one of the first user to the third user after the conversation is completed.
  • the first aspect of this embodiment is An information processing device (for example, a host device 30) that supports conversations between a plurality of users.
  • a means for acquiring the first face image of the first user (processor 32 that executes the process of step S130) is provided.
  • a means for acquiring the first voice of the first user (processor 32 that executes the process of step S133) is provided.
  • a means for acquiring a second face image of a second user is provided.
  • a means for acquiring the second voice of the second user is provided.
  • a means for presenting the first predicted face image to the second user (processor 32 that executes the process of step S132) is provided.
  • a means for presenting the second predicted face image to the first user (processor 32 that executes the process of step S132) is provided.
  • a means for generating a message prompting at least one of the first user and the second user to speak (processor 32 that executes the process of step S132) is provided.
  • a means for presenting a message to at least one of a first user and a second user (processor 32 that executes the process of step S132) is provided.
  • a means (processor 32 that executes the process of step S133) for presenting the second voice corresponding to the message to the first user is provided.
  • a means (processor 32 that executes the process of step S133) for presenting the first voice corresponding to the message to the second user is provided. It is an information processing device.
  • the first predicted image corresponding to the first designated age of the first user is presented to the second user
  • the second predicted image corresponding to the second designated age of the second user is presented to the first user.
  • the second aspect of this embodiment is A means for analyzing the emotion of the first user based on the first face image (processor 32 that executes the process of step S133) is provided.
  • a means for analyzing the emotion of the second user based on the second face image (processor 32 that executes the process of step S133) is provided.
  • the means for generating a message generates a message according to at least one of the emotions of the first user and the emotions of the second user. It is an information processing device.
  • a message corresponding to at least one of the emotions of the first user and the emotions of the second user is presented to at least one of the first user and the second user.
  • the conversation between the first user and the second user can be further promoted.
  • the third aspect of this embodiment is A means for analyzing the emotion of the first user based on the first voice (processor 32 that executes the process of step S133) is provided.
  • the means for generating a message generates a message according to at least one of the emotions of the first user and the emotions of the second user. It is an information processing device.
  • a message corresponding to at least one of the emotions of the first user and the emotions of the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the fourth aspect of this embodiment is A means for analyzing the context of the first voice based on the first voice (for example, the processor 32 of the first modification) is provided.
  • a means for analyzing the context of the second voice based on the second voice (for example, the processor 32 of the first modification) is provided.
  • the means for generating a message generates a message according to at least one of a first voice context and a second voice context. It is an information processing device.
  • a message according to the context of the voice spoken by at least one of the first user and the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • a fifth aspect of this embodiment is With a means to acquire the first environment information about the environment of the first user, With a means to acquire the second environment information about the environment of the second user, The means for generating a message generates a message corresponding to at least one of the first environment information and the second environment information. It is an information processing device.
  • a message corresponding to the environment of at least one of the first user and the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the sixth aspect of this embodiment is The first environmental information and the second environmental information include at least one of temperature information, humidity information, and ultraviolet light information. It is an information processing device.
  • a message corresponding to at least one of the temperature, humidity, and ultraviolet rays in the environment of the first user and the temperature, humidity, and ultraviolet rays in the environment of the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the seventh aspect of this embodiment is The means for generating the message generates a message according to the combination of the attributes of the first user and the attributes of the second user. It is an information processing device.
  • a message corresponding to the combination of the attributes of the first user and the attributes of the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the eighth aspect of this embodiment is The attribute includes at least one of the user's age and gender, It is an information processing device.
  • a message corresponding to a combination of at least one of the age and gender of the first user and at least one of the age and gender of the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the ninth aspect of this embodiment is The means for generating a message generates a message according to the relationship between the first user and the second user. It is an information processing device.
  • a message corresponding to the relationship between the first user and the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the tenth aspect of this embodiment is The means for generating a message generates a message according to a combination of a first designated age and a second designated age. It is an information processing device.
  • a message corresponding to the combination of the first designated age of the first user and the second designated age of the second user is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • the eleventh aspect of this embodiment is The means for generating the message presents a message according to at least one of the magnitude relationship between the first designated age and the second designated age and the designated age difference. It is an information processing device.
  • a message corresponding to at least one of the magnitude relationship between the first designated age of the first user and the second designated age of the second user and the designated age difference is presented. Thereby, the conversation between the first user and the second user can be further promoted.
  • a shielding unit W1 is provided at a position that blocks the line of sight of the first user with respect to the second user and the line of sight of the second user with respect to the first user. It is an information processing device.
  • the shielding unit W1 constructs an environment in which the first user and the second user cannot directly recognize each other's facial expressions.
  • the first user and the second user are guided to talk while looking at the second predicted image and the first predicted image, respectively. Thereby, the conversation between the first user and the second user can be further promoted.
  • the thirteenth aspect of this embodiment is The log data including the combination of the first face image, the second face image, the first predicted face image, the second predicted face image, the first voice, and the second voice is stored in association with the elapsed time of the conversation.
  • the means for example, the processor 32 that executes step S135) is provided.
  • a means for generating a replay video using the log data associated with the elapsed time (for example, the processor 32 of the modification 6) is provided.
  • a means for presenting to at least one of a first user, a second user, and a third user other than the first user and the second user (for example, the processor 32 of the modification 6) is provided. It is an information processing device.
  • the replay video is provided.
  • at least one of the first user to the third user can view the state of the conversation after the conversation is completed.
  • the fourteenth aspect of this embodiment is A means for acquiring the first commemorative image of the first user and the second commemorative image of the second user after the end of the conversation (for example, the processor 32 that executes step S135) is provided. It is an information processing device.
  • a combination of the first commemorative image of the first user and the second commemorative image of the second user is acquired. This makes it possible to save a commemorative image of the conversation.
  • the fifteenth aspect of this embodiment is A means for presenting the first predicted face image and the second predicted face image to the first user and the second user (for example, the processor 32 that executes step S130) is provided. It is an information processing device.
  • the combination of the first predicted image and the second predicted image is presented to the first user and the second user. Thereby, the conversation between the first user and the second user can be further promoted.
  • the sixteenth aspect of this embodiment is The sixteenth aspect of this embodiment is An information processing method that supports conversations between multiple users.
  • the step (step S130) for acquiring the first face image of the first user is provided.
  • the step (step S133) for acquiring the first voice of the first user is provided.
  • the step (step S130) for acquiring the second face image of the second user is provided.
  • the step (step S133) for acquiring the second voice of the second user is provided.
  • a step (step S131) of generating a first predicted face image of a first user at a first designated age that has been aged or decreased from the age of the first user based on the first face image is provided.
  • a step (step S131) of generating a second predicted face image of a second user at a second designated age that has been aged or decreased from the age of the second user based on the second face image is provided.
  • a step (step S132) of presenting the first predicted face image to the second user is provided.
  • a step (step S132) of presenting the second predicted face image to the first user is provided.
  • a step (step S132) of generating a message prompting at least one of the first user and the second user to speak is provided.
  • a step (step S132) of presenting a message to at least one of a first user and a second user is provided.
  • the step (step S133) of presenting the second voice corresponding to the message to the first user is provided.
  • a step (step S133) of presenting the first voice corresponding to the message to the second user is provided. It is an information processing method.
  • the seventeenth aspect of this embodiment is A program for causing a computer (for example, at least one of processors 12, 22, and 32) to function as each of the above means.
  • the storage device 11 may be connected to the first client device 10 via the network NW.
  • the storage device 31 may be connected to the host device 30 via the network NW.
  • Each step of the information processing host device 30 described above can also be executed by the first client device 10 and the second client device 20.
  • the first client device 10 and the second client device 20 function as information processing devices that operate standalone without transmitting a request to the host device 30.
  • the host device 30 may be a server (for example, a web server).
  • the first client device 10 and the second client device 20 are connected to the server via, for example, the Internet.
  • Information processing system 10 First client device 11: Storage device 12: Processor 13: Input / output interface 14: Communication interface 15: Camera 16: Display 17: Microphone 18: Speaker 20: Second client device 21: Storage device 22 : Processor 23: Input / output interface 24: Communication interface 25: Camera 26: Display 27: Microphone 28: Speaker 30: Host device 31: Storage device 32: Processor 33: Input / output interface 34: Communication interface

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Psychiatry (AREA)
  • Hospice & Palliative Care (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Child & Adolescent Psychology (AREA)
  • User Interface Of Digital Computer (AREA)
  • Image Analysis (AREA)
  • Processing Or Creating Images (AREA)
  • Image Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

L'invention concerne un dispositif de traitement d'informations pour prendre en charge une conversation entre une pluralité d'utilisateurs, qui est pourvu : d'un moyen pour acquérir une première image faciale d'un premier utilisateur ; d'un moyen pour acquérir une première voix du premier utilisateur ; d'un moyen pour acquérir une seconde image faciale d'un second utilisateur ; d'un moyen pour acquérir une seconde voix du second utilisateur ; d'un moyen pour générer une première image faciale prédite du premier utilisateur à un premier âge désigné, diminué ou augmenté à partir de l'âge du premier utilisateur, sur la base de la première image faciale ; d'un moyen pour générer une seconde image faciale prédite du second utilisateur à un second âge désigné, diminué ou augmenté à partir de l'âge du second utilisateur, sur la base de la seconde image faciale ; d'un moyen pour présenter la première image faciale prédite au second utilisateur ; d'un moyen pour présenter la seconde image faciale prédite au premier utilisateur ; d'un moyen pour générer un message invitant le premier utilisateur et/ou le second utilisateur à parler ; d'un moyen pour présenter le message au premier utilisateur et/ou au second utilisateur ; d'un moyen pour présenter la seconde voix correspondant au message au premier utilisateur ; et d'un moyen pour présenter la première voix correspondant au message au second utilisateur.
PCT/JP2020/040558 2019-11-08 2020-10-29 Dispositif de traitement d'informations, procédé de traitement d'informations et programme WO2021090750A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2019-202887 2019-11-08
JP2019202887A JP2022176415A (ja) 2019-11-08 2019-11-08 情報処理装置及びプログラム

Publications (1)

Publication Number Publication Date
WO2021090750A1 true WO2021090750A1 (fr) 2021-05-14

Family

ID=75849686

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/040558 WO2021090750A1 (fr) 2019-11-08 2020-10-29 Dispositif de traitement d'informations, procédé de traitement d'informations et programme

Country Status (2)

Country Link
JP (1) JP2022176415A (fr)
WO (1) WO2021090750A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115410262A (zh) * 2022-10-09 2022-11-29 刘鹏 面部图像信息预测系统

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH07255044A (ja) * 1993-12-17 1995-10-03 Mitsubishi Electric Corp アニメーション式電子会議室及びビデオ会議システム及びビデオ会議方法
JPH09135427A (ja) * 1995-11-10 1997-05-20 Sharp Corp テレビカメラ装置
JP2009044299A (ja) * 2007-08-07 2009-02-26 Nippon Telegr & Teleph Corp <Ntt> 自動発呼システム、自動発呼制御方法、端末、端末制御方法、サービス制御装置、およびサービス制御方法
JP2010224802A (ja) * 2009-03-23 2010-10-07 Hironobu Otoi 仮想顔画像作成装置及び仮想顔画像作成方法
JP2016048569A (ja) * 2015-10-16 2016-04-07 洋彰 宮崎 言語入力により自律的に知識を拡大する人工知能装置
WO2018074516A1 (fr) * 2016-10-21 2018-04-26 株式会社Myth Système de traitement d'informations
JP2019067433A (ja) * 2012-12-28 2019-04-25 株式会社ユニバーサルエンターテインメント 話題提供システム
JP2019159337A (ja) * 2008-11-28 2019-09-19 株式会社半導体エネルギー研究所 表示装置

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH07255044A (ja) * 1993-12-17 1995-10-03 Mitsubishi Electric Corp アニメーション式電子会議室及びビデオ会議システム及びビデオ会議方法
JPH09135427A (ja) * 1995-11-10 1997-05-20 Sharp Corp テレビカメラ装置
JP2009044299A (ja) * 2007-08-07 2009-02-26 Nippon Telegr & Teleph Corp <Ntt> 自動発呼システム、自動発呼制御方法、端末、端末制御方法、サービス制御装置、およびサービス制御方法
JP2019159337A (ja) * 2008-11-28 2019-09-19 株式会社半導体エネルギー研究所 表示装置
JP2010224802A (ja) * 2009-03-23 2010-10-07 Hironobu Otoi 仮想顔画像作成装置及び仮想顔画像作成方法
JP2019067433A (ja) * 2012-12-28 2019-04-25 株式会社ユニバーサルエンターテインメント 話題提供システム
JP2016048569A (ja) * 2015-10-16 2016-04-07 洋彰 宮崎 言語入力により自律的に知識を拡大する人工知能装置
WO2018074516A1 (fr) * 2016-10-21 2018-04-26 株式会社Myth Système de traitement d'informations

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115410262A (zh) * 2022-10-09 2022-11-29 刘鹏 面部图像信息预测系统

Also Published As

Publication number Publication date
JP2022176415A (ja) 2022-11-29

Similar Documents

Publication Publication Date Title
JP6969653B2 (ja) 電子会議システム
JP6866860B2 (ja) 電子会議システム
US11061700B2 (en) Virtual assistants using state-based artificial intelligence
US10257241B2 (en) Multimodal stream processing-based cognitive collaboration system
CN102771082B (zh) 具有混合能力的设备和接口之间的通信会话
KR102428368B1 (ko) 선택가능한 그래픽 엘리먼트를 통해 자동화된 에이전트로 대화를 초기화하기
CN110730952A (zh) 处理网络上的音频通信的方法和系统
WO2019165877A1 (fr) Procédé, appareil, et dispositif de pousser de message, et support de stockage
WO2013091550A1 (fr) Procédé et dispositif permettant de poster un message de microblogue
CN103455592A (zh) 一种问答方法、装置及系统
KR102199434B1 (ko) 대화형 애플리케이션의 메시지 공유 시스템과 방법
US20240169989A1 (en) Multimodal responses
CN116595148B (zh) 一种利用大型语言模型实现对话流程的方法及系统
WO2021090750A1 (fr) Dispositif de traitement d&#39;informations, procédé de traitement d&#39;informations et programme
JP5729844B1 (ja) コンテンツの評価装置、システム、サーバ装置及び端末装置
CN114064943A (zh) 会议管理方法、装置、存储介质及电子设备
CN107783650A (zh) 一种基于虚拟机器人的人机交互方法及装置
WO2013123853A1 (fr) Procédé et dispositif de conversation homme-machine
US11164576B2 (en) Multimodal responses
WO2022119737A1 (fr) Assistant automatisé conçu pour faciliter des communications par l&#39;intermédiaire de fonctions de messagerie dissemblables de différentes applications
CN110989910A (zh) 交互方法、系统、装置、电子设备及存储介质
CN113767379A (zh) 使用内容代理和/或存储的内容参数来渲染内容
CN113867538A (zh) 交互方法、装置、计算机设备及计算机可读存储介质
Swati et al. Enhanced Accessibility of Facebook Messenger for Blind Users
CN116245974A (zh) 基于语音的绘制方法、装置、设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20883810

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20883810

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP