WO2023152976A1 - Communication system - Google Patents

Communication system Download PDF

Info

Publication number
WO2023152976A1
WO2023152976A1 PCT/JP2022/005691 JP2022005691W WO2023152976A1 WO 2023152976 A1 WO2023152976 A1 WO 2023152976A1 JP 2022005691 W JP2022005691 W JP 2022005691W WO 2023152976 A1 WO2023152976 A1 WO 2023152976A1
Authority
WO
WIPO (PCT)
Prior art keywords
virtual space
information
user
space construction
construction information
Prior art date
Application number
PCT/JP2022/005691
Other languages
French (fr)
Japanese (ja)
Inventor
千春 梶塚
美美 梶塚
琴未 梶塚
時央 梶塚
Original Assignee
千春 梶塚
美美 梶塚
琴未 梶塚
時央 梶塚
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 千春 梶塚, 美美 梶塚, 琴未 梶塚, 時央 梶塚 filed Critical 千春 梶塚
Priority to PCT/JP2022/005691 priority Critical patent/WO2023152976A1/en
Publication of WO2023152976A1 publication Critical patent/WO2023152976A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • the present disclosure relates to a communication system in which a user's communication method is executed in an experience-based service using a virtual space such as the Metaverse, in which at least one or more users participate, and a community service using a virtual space.
  • a virtual space such as the Metaverse
  • Metaverse A three-dimensional virtual space built on computers and networks that mimics the real world or is different from the real world.
  • users communicate with each other using avatars that represent users (see Patent Documents 1 and 2, for example).
  • Metaverse is highly synchronous, just like online games.
  • “high synchronism” means that when users interact with each other, they need to be online at the same time and in close proximity to each other in the same virtual space.
  • the system constructs and controls a virtual space based on multimodal information from all users, and distributes information about the constructed virtual space to users in the same virtual space. Based on the information, the user terminal displays the virtual space mainly for audiovisual purposes using computer graphics or the like.
  • Patent Document 1 Japanese Patent Application Laid-Open No. 2000-502266 makes it possible to record, play back, and edit past meetings in online meetings. The events that occur in continue without ending. It is desirable to make the means for storing/updating events and the means for replaying events independent processes so that replaying can be started in parallel while storing continues.
  • Patent Document 2 Japanese Patent Application Laid-Open No. 2002-123841
  • Patent Document 1 Japanese Patent Application Laid-Open No. 2002-123841
  • Patent Document 1 does not assume such a wide space.
  • the present disclosure saves multimodal (various information transmission means) inputs regarding the actions and conversations of users and avatars in past events in the Metaverse, and enables them to be played back later, thereby enabling real-world time It allows you to relive the past of the Metaverse as if you were riding in a machine and participating in an interesting event in the past.
  • users who join later can modify past events. Systematize a parallel world-like "what-if world” such as "what if the user was there” and “what if the user was acting differently” for events that occurred in the virtual space from the original world. It is possible to realize it as a world that spreads like a tree.
  • the communication system of the present disclosure is a communication system comprising one or more computers, comprising a first part, a second part, a third part, and a fourth and a portion of In a virtual space environment or an information space environment corresponding to a real space in which at least one or more users participate, the first part includes text, voice, still images, moving images, and avatar operation commands input from user terminals.
  • the first part includes text, voice, still images, moving images, and avatar operation commands input from user terminals.
  • multimodal information including any of the above, virtual space information, user information, and avatar information acquired based on the input information, and information from the third part, multiple virtual spaces are simultaneously constructed and controlled.
  • the third part reproduces and transmits to the first part information stored or updated by the second part
  • the fourth part reproduces information stored or updated by the second part. phylogenetically manages the history of the user's visits, and indicates to the third portion which information has been saved or updated to be reproduced according to the user's selection.
  • the problem caused by highly synchronous services in experience-based services using virtual space such as Metaverse, and community services using virtual space, is solved.
  • users can efficiently experience the Metaverse according to their limited time.
  • FIG. 1 is a conceptual diagram showing a configuration example of a communication system according to a first embodiment
  • FIG. FIG. 11 is a conceptual diagram showing a configuration example of a communication system according to a second embodiment
  • FIG. 10 is an explanatory diagram for explaining an example of a multimodal input procedure for the Metaverse using a smartphone
  • FIG. 10 is an explanatory diagram for explaining an example of a multimodal input procedure for the Metaverse using a smartphone
  • FIG. 10 is an explanatory diagram for explaining an example of a procedure for storing virtual space construction information on the basis of space
  • FIG. 10 is an explanatory diagram for explaining an example of a procedure for storing virtual space construction information based on users and avatars
  • FIG. 4 is an explanatory diagram for explaining a phylogenetic tree of virtual space construction information
  • FIG. 1 is a conceptual diagram showing a configuration example of a communication system 100 according to the first embodiment of the present disclosure.
  • a communication system 100 according to the first embodiment includes a plurality of user terminals 101, a first portion of a plurality of virtual space construction/distribution units 200, and a second portion of A virtual space construction information holding/updating unit 231, a virtual space construction information reproduction unit 232 as the third part, a virtual space construction information history management/selection unit 233 as the fourth part, and data stored in memory.
  • It comprises a multimodal document 234 which is a structure, a spatial information acquisition section 401 and a user information/avatar information acquisition section 501 .
  • Reference numerals "1100", “1101", “1102", and "1301" in FIG. 1 indicate the flow of information transmitted and received between units.
  • the multimodal document 234, the spatial information acquisition unit 401, and the user information/avatar information acquisition unit 501 are installed on the server 110, which is an information processing device, but the configuration is not limited to this.
  • these may be installed in separate servers, may be installed in the cloud or in the user terminal 101, or may be installed in a distributed manner using blockchain technology.
  • the server 110 By executing the communication program stored in the memory, the server 110 performs multiple virtual space construction/distribution unit 200, virtual space construction information holding/updating unit 231, virtual space construction information reproduction unit 232, virtual space construction information history. It functions as a management/selection unit 233 , a space information acquisition unit 401 and a user information/avatar information acquisition unit 501 .
  • the user terminal 101 is a device held by a user who participates in the Metaverse provided by the communication system 100 according to the first embodiment. Users participate in the Metaverse using various user terminals 101 such as personal computers, smartphones, and head-mounted displays. There are various types of Metaverses, but all of them use the three-dimensional positional information of the avatar in the virtual space or the positional information of the user in the real space. It determines the range of physical perception.
  • the user may send stamps (stickers) and photos during conversation, send attached documents, or use a slide show. Data that accompanies these conversations is also transmitted as multimodal information.
  • the multimodal information of the user terminal 101 is transmitted to the multiple virtual space construction/distribution unit 200, which is the first part.
  • This multiple virtual space construction/distribution unit 200 is installed on the server 110 in the first embodiment, but is not limited to this configuration, and may be installed on the cloud or installed on the user terminal 101.
  • each user terminal 101 may be connected by P2P and communicate with each other.
  • the multiple virtual space construction/distribution unit 200 puts together multimodal information transmitted by a user in a certain virtual space and multimodal information of other users in the virtual space, and the user terminal 101 Information necessary to construct a virtual space is processed in real time, and the processing result is delivered to the user terminal 101. - ⁇ At the same time, the multiple virtual space construction/distribution unit 200 performs the same processing for users in another virtual space, and distributes the processing result to the user terminal 101 .
  • the multiple virtual space construction/distribution unit 200 simultaneously performs parallel processing of the virtual spaces in which each user exists for all users, and sequentially distributes the processing results.
  • the multiple virtual space construction/distribution unit 200 also processes events occurring in each virtual space and distributes them to the user terminal 101 .
  • judgments necessary for the game such as judgment of contact between users and judgment of victory or defeat, are performed and distributed to each user terminal 101 .
  • the multiple virtual space construction/distribution unit 200 processes information about what happens in the virtual space and distributes it to each user terminal 101 .
  • the user terminal 101 renders a virtual space image based on the distributed information, and reproduces audio and video.
  • the space, objects in the space, and graphic data and sound data of the avatar may be distributed by the multiple virtual space construction/distribution unit 200, or may be stored on another system, service, cloud, or blockchain and provided to the user.
  • the information may be acquired by the terminal 101 as necessary, or may be included in advance in the program of the user terminal 101, or the above methods may be combined.
  • the virtual space construction information holding/updating unit 231 which is the second means, distributes to the user terminal 101 and the virtual space.
  • Information similar to the multimodal information that constitutes what is happening in the multimodal document 234 is stored along with time course information.
  • a "multimodal document 234" is a data structure stored in memory. These "information representing what happened in the virtual space at a specific time" to be saved or updated are called “virtual space construction information”.
  • the storage of the "virtual space construction information" by the virtual space construction information holding/updating unit 231 may be instructed by the user from the user terminal 101, or may be performed by the virtual space administrator or system administrator from the administrator terminal. You can instruct me to do it.
  • virtual space construction information for a specific virtual space.
  • virtual space construction information spanning multiple virtual spaces may be collectively saved in accordance with a user who moves between multiple virtual spaces.
  • the former is called “storing the virtual space construction information based on the space”
  • the latter is called “storing the virtual space construction information based on the user or avatar”.
  • the “virtual space building information” is written in the same format as the multimodal information from the user terminal 101 .
  • a virtual space construction information reproduction unit 232 which is a third means, reproduces specific "virtual space construction information" stored in the multimodal document 234 based on time information like video recording and reproduction, and reproduces a plurality of virtual spaces. It sequentially transmits to the construction/distribution unit 200 (see reference numeral 1102 in FIG. 1).
  • the multiple virtual space constructing/distributing unit 200 distributes the information to the user terminal 101, so that the user can re-experience what happened in a specific virtual space at a certain time in the past as if time-tripping into the past.
  • the multiple virtual space construction/distribution unit 200 or the virtual space construction information reproduction unit 232 changes the reproduction speed or skips the reproduction information so as to meet the user's request. You can fast-forward, rewind, pause, and skip the event occurrence time in the virtual space according to the time.
  • the processing of the virtual space construction information holding/updating unit 231 and the processing of the virtual space construction information reproducing unit 232 may be performed independently. As a result, the virtual space construction information reproducing unit 232 can start reproducing information at the same time as the processing by the virtual space construction information holding/updating unit 231 or after the completion of this processing.
  • the communication system 100 not only re-experiences what happened in the past virtual space as a single user, but also allows a plurality of users to form a group, register the group in the system, and participate as group members. It may be configured so that all members can play back the "virtual space construction information" at the same time and act together in the past virtual space.
  • the user terminal 101 transmits multimodal information of the user's behavior in the reproduced past virtual space to the multiple virtual space construction/distribution unit 200, and the new information from the user terminal 101 and the past " "Virtual space construction information" can be processed and distributed together.
  • the user not only sees and hears the virtual space in the past, but also adds new actions, adds new remarks, replaces past actions and remarks, and updates what happened in the past virtual space. can.
  • the virtual space construction information storage/update unit 231 may store in the multimodal document 234 what happened in the updated virtual space. At that time, the past "virtual space construction information" may be reproduced and updated from the start to the end, or only a part of the time may be reproduced and updated.
  • the first saved virtual space construction information is called “original virtual space construction information", and the updated virtual space construction information is called “updated virtual space construction information”.
  • the “updated virtual space construction information” may be further updated and saved. Updated “Original virtual space construction information” to “Updated virtual space construction information” Updated "Virtual space construction information” which further updated “Updated virtual space construction information” of the 1st generation and updated 1st generation 2
  • the “virtual space construction information” obtained by updating the “virtual space construction information" of the update (n)th generation is referred to as the update (n+1)th generation.
  • the virtual space construction information saving/updating unit 231 saves only difference information from the "original virtual space construction information" when saving the "updated virtual space construction information". may be saved to avoid tying up storage.
  • the multiple virtual space construction/distribution unit 200 treats "original virtual space construction information" and "updated virtual space construction information" as belonging to different independent worlds such as parallel worlds. Users in the original virtual space cannot see users or avatars who join later in the updated virtual space. On the other hand, users who join the updated virtual space later can see the users and avatars in the original virtual space.
  • the virtual space construction information reproducing unit 232 starts reproducing the original virtual space construction with a slight delay before the saving of the original virtual space construction is completed, and saves the original virtual space construction and updates the virtual space. Storage of spatial information may be performed in parallel at the same time. As a result, a state is created in which a parallel world with a slight time lag exists.
  • the multiple virtual space construction/distribution unit 200 processes virtual spaces belonging to different time axes and world lines as different virtual spaces in the same way as parallel processing of multiple virtual spaces at the same time.
  • the multiple virtual space construction/distribution unit 200 acquires space information. Information necessary to reproduce the space and avatar at that time is acquired via the unit 401 and the user information/avatar information acquisition unit 501 (see reference numeral 1301 in FIG. 1) and stored in the multimodal document 234 .
  • the multiple virtual space construction/distribution unit 200 archives them.
  • the author or distributor of each data may be allowed to set user authority as to whether or not to archive the data.
  • the user sends a request for playback time control from the user terminal 101 to the multiple virtual space building/distribution unit 200 or the virtual space building information playback unit 232, and time-controls what happens in the virtual space like fast-forwarding or rewinding playback of a moving image.
  • the virtual space construction information history management/selection unit 233 which is the fourth part, ID pointing to "virtual space building information", ID pointing to the original "virtual space building information” to be updated if the "virtual space building information" has been updated, information such as the time when it was saved or updated is saved in the multimodal document 234 as an update history and managed.
  • the history information managed by the virtual space construction information history management/selection unit 233 is made accessible 1100 from the user terminal 101 so that the user can select past "virtual space construction information". According to the user's selection, the virtual space construction information history management/selection unit 233 instructs the virtual space construction information reproduction unit 232 to reproduce the "virtual space construction information" (see reference numeral 1101 in FIG. 1). By doing so, the user can designate reproduction of the "virtual space construction information". The user or administrator who designates the storage of the “virtual space construction information” may be allowed to designate the user access authority for browsing and reproduction in the virtual space construction information history management/selection section 233 .
  • the virtual space construction information history management/selection unit 233 treats the "updated virtual space construction information" derived from the "original virtual space construction information" as a phylogenetic tree starting from the "original virtual space construction information”. structure, and displayed to the user via the user terminal 101 . By displaying like a phylogenetic tree in this way, the user can trace how what happened in the virtual space has been altered.
  • the phylogenetic tree and the difference information of each update may be simply displayed in text or the like so that the change process can be viewed easily.
  • the virtual space construction information history management/selection unit 233 with an AI equipped with appropriate judgment means, it is possible to detect tampering with malicious events.
  • the virtual space construction information reproduction unit 232 adds different audiovisual effects to the users and avatars who were in the original virtual space and the users and avatars added in the updated virtual space. can be identified.
  • the updated utterance may be identified by changing the text display color of the utterance or by adding a different sound effect to the voice. It is also possible to display a numerical value indicating how many generations the update corresponds to next to the balloon that displays the avatar or remarks.
  • a dedicated editor may be used to edit the "virtual space construction information" to create “updated virtual space information” without taking any action in the virtual space.
  • the "updated virtual space information” may be automatically processed by some algorithm, for example, updated such as removing a specific avatar from past events.
  • some algorithm for example, updated such as removing a specific avatar from past events.
  • the virtual space construction information history management/selection unit 233 strictly manages the history.
  • history data may be distributed and managed by blockchain so that even system administrators cannot easily modify it.
  • FIG. 2 is a conceptual diagram showing a configuration example of a communication system 100A according to the second embodiment of the present disclosure.
  • the communication system 100A according to the second embodiment includes a plurality of user terminals 101, a plurality of data format conversion units 301 provided corresponding to each user terminal 101, and a first part.
  • Multiple virtual space construction/distribution unit 200 includes an input interface 201, an output interface 202, a space sorting unit 203, and multiple virtual space constructing units 221-223. Reference numerals "1100”, “1101”, “1102”, “1200”, “1201", “1301”, etc. in FIG. 1 indicate the flow of information transmitted and received between the units.
  • a multiple virtual space construction/distribution unit 200 a virtual space construction information storage/update unit 231, a virtual space construction information reproduction unit 232, a virtual space construction information history management/selection unit 233, and a multimodal
  • the document 234, the spatial information acquisition unit 401, the user information/avatar information acquisition unit 501, the virtual space construction information learning unit 261, and the AI avatar/AI event generation unit 271 are installed on the server 110. , but not limited to this configuration. For example, these may be installed in separate servers, may be installed in the cloud or in the user terminal 101, or may be installed in a distributed manner using blockchain technology.
  • a user operates an avatar in a virtual space with the user terminal 101 to have a conversation, or moves in real space holding the user terminal 101, tilts the user terminal 101, and specifies in the direction of
  • a user terminal 101 includes a camera, a microphone, a GPS, and the like.
  • User terminal 101 transmits multimodal information to input interface 201 of multiple virtual space construction/distribution unit 200 . Also, the processing result of the multiple virtual space information constructing/distributing unit 200 is distributed to the user terminal 101 through the output interface 202 .
  • the data format conversion unit 301 converts data in a format that can be input by the user or a format that is highly convenient for the user into a format that is easy for the system to handle. Convert.
  • the data format conversion unit 301 converts multimodal information into a format that facilitates temporal manipulation by the virtual space construction information reproduction unit 232, a format that facilitates updating later, and a format that reduces the amount of data and does not impose a load on the system. do.
  • the information delivered from the output interface 202 is also converted by the data format conversion unit 301 into a format that is easy for the user to utilize.
  • the data format conversion unit 301 converts the text so that it can be conveyed to other users by voice, and the other user's voice is converted to text. Convert.
  • the data format conversion unit 301 or the virtual space construction information reproduction unit 232 generates a video of the conversation according to the conversation based on the face photo prepared in advance by AI or the like. and may be transmitted to the input interface 201 .
  • the data format conversion unit 301 may translate.
  • the data format conversion unit 301 converts the voice conversation of another user into text and displays it, so that the text-inputted conversation can be conveyed to the other user by synthesized speech.
  • the data format conversion unit 301 converts text chat of other users into voice.
  • Such a data format conversion unit 301 may be provided in each user terminal 101 and data conversion processing may be performed on the client side, or may be provided on the server 110 or cloud side and data conversion processing may be performed on these sides. If the user terminal 101 has sufficient processing power, placing the data format conversion unit 301 on the client side will balance the load of the entire system. If an excellent system such as translation conversion already exists, the data format conversion unit 301 may use the function of such other system or service to perform the conversion. Details of the data format conversion will be described later with reference to FIGS.
  • the input interface 201 and the output interface 202 are installed on the server 110, they may be installed on the cloud, or may be installed on each user terminal 101 so that the user terminals 101 communicate with each other. Also good.
  • the input interface 201 handles the multimodal information from the user terminal 101, the multimodal information from the virtual space construction information reproduction unit 232, and the multimodal information from the AI avatar/AI event generation unit 271, which will be described later.
  • Multimodal information received by the input interface 201 is distributed to a plurality of virtual space construction units 221 to 223 by a space distribution unit 203 .
  • the spatial distribution unit 203 is a mechanism for distributing loads by parallel processing similar to the load balancer of the server 110, but the distribution logic is different from that of the load balancer.
  • the space distribution unit 203 distributes users in the same virtual space to the same virtual space construction units 221-223.
  • the space distribution unit 203 distributes users in the same virtual space to different virtual space construction units 221 to 223. If the user has designated to act in a group, the space distribution unit 203 preferentially distributes to the same virtual space construction units 221 to 223 .
  • the space sorting section 203 sorts the users acting in groups to the same virtual space building sections 221-223.
  • the space distribution unit 203 distributes the received users to the same virtual space construction units 221 to 223 . Even if they are the same virtual space, it is not necessary to distribute the original virtual space, the updated virtual space, and the virtual spaces of different update generations to the same virtual space constructing units 221 to 223 .
  • the system needs to handle many virtual spaces at the same time, and the load is too high for a single server or process to handle. Therefore, the virtual space constructing units 221 to 223 process a plurality of virtual spaces in parallel.
  • the virtual space constructing units 221 to 223 can be increased or decreased according to the load of the entire system and the number of active users in the virtual space.
  • a thread for executing the virtual space construction units 221 to 223 is automatically created when a user or avatar enters a specific empty virtual space, and when there is no user or avatar left in the virtual space. Allow threads to terminate automatically.
  • three virtual space construction units 221, 222, and 223 are displayed, and two rectangles marked with "" are displayed between them. represents what is possible.
  • the virtual space construction units 221 to 223 are installed on a specific server 110, but they may be installed virtually on the cloud or distributed on each user terminal 101 connected by P2P connection.
  • the space sorting unit 203 is designed to be able to grasp when the virtual space constructing units 221 to 223 are newly opened or closed.
  • the virtual space construction information storage/update unit 231 stores or updates what happened in the virtual space in the multimodal document 234 .
  • the virtual space construction information storage/update unit 231 may perform parallel processing in the same way as the virtual space construction units 221 to 223 to distribute the load. Since "virtual space information" that spans multiple virtual spaces may be saved, it is installed independently of the virtual space constructing units 221-223.
  • the virtual space construction information storage/update unit 231 acquires graphic data required for the virtual space through the space information acquisition unit 401 and the user information/avatar information acquisition unit 501, and stores them in the multimodal document 234 (1301). .
  • the virtual space construction information storage/update unit 231 acquires graphic data via the space information acquisition unit 401 and the user information/avatar information acquisition unit 501 so that the user terminal 101 can reproduce the virtual space at that time. Good (see reference numeral 1303 shown in FIG. 2).
  • the "virtual space construction information” is stored in the multimodal document 234, and this information is also information about what actions the user or avatar took in a specific situation.
  • the multimodal document 234 stores a large amount of "virtual space construction information”, and the virtual space construction information learning unit 261 uses machine learning to learn data representing necessary and sufficient behavior of the user or avatar under a specific situation. user or avatar behavior can be predicted and behavior can be generated. Similarly, the virtual space construction information learning unit 261 can also predict and generate what kind of event will occur in the virtual space if users or avatars with what attributes gather.
  • the communication system 100A can construct a Metaverse that is crowded with many avatars and various events occur even though no user actually participates.
  • the communication system 100A can construct a village-like village that is filled with NPC (non-player character) villagers in an adventure game and that conveys information to the user.
  • NPC non-player character
  • the learning result of the virtual space construction information learning unit 261 is transmitted to the AI avatar/AI event generation unit 271, and the AI avatar/AI event generation unit 271 generates an avatar acting by AI and an event controlled by AI.
  • These avatars and events are described in the same format as multimodal information and "virtual space construction information" from the user terminal 101, and are transmitted from the AI avatar/AI event generation unit 271 to the input interface 201 (Fig. 2 1201) can be handled by the multiple virtual space construction/distribution unit 200 in the same way as the user's avatar.
  • the AI avatar/AI event generation unit 271 changes the response according to the user's multimodal information, thereby generating an avatar and a situation that appropriately responds to the user's behavior. For example, generated avatars and situations can be used to build dynamic learning environments in conversational style.
  • a user can normally operate only one avatar at a time, but if an AI avatar with a personality that comprehends the user's behavior becomes possible, it will be possible for the user to act in multiple virtual spaces at the same time. Possessing an avatar, you can experience what happened in each virtual space later in fast-forward, and you will be able to use your limited time effectively. Also, even if the user is offline, the avatar of the user's alter ego will perform the task or mission of the user.
  • Such possibilities also provide a solution to the problem of the Metaverse, in which the time in the real world and the time in the Metaverse progress in a one-to-one relationship, and the problem of the user's time being consumed more and more.
  • the user can operate the setting of the virtual space construction information learning unit 261, it becomes possible to customize AI avatars and AI events.
  • Users can buy and sell customized AI avatars and AI events.
  • NFT Non-fungible token
  • non-fungible token non-fungible token
  • FIGS. 3 and 4 are explanatory diagrams for explaining one example and another example of a multimodal input procedure for the Metaverse using a smartphone.
  • User operations at the user terminal 101 and data conversion at the data format conversion unit 301 will be described below with reference to FIGS. 3 and 4.
  • FIG. 3
  • the user terminal 101 is not limited to a smart phone.
  • the user terminal 101 is provided with a camera 102 and a microphone 103, and incorporates a GPS function for acquiring the user position in real space and various sensors (not shown) for detecting the tilt and movement of the smartphone.
  • the smartphone screen is divided into upper and lower parts, the virtual space 1001 by computer graphics is displayed in the upper part, and the screen corresponding to each input is displayed in the lower part, but this is just an example. It does not limit interface design.
  • the user terminal 101 may express the position of the user, the surrounding environment, the positions of other users, and the like with stereoscopically localized sounds without visually displaying the virtual space.
  • an avatar 1002 of the user and avatars 1003 of other users in the same space are displayed.
  • FIG. 3(a) is an example of conversation input by text input.
  • a text input screen 1010 is displayed at the bottom of the screen of the user terminal 101 , and the user inputs characters using a keyboard (touch panel) and transmits a series of characters input using a send button 1011 .
  • the text "Hello” is sent, followed by a sticker with an illustration image.
  • the user makes a series of 'tap inputs' 3012 on the screen along the 'time lapse' 3000 .
  • “H”, “e”, “l”, “l”, “o”, the send button, and the stamp are tapped in order. In the explanation, the operation of the stamp is simplified.
  • the data to be transmitted is processed as two data groups of “text “Hello”” 3013 and “stamp ID” 3014 .
  • Communication text data 3001 is encrypted and transmitted together with “text “Hello”” 3013 and “stamp ID” 3014.
  • common text data a user ID, an avatar ID, the local time of the user terminal 101, the user's position in real space, and the like are transmitted.
  • FIG. 3 is an example of voice conversation input.
  • a “send voice button” 1020 is displayed below the “voice chat screen” 1021 .
  • the user speaks while pressing the “send voice button” 1020 for a long time.
  • "Voice” 3023 of "from start to end of long press” 3022 of "send voice button” 1020 is processed as a series of speech data.
  • This “speech” is subjected to “speech text conversion processing” 3024 by the data format conversion unit 301 .
  • a series of “conversation voice data” 3025 is composed of text converted into voice-to-text, ID indicating voice data, original “voice data” 3026 , and “common text data” 3001 .
  • the virtual space construction information storage/update unit 231 can text-edit the past conversation in the "updated virtual space construction information".
  • the data format conversion unit 301 can also convert the edited text into text-to-speech and replace the past conversation speech. If AI-synthesized voice mimics the user's voice, the data format conversion unit 301 can replace the past conversation in the virtual space as if the user had spoken at that time.
  • FIG. 3 is an example of a conversation input using a moving image. This example is similar to the voice chat example. The user converses while pressing and holding the 'video transmission button' 1030 on the 'video chat screen' 1031 displayed on the screen of the user terminal 101 . In the case of video chat, since hands-free operation is often desired, the user terminal 101 may identify the start and end of speech by volume.
  • "Audio” 3033 is subjected to "speech text conversion” 3024 by the data format conversion unit 301
  • "moving image” 3034 is subjected to "moving image compression” 3035
  • text converted as a series of data 3036, "sound data” 3023, "moving image data 3037 and “common text data” 3001 are transmitted from the user terminal 101 to the server 110 .
  • the data format conversion unit 301 uses a general moving image compression algorithm that reduces the data transfer load, and also converts moving images into frame images of faces corresponding to vowels of speech and characteristic frame images with movement. and send only those keyframe images.
  • the user terminal 101 of another user interpolates the keyframe image into a moving image to generate a moving image of the user speaking.
  • the data format conversion unit 301 replaces and interpolates the key frame images to create a video, thereby converting the speech in the past virtual space at that point in time. It is replaced as if the user had spoken.
  • FIG. 4(a) relates to movement of the avatar in space.
  • an "avatar movement operation screen” 1040 is displayed obliquely from above (quarter view) to clearly display the position of the avatar in the virtual space.
  • a display viewed from directly above (bird's eye view) like a map application may be used, or an avatar movement operation may be performed using a dedicated input terminal for a head-mounted display.
  • "avatar movement operation screen” 1040 "own avatar” 1043 and “another user's avatar” 1042 are drawn.
  • a “virtual space where the user is present” 1041 and an “adjacent virtual space” 1042 are displayed.
  • Patent Document 2 Japanese Unexamined Patent Application Publication No. 2002-123841
  • Patent Document 2 Japanese Unexamined Patent Application Publication No. 2002-123841
  • the user taps the "avatar movement operation screen” 1040 to move the avatar.
  • an animation of your avatar walking to that location is played.
  • the user continuously taps a plurality of points a, b, c, and d on the "avatar movement operation screen” 1040 (3044) to move the avatar.
  • b tap twice to stop.
  • the user terminal 101 reduces the communication load and the processing load of the server 110 by transmitting only the movement position to the server 110, but may transmit detailed movement data like a game.
  • the avatar's position information follows the general method of indicating the position of an object in computer graphics.
  • the angle (x, y, z) and the scaling factor (x, y, z) may be represented by a set of numeric text data.
  • it sends how to interpolate the movement animation between each position and the next position.
  • Common specification methods for keyframe animation include "linear”, which connects specified positions in a straight line, and "spline", which connects specified positions with a smooth curve.
  • the location information is transmitted together with the “common text data” 3001 .
  • "P1" 3045 of the transmission data represents movement within the same virtual space
  • "P2” 3046 represents movement accompanied by movement to a different virtual space.
  • information indicating the destination virtual space such as an ID uniquely held by the Metaverse system and the URL of the virtual space data, is sent from the user terminal 101 to the server 110 .
  • FIG. 4 relates to avatar operations other than movement.
  • motion and facial expressions are explained as an example.
  • An avatar “motion designation screen” 1050 is displayed at the bottom of the screen of the user terminal 101 .
  • Several “thumbnails of poses” 1051 are displayed at the top of the screen, and “selected poses” 1050 are displayed at the bottom of the screen.
  • the user taps the "thumbnail of pose” 1051 according to the "passage of time” 3000, connects poses, and designates a motion (3052).
  • the user terminal 101 displays options for facial expressions on the screen, and the user taps them to designate facial expressions (3053).
  • This is a simple method of specifying motion and facial expressions, but there is also a method in which the user attaches a sensor device to the body and records the movement (motion capture) and transmits it in real time.
  • the user terminal 101 transmits "common text data" 3001 together with motion and facial expression data designated by various methods.
  • a motion or facial expression may be linked to the conversation stamp.
  • (c) of FIG. 4 is a case of transmitting continuous sound and moving images.
  • the user continues talking into the camera 102 and the microphone 103 of the user terminal 101 without operating the screen, and confirms his own "camera image” 1060 .
  • the user terminal 101 divides the continuous “speech” 3061 and “moving image” 3062 accompanying the “passage of time” 3000 at regular time intervals so that the server 110 can easily handle them, and divides them into “speech-to-text conversion” 3024 and “moving image compression” 3024 .
  • After performing data format conversion such as 3035 it is transmitted to the server 110 as a continuous data group together with the converted text and "common text data" 3001 (3061).
  • the virtual space construction information storage/update unit 231 stores the multimodal information from the “user terminal 101” as a series of “virtual space construction information”. and how the virtual space construction control information history management/selection unit 233 manages them.
  • FIG. 5 is an explanatory diagram for explaining an example of a procedure for "saving virtual space construction information based on space”.
  • FIG. 6 is an explanatory diagram for explaining an example of a procedure for "saving virtual space construction information based on a user or avatar”.
  • “Start Saving” 2001 is the point at which the virtual space construction information storage/update unit 231 starts saving the "virtual space construction information”
  • “End Saving” 2004 is the point at which the saving ends. treated as a set of "virtual space construction information”.
  • 5 to 7 the symbols "2000", “2010” to "2019” shown on the left side of the page indicate a series of data groups sent to the virtual space construction information storage/update unit 231 over time.
  • “recorded in space A” 2000 is a data group recorded in space A and transmitted to the virtual space construction information storage/update unit 231 .
  • Codes “2100” to “2103” indicate “updated virtual space construction information” which is the result of putting these data groups together by the virtual space construction information storage/update unit 231 .
  • the symbols shown on the right side of the paper indicate the time (point of time) at which an event occurs or the process to be executed.
  • the “original virtual space construction information” shown in FIG. 5 will be described.
  • some event starts at “Start Event” 2002 and ends at “End Event” 2003 .
  • a guitarist's avatar appears in the virtual space at the "Start Event” and performs, and then exits the space at the "End Event”.
  • users B and C converse in the space and wait for the event, and user A appears in the space just before the start of the event. User A leaves the virtual space shortly after the event ends.
  • the “updated first generation+user D” shown in FIG. 5 will be described.
  • User D updates the “original virtual space construction information” and saves the update result as “updated virtual space construction information” 2101 .
  • User D starts the reproduction from the beginning of the "original virtual space construction information” 2007, conducts several voice conversations, selects an avatar action such as clapping for the guitarist, and then ends the reproduction 2008.
  • the organizer of the event has specified to automatically update the 'updated virtual space construction information', and the 'updated virtual space construction information' 2101 has been updated. This is the first updated generation.
  • the “update first generation + user E, user B (revisit)” shown in FIG. 5 will be described.
  • User E forms a group with user B, reproduces the "virtual space construction information", and acts in it.
  • User B's avatar also participates in the "original virtual space construction information", but participates in the event again with user E.
  • User B joins late, talks with user E, and leaves early. Due to the update by user B, part of the "input of user B" in the "original virtual space construction information” is rewritten.
  • the updated utterance overwrites the previously spoken utterance, or is inserted as a new utterance.
  • the standing position and clothes of user B's avatar may be different from those of the "original virtual space construction information”. In such a case, the original user B's avatar may be replaced only while user B is participating. Alternatively, like the movie "Back to the Future", the original avatar of user B and the updated avatar of user B1 may be displayed simultaneously.
  • FIG. 6 is an explanatory diagram for explaining an example of a procedure for "saving virtual space construction information based on users and avatars".
  • the “original virtual space construction information” shown in FIG. 6 will be described.
  • User A first visits space B, then moves to space A to participate in an event, and moves to space C.
  • Different "virtual space construction information” are summarized as "stored virtual space construction information" 2100 as a result of user A's actions processed in space A, space B, and space C.
  • User B and user C are in space A, and user A can talk with their avatars while staying in space A.
  • User G visits space A halfway through, meets user A and has a conversation with him, then moves to space C with user A, and continues the conversation with user A.
  • the “updated first generation+user H, user I” shown in FIG. 6 will be described.
  • user H and user I reproduce the "original virtual space building information" as a group, and move in space B, space A, and space C as if accompanied by user A.
  • User H moves in the space without leaving user A, but user I moves in the space with a slight delay. Therefore, in updating the virtual space information stored with user A as a reference, user H is temporarily out of the space where user A is, and disappears from the update information.
  • the participating users are temporarily divided into a plurality of spaces in this way, the information about the plurality of spaces may be included in the "updated virtual space construction information" at the same time.
  • FIG. 7 is an explanatory diagram for explaining the phylogenetic tree of the virtual space construction information. Based on the example of FIG. 5, this FIG. 3103 is updated, and a phylogenetic branch is formed.
  • This disclosure can be used for experience-based services and community services using virtual spaces such as the Metaverse.
  • This disclosure can also be used to create interactive educational content that makes use of conversations within the Metaverse.
  • the content provided by the communication system of the present disclosure is different from mere interactive educational content. Active participation as if you traveled back in time to a past event, such as walking around in a group and listening to an explanation of how it works. becomes possible.
  • FIG. 6 by storing the space construction information from the viewpoint of the user who moves across multiple virtual spaces, for example, it is possible not only to explain in a nuclear reactor in a single space, but also to You can experience explanations while moving around the entire space of the power plant, or while moving from the power plant space to the virtual space of a wide area such as the urban space where electricity is used.
  • AR Augmented Reality
  • users can see historical sites where historical battles were fought, for example. While actually visiting various places in the historic site, you will be able to actively participate by listening to explanations from avatars and asking questions.
  • the facilitator checks the added avatar questions, further updates the virtual space, and adds answers to the questions.
  • the communication system of the present disclosure accumulates many dialogues, cuts and compiles multiple pieces of "virtual space construction information", eliminates redundant parts, and creates new content such as a collection of dialogues to experience. can.
  • the communication system of the present disclosure can also be used to make a monologue in the virtual space and update it later to add a comment.
  • the communication system of the present disclosure may be equipped with artificial intelligence, and with the development of artificial intelligence technology, it becomes possible for the avatar to behave more highly human-like, allowing the user to accurately respond to requests, If a part of the user's knowledge and way of thinking is inherited and it becomes an existence that acts as an alter ego that cuts out a certain characteristic of the user, it will be possible to realize a new service for sharing knowledge. Since such an AI avatar resides in the reproduced virtual space, it is possible to respond to new participants as if the original user were there. Furthermore, if users can nurture avatars with their own specialized knowledge and ways of thinking, the usage rights can be attached to virtual space dialogue events and sold or lent. The “knowledge market” that buys and sells such expertise creates value.
  • the communication system of this disclosure does not keep avatars in the virtual world, but when combined with robots and self-driving cars, it will be possible to participate in dialogues in each place while acting in the real world.
  • the present disclosure can be used as a mechanism for creating new uses of the Metaverse in industrial fields such as education and entertainment.

Abstract

Provided is a communication system that enables a vicarious experience of a past in a metaverse as if riding in a time machine in a real world to participate in a scene of an interesting event in the past. A communication system (100) includes: a multiple virtual spaces construction/distribution unit (200) as a first section; a virtual space construction information storing/updating unit (231) as a second section; a virtual space construction information reproduction unit (232) as a third section; and a virtual space construction information history managing/selecting unit (233) as a fourth section. The multiple virtual spaces construction/distribution unit (200) simultaneously constructs and controls a plurality of virtual spaces on the basis of multimodal information. The virtual space construction information storing/updating unit (231) stores or updates virtual space construction information. The virtual space construction information reproduction unit (232) reproduces virtual space construction information. The virtual space construction information history managing/selecting unit (233) stores and manages an update history of the virtual space construction information.

Description

コミュニケーションシステムcommunication system
 本開示は、少なくとも一人以上のユーザが参加する、メタバースをはじめとする仮想空間を使った体験型のサービス、仮想空間を使ったコミュニティ・サービスにおける、ユーザのコミュニケーション方法が実行されるコミュニケーションシステムに関する。 The present disclosure relates to a communication system in which a user's communication method is executed in an experience-based service using a virtual space such as the Metaverse, in which at least one or more users participate, and a community service using a virtual space.
 コンピュータやネットワーク上に構築され、現実社会を模した、又は現実社会とは異なる三次元の仮想空間は、メタバースと呼ばれる。メタバースでは、ユーザの分身であるアバタを用いて、ユーザ同士がコミュニケーションすることが行われる(例えば、特許文献1、2参照)。 A three-dimensional virtual space built on computers and networks that mimics the real world or is different from the real world is called the Metaverse. In the Metaverse, users communicate with each other using avatars that represent users (see Patent Documents 1 and 2, for example).
 メタバースはオンラインゲームと同様に同期性が高い。ここで「同期性が高い」とは、ユーザが交流する際に同じ時間、同じ仮想空間内のお互いに近い場所にオンライン状態にいる必要があるということを指す。  The Metaverse is highly synchronous, just like online games. Here, "high synchronism" means that when users interact with each other, they need to be online at the same time and in close proximity to each other in the same virtual space.
 ユーザは、さまざまなユーザ端末からマルチモーダル(様々な情報伝達手段)な情報を送信してメタバースに参加する。システムは全ユーザからのマルチモーダルな情報をもとに仮想空間を構築制御し、同一の仮想空間にいるユーザに対して、構築された仮想空間の情報を配信する。ユーザ端末はその情報に基づき、コンピュータグラフックスなどを用いて仮想空間を主に視聴覚に対して表示する。 Users participate in the Metaverse by transmitting multimodal information (various information transmission means) from various user terminals. The system constructs and controls a virtual space based on multimodal information from all users, and distributes information about the constructed virtual space to users in the same virtual space. Based on the information, the user terminal displays the virtual space mainly for audiovisual purposes using computer graphics or the like.
特開2000-50226号公報JP-A-2000-50226 特開2002-123841号公報JP-A-2002-123841
 一方で同期性が重要でない「ツイッター」や「フェイスブック」などのソーシャル・ネットワーク・サービスでは多くのユーザが書き込んだ人気の話題は蓄積され、後から参加したユーザも閲覧しコメントすることができる。しかしメタバースでは、多くのユーザが集まった人気のイベントでも後からその空間を訪れると誰もいない寂しい状態になっている。こうしたメタバースにおけるイベント終了後のユーザ減少の可視化は、これまでの様々なメタバース・サービスが衰退する一因ともなっている。 On the other hand, in social network services such as "Twitter" and "Facebook" where synchronism is not important, popular topics written by many users are accumulated, and users who join later can view and comment. However, in the Metaverse, even if a popular event attracts a large number of users, when you visit the space later, it will be empty and lonely. The visualization of the decrease in users after events in the Metaverse is one of the reasons why various Metaverse services have declined.
 さらに、SNSでは気にいった書き込みは何度でも読み返せるが、メタバースでは過去のイベントを収録された動画で見るくらいでしか追体験できない。さまざまな現実世界の制約をとりのぞくことを特徴とするメタバースにも関わらず、現実の世界でタイムマシンによって過去に行けないのと同様に、メタバースでも過去に戻ることができない。 Furthermore, on SNS, you can reread what you like as many times as you like, but on the Metaverse, you can only relive past events by watching the recorded videos. Despite the fact that the Metaverse is characterized by the removal of various real-world constraints, you cannot go back to the past in the Metaverse any more than you can go to the past by a time machine in the real world.
 特許文献1(特開2000-50226号公報)に記載の発明は、オンライン会議において過去の会議の録画と再生、編集を実現できるようにするものだが、開始と終了が明確な会議と異なり、メタバースで起きる事象は終了することなく継続する。事象の保存/更新手段と再生手段を独立したプロセスとし、保存が継続して行われている間にも並行して再生が開始できるようにするのが望ましい。 The invention described in Patent Document 1 (Japanese Patent Application Laid-Open No. 2000-50226) makes it possible to record, play back, and edit past meetings in online meetings. The events that occur in continue without ending. It is desirable to make the means for storing/updating events and the means for replaying events independent processes so that replaying can be started in parallel while storing continues.
 また、特許文献2(特開2002-123841号公報)に記載の発明は、会議室のような限定された広さの仮想空間ではなく多数のユーザが同時に参加する広大な仮想空間を想定しているが、特許文献1は、こうした広い空間を想定したものではない。 In addition, the invention described in Patent Document 2 (Japanese Patent Application Laid-Open No. 2002-123841) assumes a vast virtual space in which many users participate simultaneously rather than a virtual space of limited size such as a conference room. However, Patent Document 1 does not assume such a wide space.
 また、昨今では、フェイクニュースに代表されるように過去の事象の改竄が問題となっている。このため、最初に起きた事象と更新された情報の履歴を管理し、ユーザが容易にその更新の過程を遡れるようにする必要がある。 In recent years, the falsification of past events, as typified by fake news, has become a problem. For this reason, it is necessary to manage the history of events that occurred first and updated information so that the user can easily trace the update process.
 そこで、本開示は、メタバースにおける過去のイベントでのユーザやアバタの行動や会話に関するマルチモーダル(様々な情報伝達手段)な入力を保存し、後から再生できるようにすることで、現実世界でタイムマシンに乗って過去の興味ある事件の場に参加するかのようにメタバースの過去を追体験できるようにするものである。加えて、後から参加したユーザが過去のイベントを改変することができるようにする。
 仮想空間で起きた出来事に対して「もしユーザがその場にいたら」「もしユーザがその場で違う行動をとっていたら」というようなパラレルワールド的な「もしもの世界」をオリジナルの世界から系統樹のように拡がっていく世界として実現することを可能とするものである。
Therefore, the present disclosure saves multimodal (various information transmission means) inputs regarding the actions and conversations of users and avatars in past events in the Metaverse, and enables them to be played back later, thereby enabling real-world time It allows you to relive the past of the Metaverse as if you were riding in a machine and participating in an interesting event in the past. In addition, users who join later can modify past events.
Systematize a parallel world-like "what-if world" such as "what if the user was there" and "what if the user was acting differently" for events that occurred in the virtual space from the original world. It is possible to realize it as a world that spreads like a tree.
 前記目的を達成するために、本開示のコミュニケーションシステムは、1又は複数のコンピュータから構成されるコミュニケーションシステムであって、第一の部分と、第二の部分と、第三の部分と、第四の部分とを備える。少なくとも1人以上のユーザが参加する仮想空間環境または実空間に対応する情報空間環境において、前記第一の部分は、ユーザ端末から入力されるテキスト、音声、静止画、動画、アバタ操作のコマンドの何れかを含むマルチモーダルな情報と、その入力情報をもとに取得した仮想空間情報、ユーザ情報ならびにアバタ情報とに加えて、前記第三の部分からの情報により複数の仮想空間を同時に構築制御し、その制御結果を前記ユーザ端末に配信し、前記第二の部分は、前記第一の部分が仮想空間を構築制御する際に使用した情報を時間経過と関連付けてメモリに保存または更新し、前記第三の部分は、前記第二の部分により保存または更新される情報を再生して前記第一の部分に送信し、前記第四の部分は、前記第二の部分が情報を保存または更新した履歴を系統樹的に管理し、ユーザの選択に応じて再生すべき保存または更新された情報を前記第三の部分に指示する。 To achieve the above object, the communication system of the present disclosure is a communication system comprising one or more computers, comprising a first part, a second part, a third part, and a fourth and a portion of In a virtual space environment or an information space environment corresponding to a real space in which at least one or more users participate, the first part includes text, voice, still images, moving images, and avatar operation commands input from user terminals. In addition to multimodal information including any of the above, virtual space information, user information, and avatar information acquired based on the input information, and information from the third part, multiple virtual spaces are simultaneously constructed and controlled. and distributes the control result to the user terminal, and the second part saves or updates the information used when the first part constructs and controls the virtual space in a memory in association with the passage of time, The third part reproduces and transmits to the first part information stored or updated by the second part, and the fourth part reproduces information stored or updated by the second part. phylogenetically manages the history of the user's visits, and indicates to the third portion which information has been saved or updated to be reproduced according to the user's selection.
 本開示によれば、メタバースをはじめとする、仮想空間を使った体験型のサービス、仮想空間を使ったコミュニティ・サービスにおいて同期性の高いサービスであったことに起因する課題を解決し、ソーシャル・ネットワーク・サービスのような非同期・蓄積型の利点を付加し、ユーザがメタバースの体験をユーザの有限な時間にあわせて効率的に行なえる。 According to the present disclosure, the problem caused by highly synchronous services in experience-based services using virtual space, such as Metaverse, and community services using virtual space, is solved. By adding asynchronous and storage type advantages such as network services, users can efficiently experience the Metaverse according to their limited time.
第1の実施の形態に係るコミュニケーションシステムの構成例を示す概念図である。1 is a conceptual diagram showing a configuration example of a communication system according to a first embodiment; FIG. 第2の実施の形態に係るコミュニケーションシステムの構成例を示す概念図である。FIG. 11 is a conceptual diagram showing a configuration example of a communication system according to a second embodiment; FIG. スマートフォンを使ったメタバースのマルチモーダルな入力手順の一例を説明するための説明図である。FIG. 10 is an explanatory diagram for explaining an example of a multimodal input procedure for the Metaverse using a smartphone; スマートフォンを使ったメタバースのマルチモーダルな入力手順の一例を説明するための説明図である。FIG. 10 is an explanatory diagram for explaining an example of a multimodal input procedure for the Metaverse using a smartphone; 空間を基準に仮想空間構築情報を保存する手順の一例を説明するための説明図である。FIG. 10 is an explanatory diagram for explaining an example of a procedure for storing virtual space construction information on the basis of space; ユーザやアバタを基準に仮想空間構築情報を保存する手順の一例を説明するための説明図である。FIG. 10 is an explanatory diagram for explaining an example of a procedure for storing virtual space construction information based on users and avatars; 仮想空間構築情報の系統樹を説明するための説明図である。FIG. 4 is an explanatory diagram for explaining a phylogenetic tree of virtual space construction information;
 以下、本開示によるメタバースにおけるコミュニケーションシステムの実施形態を、図面を参照しながら説明する。 An embodiment of a communication system in the Metaverse according to the present disclosure will be described below with reference to the drawings.
 (第1の実施の形態)
 図1は、本開示の第1の実施の形態に係るコミュニケーションシステム100の構成例を示す概念図である。この図1に示すように、第1の実施の形態に係るコミュニケーションシステム100は、複数のユーザ端末101と、第一の部分である複数仮想空間構築/配信部200と、第二の部分である仮想空間構築情報保持/更新部231と、第三の部分である仮想空間構築情報再生部232と、第四の部分である仮想空間構築情報履歴管理/選択部233と、メモリに記憶されたデータ構造であるマルチモーダルドキュメント234と、空間情報取得部401と、ユーザ情報/アバタ情報取得部501と、を備える。なお、図1中の符号「1100」、「1101」「1102」、「1301」は、各部間で送受信される情報の流れを示している。
(First embodiment)
FIG. 1 is a conceptual diagram showing a configuration example of a communication system 100 according to the first embodiment of the present disclosure. As shown in FIG. 1, a communication system 100 according to the first embodiment includes a plurality of user terminals 101, a first portion of a plurality of virtual space construction/distribution units 200, and a second portion of A virtual space construction information holding/updating unit 231, a virtual space construction information reproduction unit 232 as the third part, a virtual space construction information history management/selection unit 233 as the fourth part, and data stored in memory. It comprises a multimodal document 234 which is a structure, a spatial information acquisition section 401 and a user information/avatar information acquisition section 501 . Reference numerals "1100", "1101", "1102", and "1301" in FIG. 1 indicate the flow of information transmitted and received between units.
 第1の実施の形態では、複数仮想空間構築/配信部200と、仮想空間構築情報保持/更新部231と、仮想空間構築情報再生部232と、仮想空間構築情報履歴管理/選択部233と、マルチモーダルドキュメント234と、空間情報取得部401と、ユーザ情報/アバタ情報取得部501は、情報処理装置であるサーバ110上に設置されているが、この構成に限定されない。例えば、これらは各々別のサーバに設置されていても良いし、クラウド上やユーザ端末101に設置されてもよいし、ブロックチェーン技術を用いて分散して設置されていても良い。サーバ110は、メモリに記憶されているコミュニケーションプログラムを実行することで、複数仮想空間構築/配信部200、仮想空間構築情報保持/更新部231、仮想空間構築情報再生部232、仮想空間構築情報履歴管理/選択部233、空間情報取得部401、ユーザ情報/アバタ情報取得部501として機能する。 In the first embodiment, a multiple virtual space construction/distribution unit 200, a virtual space construction information holding/update unit 231, a virtual space construction information reproduction unit 232, a virtual space construction information history management/selection unit 233, The multimodal document 234, the spatial information acquisition unit 401, and the user information/avatar information acquisition unit 501 are installed on the server 110, which is an information processing device, but the configuration is not limited to this. For example, these may be installed in separate servers, may be installed in the cloud or in the user terminal 101, or may be installed in a distributed manner using blockchain technology. By executing the communication program stored in the memory, the server 110 performs multiple virtual space construction/distribution unit 200, virtual space construction information holding/updating unit 231, virtual space construction information reproduction unit 232, virtual space construction information history. It functions as a management/selection unit 233 , a space information acquisition unit 401 and a user information/avatar information acquisition unit 501 .
 ユーザ端末101は、第1実施形態に係るコミュニケーションシステム100が提供するメタバースに参加するユーザが保持する機器である。ユーザは、パソコンやスマートフォン、ヘッドマウントディスプレイなどのさまざまなユーザ端末101を用いてメタバースに参加する。メタバースには様々なものがあるが、いずれの場合もアバタの仮想空間での三次元的な位置情報またはユーザの実空間における位置情報を用い、相互の距離や場所によって交流できる範囲や度合い、仮想的な知覚が及ぶ範囲を決めている。 The user terminal 101 is a device held by a user who participates in the Metaverse provided by the communication system 100 according to the first embodiment. Users participate in the Metaverse using various user terminals 101 such as personal computers, smartphones, and head-mounted displays. There are various types of Metaverses, but all of them use the three-dimensional positional information of the avatar in the virtual space or the positional information of the user in the real space. It determines the range of physical perception.
 ユーザ端末101からは、ユーザのIDやユーザが使用するアバタのIDと併せて、会話のテキスト情報や音声情報や動画情報、アバタの動きや表情や服装を指定するアバタ操作情報、アバタまたはユーザの空間内での三次元的な座標位置や移動変化を表す空間移動情報などのマルチモーダルな(様々な情報伝達手段による)情報が、複数仮想空間構築/配信部200に向けて送信される。 From the user terminal 101, together with the ID of the user and the ID of the avatar used by the user, text information, audio information, and video information of the conversation, avatar operation information specifying the movement, facial expression, and clothing of the avatar, avatar or user's Multimodal information (via various information transmission means) such as three-dimensional coordinate positions in space and spatial movement information representing changes in movement is transmitted to multiple virtual space construction/distribution unit 200 .
 ユーザは、ユーザ端末101から、会話の際にスタンプ(スティッカー)や写真を送信し、添付書類を送り、あるいはスライドショーを用いても良い。こうした会話に付属するデータもマルチモーダルな情報として送信される。 From the user terminal 101, the user may send stamps (stickers) and photos during conversation, send attached documents, or use a slide show. Data that accompanies these conversations is also transmitted as multimodal information.
 ユーザ端末101のマルチモーダルな情報は、第一の部分である複数仮想空間構築/配信部200に送信される。この複数仮想空間構築/配信部200は、第1の実施の形態ではサーバ110上に設置されているが、この構成に限定されず、クラウド上に設置されても良いし、ユーザ端末101に設置し、それぞれのユーザ端末101がP2Pで接続されて相互に通信するようにしても良い。 The multimodal information of the user terminal 101 is transmitted to the multiple virtual space construction/distribution unit 200, which is the first part. This multiple virtual space construction/distribution unit 200 is installed on the server 110 in the first embodiment, but is not limited to this configuration, and may be installed on the cloud or installed on the user terminal 101. However, each user terminal 101 may be connected by P2P and communicate with each other.
 複数仮想空間構築/配信部200は、ある仮想空間にいるユーザに対して、ユーザが送信したマルチモーダルな情報と当該仮想空間にいる他のユーザのマルチモーダルな情報をまとめて、ユーザ端末101が仮想空間を構築するのに必要な情報をリアルタイムに処理し、その処理結果をユーザ端末101に配信する。同時に、複数仮想空間構築/配信部200は、別な仮想空間にいるユーザに対しても同様の処理を行ない、その処理結果をユーザ端末101に配信する。複数仮想空間構築/配信部200は、すべてのユーザに対してそれぞれのユーザがいる仮想空間を同時に並列処理し、その処理結果を遂次配信する。 The multiple virtual space construction/distribution unit 200 puts together multimodal information transmitted by a user in a certain virtual space and multimodal information of other users in the virtual space, and the user terminal 101 Information necessary to construct a virtual space is processed in real time, and the processing result is delivered to the user terminal 101. - 特許庁At the same time, the multiple virtual space construction/distribution unit 200 performs the same processing for users in another virtual space, and distributes the processing result to the user terminal 101 . The multiple virtual space construction/distribution unit 200 simultaneously performs parallel processing of the virtual spaces in which each user exists for all users, and sequentially distributes the processing results.
 複数仮想空間構築/配信部200は、各仮想空間で起きるイベントも処理し、ユーザ端末101に配信する。ユーザ同士でゲームを競う場合には、ユーザ間の接触判定や勝敗判定などのゲームに必要な判定を行ない、各ユーザ端末101に配信する。以上のように、複数仮想空間構築/配信部200は仮想空間で起きることに関する情報を処理し、各ユーザ端末101に配信する。 The multiple virtual space construction/distribution unit 200 also processes events occurring in each virtual space and distributes them to the user terminal 101 . When users compete in a game, judgments necessary for the game, such as judgment of contact between users and judgment of victory or defeat, are performed and distributed to each user terminal 101 . As described above, the multiple virtual space construction/distribution unit 200 processes information about what happens in the virtual space and distributes it to each user terminal 101 .
 ユーザ端末101は配信された情報に基づき仮想空間画像をレンダリングし、音声や動画を再生する。
 空間、空間内のオブジェクトおよびアバタのグラフィックデータやサウンドデータは、複数仮想空間構築/配信部200が配信するようにしても良いし、別なシステム、サービス、クラウド、ブロックチェーン上に保持してユーザ端末101が必要に応じて取得するようにしても良いし、あるいはユーザ端末101のプログラムに予め含めても良いし、以上の方法を組みあわせても良い。
The user terminal 101 renders a virtual space image based on the distributed information, and reproduces audio and video.
The space, objects in the space, and graphic data and sound data of the avatar may be distributed by the multiple virtual space construction/distribution unit 200, or may be stored on another system, service, cloud, or blockchain and provided to the user. The information may be acquired by the terminal 101 as necessary, or may be included in advance in the program of the user terminal 101, or the above methods may be combined.
 複数仮想空間構築/配信部200が仮想空間で起きるイベントを管理し配信するのと同時に、第二の手段である仮想空間構築情報保持/更新部231は、ユーザ端末101に配信するのと仮想空間で起きていることを構成するマルチモーダルな情報と同様の情報をマルチモーダルドキュメント234に時間経過情報とともに保存する。「マルチモーダルドキュメント234」は、メモリに記憶されたデータ構造である。
 これらの、保存または更新される「仮想空間で特定の時間に起きたことを表す情報」を「仮想空間構築情報」と呼ぶ。仮想空間構築情報保持/更新部231による「仮想空間構築情報」の保存はユーザがユーザ端末101から指示して行わせても良いし、あるいは仮想空間の管理者やシステム管理者が管理者端末から指示して行わせても良い。
At the same time that the multiple virtual space construction/distribution unit 200 manages and distributes events that occur in the virtual space, the virtual space construction information holding/updating unit 231, which is the second means, distributes to the user terminal 101 and the virtual space. Information similar to the multimodal information that constitutes what is happening in the multimodal document 234 is stored along with time course information. A "multimodal document 234" is a data structure stored in memory.
These "information representing what happened in the virtual space at a specific time" to be saved or updated are called "virtual space construction information". The storage of the "virtual space construction information" by the virtual space construction information holding/updating unit 231 may be instructed by the user from the user terminal 101, or may be performed by the virtual space administrator or system administrator from the administrator terminal. You can instruct me to do it.
 ある特定の仮想空間について「仮想空間構築情報」を保存できるようにしても良い。あるいは、複数の仮想空間を移動するユーザにあわせて、複数の仮想空間にまたがる「仮想空間構築情報」をまとめて保存できるようにしても良い。
 前者を「空間を基準に仮想空間構築情報を保存する」、後者を「ユーザやアバタを基準に仮想空間構築情報を保存する」と呼ぶ。
It may be possible to store "virtual space construction information" for a specific virtual space. Alternatively, "virtual space construction information" spanning multiple virtual spaces may be collectively saved in accordance with a user who moves between multiple virtual spaces.
The former is called "storing the virtual space construction information based on the space", and the latter is called "storing the virtual space construction information based on the user or avatar".
「仮想空間構築情報」の保存開始と終了のタイミングについては様々な方式が考えられる。例えば、オンライン会議の録画のようにユーザが明示的に開始と終了を指定する方式、仮想空間の管理者やシステム管理者が一定時間ごとの開始時間と終了時間を設定する方式、誰もいない仮想空間にアバタが入ってきた時点で自動的に開始されアバタが一人もいなくなった時点で終了する方式などが考えられる。 Various methods are conceivable for the timing of starting and ending saving of the “virtual space construction information”. For example, a method in which the user explicitly specifies the start and end of an online conference, a method in which the virtual space administrator or system administrator sets the start and end times at regular intervals, a method in which no one A method of automatically starting when an avatar enters the space and ending when there are no more avatars can be considered.
 「仮想空間構築情報」は、ユーザ端末101からのマルチモーダルな情報と同様の書式で記述する。第三の手段である仮想空間構築情報再生部232は、マルチモーダルドキュメント234に保存された特定の「仮想空間構築情報」を動画の録画再生のように時間情報に基づいて再生し、複数仮想空間構築/配信部200に遂次送信する(図1の符号1102参照)。複数仮想空間構築/配信部200がその情報をユーザ端末101に配信することで、ユーザは過去の一定時間にある特定の仮想空間で起きたことを過去にタイムトリップするように再体験できる。
 また、複数仮想空間構築/配信部200または仮想空間構築情報再生部232は、「仮想空間構築情報」を再生する際に、その再生速度を変え、あるいは再生情報をスキップすることにより、ユーザの求めに応じて仮想空間の事象発生時間の早送り、巻き戻し、一時停止、スキップを行うことができる。
 また、仮想空間構築情報保持/更新部231の処理と仮想空間構築情報再生部232の処理は、独立して行なわれるようにしても良い。これにより、仮想空間構築情報再生部232は、仮想空間構築情報保持/更新部231による処理と同時に、またはこの処理の完了後に、情報の再生を開始することが可能である。
The “virtual space building information” is written in the same format as the multimodal information from the user terminal 101 . A virtual space construction information reproduction unit 232, which is a third means, reproduces specific "virtual space construction information" stored in the multimodal document 234 based on time information like video recording and reproduction, and reproduces a plurality of virtual spaces. It sequentially transmits to the construction/distribution unit 200 (see reference numeral 1102 in FIG. 1). The multiple virtual space constructing/distributing unit 200 distributes the information to the user terminal 101, so that the user can re-experience what happened in a specific virtual space at a certain time in the past as if time-tripping into the past.
In addition, when reproducing the "virtual space construction information", the multiple virtual space construction/distribution unit 200 or the virtual space construction information reproduction unit 232 changes the reproduction speed or skips the reproduction information so as to meet the user's request. You can fast-forward, rewind, pause, and skip the event occurrence time in the virtual space according to the time.
Also, the processing of the virtual space construction information holding/updating unit 231 and the processing of the virtual space construction information reproducing unit 232 may be performed independently. As a result, the virtual space construction information reproducing unit 232 can start reproducing information at the same time as the processing by the virtual space construction information holding/updating unit 231 or after the completion of this processing.
 第1の実施の形態のコミュニケーションシステム100は、単独のユーザとして過去の仮想空間で起きたことを再体験するだけでなく、複数ユーザがグループを組み、そのグループをシステムに登録して、グループメンバー全員で同時に「仮想空間構築情報」を再生し、過去の仮想空間で一緒に行動できるように、構成されても良い。 The communication system 100 according to the first embodiment not only re-experiences what happened in the past virtual space as a single user, but also allows a plurality of users to form a group, register the group in the system, and participate as group members. It may be configured so that all members can play back the "virtual space construction information" at the same time and act together in the past virtual space.
 さらに、再生される過去の仮想空間の中でのユーザの行動のマルチモーダルな情報をユーザ端末101から複数仮想空間構築/配信部200に送信し、ユーザ端末101からの新たな情報と過去の「仮想空間構築情報」とを併せて処理し配信することもできる。これにより、ユーザは過去の仮想空間を見聞きするだけでなく、新たな行動を付け加え、新たな発言を付け加え、あるいは過去の行動や発言を差し替え、過去の仮想空間で起きたことを更新することができる。 Furthermore, the user terminal 101 transmits multimodal information of the user's behavior in the reproduced past virtual space to the multiple virtual space construction/distribution unit 200, and the new information from the user terminal 101 and the past " "Virtual space construction information" can be processed and distributed together. As a result, the user not only sees and hears the virtual space in the past, but also adds new actions, adds new remarks, replaces past actions and remarks, and updates what happened in the past virtual space. can.
 仮想空間構築情報保存/更新部231は、更新された仮想空間で起きたことを、マルチモーダルドキュメント234に保存しても良い。その際に、過去の「仮想空間構築情報」の開始から終了までを再生し、更新しても良いし、その一部の時間だけを再生し、更新しても良い。 The virtual space construction information storage/update unit 231 may store in the multimodal document 234 what happened in the updated virtual space. At that time, the past "virtual space construction information" may be reproduced and updated from the start to the end, or only a part of the time may be reproduced and updated.
 最初に保存された仮想空間構築情報を「オリジナルの仮想空間構築情報」、更新された仮想空間構築情報を「更新された仮想空間構築情報」と呼ぶ。「更新された仮想空間構築情報」をさらに更新して保存しても良い。「オリジナルの仮想空間構築情報」を「更新した仮想空間構築情報」を更新1世代目、更新1世代目の「更新された仮想空間構築情報」をさらに更新した「仮想空間構築情報」を更新2世代目、更新(n)世代目の「仮想空間構築情報」を更新した「仮想空間構築情報」を更新(n+1)世代目と呼ぶ。 The first saved virtual space construction information is called "original virtual space construction information", and the updated virtual space construction information is called "updated virtual space construction information". The "updated virtual space construction information" may be further updated and saved. Updated "Original virtual space construction information" to "Updated virtual space construction information" Updated "Virtual space construction information" which further updated "Updated virtual space construction information" of the 1st generation and updated 1st generation 2 The "virtual space construction information" obtained by updating the "virtual space construction information" of the update (n)th generation is referred to as the update (n+1)th generation.
 仮想空間構築情報保存/更新部231は、「更新された仮想空間構築情報」を保存する際に、「オリジナルの仮想空間構築情報」との差分情報だけを保存するようにして過剰に膨大な情報が保存され記憶装置を逼迫するのを回避しても良い。 The virtual space construction information saving/updating unit 231 saves only difference information from the "original virtual space construction information" when saving the "updated virtual space construction information". may be saved to avoid tying up storage.
 複数仮想空間構築/配信部200は、「オリジナルの仮想空間構築情報」と「更新された仮想空間構築情報」とをパラレルワールドのような別の独立した世界に属するものとして扱う。オリジナルの仮想空間にいるユーザは更新される仮想空間に後から参加するユーザやアバタを見ることはできない。一方、更新される仮想空間に後から参加したユーザはオリジナルの仮想空間にいるユーザやアバタを見ることができる。 The multiple virtual space construction/distribution unit 200 treats "original virtual space construction information" and "updated virtual space construction information" as belonging to different independent worlds such as parallel worlds. Users in the original virtual space cannot see users or avatars who join later in the updated virtual space. On the other hand, users who join the updated virtual space later can see the users and avatars in the original virtual space.
 仮想空間構築情報再生部232は、オリジナルの仮想空間構築の保存が終了する前に、時間を多少遅らせてオリジナルの仮想空間構築の再生を開始し、オリジナルの仮想空間構築の保存と更新された仮想空間情報の保存とを同時に並列して行なえるようにしても良い。これにより、少しずつ時間がずれたパラレルワールドが存在するような状態になる。複数仮想空間構築/配信部200は、同時に複数の仮想空間を並列処理するのと同様に、異なる時間軸、世界線に属する仮想空間を異なる仮想空間として処理する。 The virtual space construction information reproducing unit 232 starts reproducing the original virtual space construction with a slight delay before the saving of the original virtual space construction is completed, and saves the original virtual space construction and updates the virtual space. Storage of spatial information may be performed in parallel at the same time. As a result, a state is created in which a parallel world with a slight time lag exists. The multiple virtual space construction/distribution unit 200 processes virtual spaces belonging to different time axes and world lines as different virtual spaces in the same way as parallel processing of multiple virtual spaces at the same time.
 空間、空間内のオブジェクト、アバタのグラフィックデータやサウンドデータが別なシステム、サービス、サイト、クラウド、ブロックチェーン上に保持されている場合には、複数仮想空間構築/配信部200は、空間情報取得部401ならびにユーザ情報/アバタ情報取得部501を介して、その時の空間やアバタを再現するのに必要な情報を取得し(図1の符号1301参照)、マルチモーダルドキュメント234に保存する。 When the space, objects in the space, avatar graphic data and sound data are held on another system, service, site, cloud, or blockchain, the multiple virtual space construction/distribution unit 200 acquires space information. Information necessary to reproduce the space and avatar at that time is acquired via the unit 401 and the user information/avatar information acquisition unit 501 (see reference numeral 1301 in FIG. 1) and stored in the multimodal document 234 .
 グラフィックデータやサウンドデータが書き換えられ、あるいは削除されることがあるため、複数仮想空間構築/配信部200は、それらをアーカイブする。グラフィックデータやサウンドデータに関しては、それぞれのデータの著作者や配信者がアーカイブするかどうかのユーザ権限を設定できるようにしても良い。 Since graphic data and sound data may be rewritten or deleted, the multiple virtual space construction/distribution unit 200 archives them. As for graphic data and sound data, the author or distributor of each data may be allowed to set user authority as to whether or not to archive the data.
 ユーザは、ユーザ端末101から複数仮想空間構築/配信部200または仮想空間構築情報再生部232に再生時間制御のリクエストを送り、仮想空間で起きることを動画の早送りや巻き戻し再生のように時間制御をできるようにしても良い。
 たとえば、ユーザは、倍速で他のアバタが行動している仮想空間の中を通常の速度で歩き回る、長時間の会話に時間短縮して効率的に参加する、といった「自分とまわりの世界の進む時間が異なる世界」を体験することもできる。
The user sends a request for playback time control from the user terminal 101 to the multiple virtual space building/distribution unit 200 or the virtual space building information playback unit 232, and time-controls what happens in the virtual space like fast-forwarding or rewinding playback of a moving image. You may allow
For example, a user can walk around at normal speed in a virtual space where other avatars are acting at double speed, or participate in a long conversation efficiently by shortening the time. You can also experience a world where time is different.
 仮想空間構築情報保存/更新部231が「仮想空間構築情報」をマルチモーダルドキュメント234に保存または更新するのと同時に、第四の部分である仮想空間構築情報履歴管理/選択部233は、その「仮想空間構築情報」を指し示すID、その「仮想空間構築情報」が更新されたものである場合は更新されるもとの「仮想空間構築情報」を指し示すID、保存や更新された時間などの情報を更新履歴としてマルチモーダルドキュメント234に保存し、管理する。 At the same time when the virtual space construction information storage/update unit 231 saves or updates the "virtual space construction information" in the multimodal document 234, the virtual space construction information history management/selection unit 233, which is the fourth part, ID pointing to "virtual space building information", ID pointing to the original "virtual space building information" to be updated if the "virtual space building information" has been updated, information such as the time when it was saved or updated is saved in the multimodal document 234 as an update history and managed.
 ユーザ端末101から仮想空間構築情報履歴管理/選択部233が管理する履歴情報にアクセス1100できるようにして、ユーザが過去の「仮想空間構築情報」を選択できるようにする。ユーザの選択に応じて、仮想空間構築情報履歴管理/選択部233は、仮想空間構築情報再生部232に、再生する「仮想空間構築情報」を指示する(図1の符号1101参照)。このようにすることで、ユーザは「仮想空間構築情報」の再生を指定可能となる。
 「仮想空間構築情報」の保存を指定したユ―ザや管理者が、仮想空間構築情報履歴管理/選択部233での閲覧や再生に関するユーザアクセス権限を指定できるようにしても良い。
The history information managed by the virtual space construction information history management/selection unit 233 is made accessible 1100 from the user terminal 101 so that the user can select past "virtual space construction information". According to the user's selection, the virtual space construction information history management/selection unit 233 instructs the virtual space construction information reproduction unit 232 to reproduce the "virtual space construction information" (see reference numeral 1101 in FIG. 1). By doing so, the user can designate reproduction of the "virtual space construction information".
The user or administrator who designates the storage of the “virtual space construction information” may be allowed to designate the user access authority for browsing and reproduction in the virtual space construction information history management/selection section 233 .
 仮想空間構築情報履歴管理/選択部233は、「オリジナルの仮想空間構築情報」から派生する「更新された仮想空間構築情報」を、「オリジナルの仮想空間構築情報」を起点とする系統樹のような構造で管理し、ユーザ端末101を介してユーザに表示する。
 このように系統樹のように表示することで、ユーザは仮想空間で起きたことがどのように改変されたかを辿ることができる。系統樹とそれぞれの更新の差分情報をテキストなどで簡易に表示し、簡便に変更の過程が閲覧できるようにしても良い。
The virtual space construction information history management/selection unit 233 treats the "updated virtual space construction information" derived from the "original virtual space construction information" as a phylogenetic tree starting from the "original virtual space construction information". structure, and displayed to the user via the user terminal 101 .
By displaying like a phylogenetic tree in this way, the user can trace how what happened in the virtual space has been altered. The phylogenetic tree and the difference information of each update may be simply displayed in text or the like so that the change process can be viewed easily.
 また、このようにユーザが仮想空間(メタバース)の改変履歴を辿れるようにすることで、仮想空間における過去の改竄がどのように行われたかをユーザが知ることができる。さらには、仮想空間構築情報履歴管理/選択部233に、適宜の判断手段を備えたAIを組み合わせれば、悪意ある事象の改竄などを検出することも可能となる。 In addition, by enabling the user to trace the alteration history of the virtual space (metaverse) in this way, the user can know how past alterations were made in the virtual space. Furthermore, by combining the virtual space construction information history management/selection unit 233 with an AI equipped with appropriate judgment means, it is possible to detect tampering with malicious events.
 仮想空間構築情報再生部232は、仮想空間を再生する際に、オリジナルの仮想空間にいたユーザやアバタと更新された仮想空間で追加されたユーザやアバタに異なる視聴覚的なエフェクトを加えて、更新が識別できるようにしても良い。更新された発言については、その発言のテキスト表示色を変える、音声に異なる効果音を加えるなどして識別できるようにしても良い。アバタや発言を表示するフキダシの横に更新何世代目にあたるかを数字で表示しても良い。 When reproducing the virtual space, the virtual space construction information reproduction unit 232 adds different audiovisual effects to the users and avatars who were in the original virtual space and the users and avatars added in the updated virtual space. can be identified. The updated utterance may be identified by changing the text display color of the utterance or by adding a different sound effect to the voice. It is also possible to display a numerical value indicating how many generations the update corresponds to next to the balloon that displays the avatar or remarks.
 専用のエディタによって「仮想空間構築情報」を編集し、仮想空間での行動せずに「更新された仮想空間情報」を作成できるようにしても良い。何らかのアルゴリズムによって「更新された仮想空間情報」を自動処理、たとえば特定のアバタを過去の出来事から除去するなどの更新を行なっても良い。
 このように第1の実施の形態のコミュニケーションシステム100においては、メタバースで過去に起きたことを自由に改変できるため、歴史改変的なフェイクが流布するのを防止する必要がある。
 このため、仮想空間構築情報履歴管理/選択部233が履歴を厳密に管理することが望ましい。さらには履歴データをブロックチェーンで分散管理し、システム管理者でさえも容易に改変を行なえないようにしても良い。
A dedicated editor may be used to edit the "virtual space construction information" to create "updated virtual space information" without taking any action in the virtual space. The "updated virtual space information" may be automatically processed by some algorithm, for example, updated such as removing a specific avatar from past events.
As described above, in the communication system 100 of the first embodiment, it is possible to freely modify what happened in the past in the Metaverse, so it is necessary to prevent the dissemination of history-altering fakes.
Therefore, it is desirable that the virtual space construction information history management/selection unit 233 strictly manages the history. Furthermore, history data may be distributed and managed by blockchain so that even system administrators cannot easily modify it.
 (第2の実施の形態)
 図2は、本開示の第2の実施の形態に係るコミュニケーションシステム100Aの構成例を示す概念図である。この図2に示すように、第2実施形態に係るコミュニケーションシステム100Aは、複数のユーザ端末101と、各ユーザ端末101に対応して複数設けられるデータ形式変換部301と、第一の部分である複数仮想空間構築/配信部200と、第二の部分である仮想空間構築情報保持/更新部231と、第三の部分である仮想空間構築情報再生部232と、第四の部分である仮想空間構築情報履歴管理/選択部233と、メモリに記憶されたデータ構造であるマルチモーダルドキュメント234と、空間情報取得部401と、ユーザ情報/アバタ情報取得部501と、仮想空間構築情報学習部261と、AIアバタ/AIイベント生成部271と、を備える。複数仮想空間構築/配信部200は、インプットインターフェース201と、アウトプットインターフェース202と、空間振り分け部203と、複数の仮想空間構築部221~223と、を備える。なお、図1中の符号「1100」、「1101」「1102」、「1200」、「1201」、「1301」等は、各部間で送受信される情報の流れを示している。
(Second embodiment)
FIG. 2 is a conceptual diagram showing a configuration example of a communication system 100A according to the second embodiment of the present disclosure. As shown in FIG. 2, the communication system 100A according to the second embodiment includes a plurality of user terminals 101, a plurality of data format conversion units 301 provided corresponding to each user terminal 101, and a first part. Multiple virtual space construction/distribution unit 200, second part virtual space construction information holding/updating unit 231, third part virtual space construction information reproducing unit 232, and fourth part virtual space A construction information history management/selection unit 233, a multimodal document 234 which is a data structure stored in memory, a space information acquisition unit 401, a user information/avatar information acquisition unit 501, and a virtual space construction information learning unit 261. , and an AI avatar/AI event generator 271 . The multiple virtual space constructing/distributing unit 200 includes an input interface 201, an output interface 202, a space sorting unit 203, and multiple virtual space constructing units 221-223. Reference numerals "1100", "1101", "1102", "1200", "1201", "1301", etc. in FIG. 1 indicate the flow of information transmitted and received between the units.
 第2実施形態では、複数仮想空間構築/配信部200と、仮想空間構築情報保持/更新部231と、仮想空間構築情報再生部232と、仮想空間構築情報履歴管理/選択部233と、マルチモーダルドキュメント234と、空間情報取得部401と、ユーザ情報/アバタ情報取得部501と、仮想空間構築情報学習部261と、AIアバタ/AIイベント生成部271とは、サーバ110上に設置されているが、この構成に限定されない。例えば、これらは各々別のサーバに設置されていても良いし、クラウド上やユーザ端末101に設置されてもよいし、ブロックチェーン技術を用いて分散して設置されていても良い。 In the second embodiment, a multiple virtual space construction/distribution unit 200, a virtual space construction information storage/update unit 231, a virtual space construction information reproduction unit 232, a virtual space construction information history management/selection unit 233, and a multimodal The document 234, the spatial information acquisition unit 401, the user information/avatar information acquisition unit 501, the virtual space construction information learning unit 261, and the AI avatar/AI event generation unit 271 are installed on the server 110. , but not limited to this configuration. For example, these may be installed in separate servers, may be installed in the cloud or in the user terminal 101, or may be installed in a distributed manner using blockchain technology.
 この第2の実施の形態のコミュニケーションシステム100Aにおいて、ユーザはユーザ端末101で仮想空間のアバタを操作し会話を行い、あるいはユーザ端末101を持って実空間を移動し、ユーザ端末101を傾け、特定の方向に向ける。ユーザ端末101は、カメラやマイク、GPSなどを備える。ユーザ端末101は、マルチモーダルな情報を、複数仮想空間構築/配信部200のインプットインターフェース201に送信する。
 また、複数仮想空間情報構築/配信部200の処理結果は、アウトプットインターフェース202を通してユーザ端末101に配信される。
In the communication system 100A of the second embodiment, a user operates an avatar in a virtual space with the user terminal 101 to have a conversation, or moves in real space holding the user terminal 101, tilts the user terminal 101, and specifies in the direction of A user terminal 101 includes a camera, a microphone, a GPS, and the like. User terminal 101 transmits multimodal information to input interface 201 of multiple virtual space construction/distribution unit 200 .
Also, the processing result of the multiple virtual space information constructing/distributing unit 200 is distributed to the user terminal 101 through the output interface 202 .
 ユーザ端末101からのマルチモーダルな情報がインプットインターフェース201に送信される過程で、データ形式変換部301は、ユーザが入力可能な形式やユーザにとって利便性の高い形式のデータをシステムにとって扱いやすい形式に変換する。たとえば、データ形式変換部301は、マルチモーダルな情報を、仮想空間構築情報再生部232で時間操作しやすい形式、後から更新が容易な形式、データ量を抑えてシステムに負荷がかからない形式に変換する。 In the process of transmitting multimodal information from the user terminal 101 to the input interface 201, the data format conversion unit 301 converts data in a format that can be input by the user or a format that is highly convenient for the user into a format that is easy for the system to handle. Convert. For example, the data format conversion unit 301 converts multimodal information into a format that facilitates temporal manipulation by the virtual space construction information reproduction unit 232, a format that facilitates updating later, and a format that reduces the amount of data and does not impose a load on the system. do.
 アウトプットインターフェース202から配信される情報も、データ形式変換部301によって、ユーザが活用しやすい形式に変換される。 The information delivered from the output interface 202 is also converted by the data format conversion unit 301 into a format that is easy for the user to utilize.
 たとえば、ユーザが静かな場所にいて声が出せない場合、テキストで会話入力すると、データ形式変換部301はテキストを音声で他のユーザに伝えられるように変換し、他のユーザの音声はテキストに変換する。
 ユーザが自撮り動画を配信できない状況では、データ形式変換部301または仮想空間構築情報再生部232は、予め用意した自分の顔写真をもとに、会話に合わせて会話する動画をAIなどによって生成し、インプットインターフェース201に送信しても良い。
 ユーザがそれぞれの別の言語で会話したときは、データ形式変換部301が翻訳を行なっても良い。
For example, if the user is in a quiet place and cannot speak, the data format conversion unit 301 converts the text so that it can be conveyed to other users by voice, and the other user's voice is converted to text. Convert.
In a situation where the user cannot distribute the self-portrait video, the data format conversion unit 301 or the virtual space construction information reproduction unit 232 generates a video of the conversation according to the conversation based on the face photo prepared in advance by AI or the like. and may be transmitted to the input interface 201 .
When users converse in different languages, the data format conversion unit 301 may translate.
 加えて、メタバース内では視覚や聴覚が不自由な人と健常者がお互いに意識せず境なく交流できるようにする。たとえば聴覚障害者の場合、データ形式変換部301は他のユーザの音声会話をテキストに変換して表示し、テキストで入力した会話を他のユーザに合成音声で伝わるようにする。視覚障害者の場合、データ形式変換部301は他のユーザのテキストチャットを音声に変換する。 In addition, in the Metaverse, visually and hearing impaired people and able-bodied people will be able to interact seamlessly without being aware of each other. For example, in the case of a hearing-impaired person, the data format conversion unit 301 converts the voice conversation of another user into text and displays it, so that the text-inputted conversation can be conveyed to the other user by synthesized speech. For the visually impaired, the data format conversion unit 301 converts text chat of other users into voice.
 こうしたデータ形式変換部301は、各ユーザ端末101に設けて、クライアント側でデータ変換処理しても良いし、サーバ110やクラウド側に設けて、これらの側でデータ変換処理しても良い。ユーザ端末101に充分な処理能力がある場合には、クライアント側にデータ形式変換部301を置いた方がシステム全体の負荷分散になる。また翻訳変換のような優れたシステムが既存する場合は、データ形式変換部301は、このような他のシステムやサービスの機能を使って変換を行っても良い。データ形式変換の詳細については、図3、図4を参照して後述する。 Such a data format conversion unit 301 may be provided in each user terminal 101 and data conversion processing may be performed on the client side, or may be provided on the server 110 or cloud side and data conversion processing may be performed on these sides. If the user terminal 101 has sufficient processing power, placing the data format conversion unit 301 on the client side will balance the load of the entire system. If an excellent system such as translation conversion already exists, the data format conversion unit 301 may use the function of such other system or service to perform the conversion. Details of the data format conversion will be described later with reference to FIGS.
 インプットインターフェース201とアウトプットインターフェース202はサーバ110に設置しているが、クラウド上に設置しても良いし、それぞれのユーザ端末101上に設置してユーザ端末101が相互に通信を行なうようにしても良い。 Although the input interface 201 and the output interface 202 are installed on the server 110, they may be installed on the cloud, or may be installed on each user terminal 101 so that the user terminals 101 communicate with each other. Also good.
 インプットインターフェース201はユーザ端末101からのマルチモーダルな情報と仮想空間構築情報再生部232からのマルチモーダルな情報や後述のAIアバタ/AIイベント生成部271からのマルチモーダルな情報を同じように扱う。 The input interface 201 handles the multimodal information from the user terminal 101, the multimodal information from the virtual space construction information reproduction unit 232, and the multimodal information from the AI avatar/AI event generation unit 271, which will be described later.
 インプットインターフェース201が受信したマルチモーダルな情報は、空間振り分け部203により複数の仮想空間構築部221~223に振り分けられる。空間振り分け部203は、サーバ110のロードバランサーに類似した並列処理による負荷分散のための仕組みだが、振り分けのロジックがロードバランサーとは異なる。
 オリジナルの仮想空間構築の場合は、空間振り分け部203は、同じ仮想空間にいるユーザを同一の仮想空間構築部221~223に振り分ける。その仮想空間構築部221~223の負荷が高まり処理が追い付かなくなった場合には、空間振り分け部203は、同じ仮想空間にいるユーザでも異なる仮想空間構築部221~223に振り分ける。ユーザがグループで行動する指定を行なっている場合には、空間振り分け部203は、優先的に同じ仮想空間構築部221~223に振り分ける。
Multimodal information received by the input interface 201 is distributed to a plurality of virtual space construction units 221 to 223 by a space distribution unit 203 . The spatial distribution unit 203 is a mechanism for distributing loads by parallel processing similar to the load balancer of the server 110, but the distribution logic is different from that of the load balancer.
In the case of the original virtual space construction, the space distribution unit 203 distributes users in the same virtual space to the same virtual space construction units 221-223. When the loads on the virtual space construction units 221 to 223 increase and the processing cannot keep up, the space distribution unit 203 distributes users in the same virtual space to different virtual space construction units 221 to 223. If the user has designated to act in a group, the space distribution unit 203 preferentially distributes to the same virtual space construction units 221 to 223 .
 更新された仮想空間構築の場合は、空間振り分け部203は、グループで行動するユーザを同じ仮想空間構築部221~223に振り分ける。あるいは、空間振り分け部203は、更新された仮想空間構築情報の配信を不特定多数のユーザに対して配信する場合には受信したユーザを同じ仮想空間構築部221~223に振り分ける。
 同じ仮想空間であっても、オリジナルの仮想空間と更新された仮想空間、異なる更新世代の仮想空間を同じ仮想空間構築部221~223に振り分ける必要はない。
In the case of the updated virtual space construction, the space sorting section 203 sorts the users acting in groups to the same virtual space building sections 221-223. Alternatively, when distributing the updated virtual space construction information to an unspecified number of users, the space distribution unit 203 distributes the received users to the same virtual space construction units 221 to 223 .
Even if they are the same virtual space, it is not necessary to distribute the original virtual space, the updated virtual space, and the virtual spaces of different update generations to the same virtual space constructing units 221 to 223 .
 メタバースにおいてシステムは同時に多数の仮想空間を取り扱う必要があり、単独のサーバやプロセスで扱うには負荷が高い。このため、仮想空間構築部221~223は複数の仮想空間を並列処理する。仮想空間構築部221~223はシステム全体の負荷、仮想空間にはいっているアクティブなユーザ数に応じて、増減できる。ある特定の誰もいない仮想空間にユーザないしアバタが入った時点をもって仮想空間構築部221~223を実行するスレッドを自動的にたて、その仮想空間から一人もユーザないしアバタがいなくなった時点をもってそのスレッドを自動的に終了するようにする。図2では221、222、223の3つの仮想空間構築部と、その間に「…」と表記された2つの矩形が表示されているが、これはシステムが許す限り多数の仮想空間構築部が存在し得ることを表している。 In the Metaverse, the system needs to handle many virtual spaces at the same time, and the load is too high for a single server or process to handle. Therefore, the virtual space constructing units 221 to 223 process a plurality of virtual spaces in parallel. The virtual space constructing units 221 to 223 can be increased or decreased according to the load of the entire system and the number of active users in the virtual space. A thread for executing the virtual space construction units 221 to 223 is automatically created when a user or avatar enters a specific empty virtual space, and when there is no user or avatar left in the virtual space. Allow threads to terminate automatically. In FIG. 2, three virtual space construction units 221, 222, and 223 are displayed, and two rectangles marked with "..." are displayed between them. represents what is possible.
 仮想空間構築部221~223は特定のサーバ110上に設置されているが、クラウド上に仮想的に設置されても、P2P接続された各ユーザ端末101上に分散して設置されても良い。空間振り分け部203は仮想空間構築部221~223が新たに開設され、あるいは閉鎖されるのを把握できるようにしておく。 The virtual space construction units 221 to 223 are installed on a specific server 110, but they may be installed virtually on the cloud or distributed on each user terminal 101 connected by P2P connection. The space sorting unit 203 is designed to be able to grasp when the virtual space constructing units 221 to 223 are newly opened or closed.
 ユーザまたはそのアバタがいるのと同一の空間にいる他のユーザの情報を含む空間で起きたことは仮想空間構築部221~223で処理され、アウトプットインターフェース202を介してユーザ端末101に配信される。同時に仮想空間構築情報保存/更新部231は仮想空間で起きたことをマルチモーダルドキュメント234に保存または更新する。 What happens in the space containing the information of other users in the same space where the user or his/her avatar is is processed by the virtual space constructing units 221 to 223 and delivered to the user terminal 101 via the output interface 202. be. At the same time, the virtual space construction information storage/update unit 231 stores or updates what happened in the virtual space in the multimodal document 234 .
 仮想空間構築情報保存/更新部231も仮想空間構築部221~223と同様に並列処理して負荷分散を図っても良いが、「ユーザやアバタを基準に仮想空間を保存する」例のように複数の仮想空間にまたがる「仮想空間情報」の保存が行われる場合があるので、仮想空間構築部221~223とは独立して設置する。 The virtual space construction information storage/update unit 231 may perform parallel processing in the same way as the virtual space construction units 221 to 223 to distribute the load. Since "virtual space information" that spans multiple virtual spaces may be saved, it is installed independently of the virtual space constructing units 221-223.
 また、仮想空間構築情報保存/更新部231は、仮想空間の際現に必要なグラフィックデータを空間情報取得部401、ユーザ情報/アバタ情報取得部501を介して取得しマルチモーダルドキュメント234に保存1301する。
 仮想空間構築情報保存/更新部231は、ユーザ端末101がその時点の仮想空間を再現できるように、空間情報取得部401、ユーザ情報/アバタ情報取得部501を介してグラフィックデータを取得しても良い(図2に示す符号1303参照)。
Also, the virtual space construction information storage/update unit 231 acquires graphic data required for the virtual space through the space information acquisition unit 401 and the user information/avatar information acquisition unit 501, and stores them in the multimodal document 234 (1301). .
The virtual space construction information storage/update unit 231 acquires graphic data via the space information acquisition unit 401 and the user information/avatar information acquisition unit 501 so that the user terminal 101 can reproduce the virtual space at that time. Good (see reference numeral 1303 shown in FIG. 2).
 マルチモーダルドキュメント234には「仮想空間構築情報」が保存されるが、これらはユーザやアバタが特定の状況でどのような行動を取ったかという情報でもある。マルチモーダルドキュメント234には多数の「仮想空間構築情報」が保存され、必要充分なユーザまたはアバタの行動を表すデータを学習データとして、仮想空間構築情報学習部261が機械学習により特定の状況下でのユーザまたはアバタの行動を予測し、行動を生成することができる。
 同様に、仮想空間構築情報学習部261は、どのような属性のユーザまたはアバタが集まれば仮想空間でどのようなイベントが起きるかについても予測し、生成できる。
 これにより、コミュニケーションシステム100Aは、実際にはユーザが一人も参加していないのに多くのアバタで賑わい、さまざまなイベントが起きるメタバースの様子を構築できる。また、コミュニケーションシステム100Aは、アドベンチャーゲームにおけるNPC(ノンプレイヤーキャラクタ)の村人で賑わう、ユーザに情報を伝えてくれる村のようなものも構築できる。
The "virtual space construction information" is stored in the multimodal document 234, and this information is also information about what actions the user or avatar took in a specific situation. The multimodal document 234 stores a large amount of "virtual space construction information", and the virtual space construction information learning unit 261 uses machine learning to learn data representing necessary and sufficient behavior of the user or avatar under a specific situation. user or avatar behavior can be predicted and behavior can be generated.
Similarly, the virtual space construction information learning unit 261 can also predict and generate what kind of event will occur in the virtual space if users or avatars with what attributes gather.
As a result, the communication system 100A can construct a Metaverse that is crowded with many avatars and various events occur even though no user actually participates. In addition, the communication system 100A can construct a village-like village that is filled with NPC (non-player character) villagers in an adventure game and that conveys information to the user.
 仮想空間構築情報学習部261の学習結果は、AIアバタ/AIイベント生成部271に伝えられ、AIアバタ/AIイベント生成部271がAIで行動するアバタやAIで制御されるイベントを生成する。このアバタやイベントは、ユーザ端末101からのマルチモーダルな情報や「仮想空間構築情報」と同様の書式で記述されてAIアバタ/AIイベント生成部271からインプットインターフェース201に送信することで(図2の符号1201参照)、ユーザのアバタと同様に複数仮想空間構築/配信部200で扱える。
 AIアバタ/AIイベント生成部271がユーザのマルチモーダルな情報に応じて対応を変えることで、ユーザの行動に適切に応対するアバタや状況が生成できる。たとえば、生成されたアバタや状況は、会話形式での動的な学習環境の構築などにも活用できる。
The learning result of the virtual space construction information learning unit 261 is transmitted to the AI avatar/AI event generation unit 271, and the AI avatar/AI event generation unit 271 generates an avatar acting by AI and an event controlled by AI. These avatars and events are described in the same format as multimodal information and "virtual space construction information" from the user terminal 101, and are transmitted from the AI avatar/AI event generation unit 271 to the input interface 201 (Fig. 2 1201) can be handled by the multiple virtual space construction/distribution unit 200 in the same way as the user's avatar.
The AI avatar/AI event generation unit 271 changes the response according to the user's multimodal information, thereby generating an avatar and a situation that appropriately responds to the user's behavior. For example, generated avatars and situations can be used to build dynamic learning environments in conversational style.
 メタバースにおいては通常、ユーザは同時に一体のアバタしか操作できないが、ユーザの行動を把握した性格のあるAIアバタが可能になれば、ユーザが同時に複数の仮想空間で行動する自分の分身のようなAIアバタを所持し、それぞれの仮想空間で起きたことを後で早送りで体験し、有限な時間を有効に使えるようになる。
 また、ユーザがオフライン状態にあっても自分の分身のアバタが自分のタスクやミッションをこなしてくれるようになる。
 こうした可能性は、実世界での時間とメタバースの時間が1対1で進行するメタバースの課題、ユーザの時間がどんどん費やされてしまう課題への解決策にもなる。
In the Metaverse, a user can normally operate only one avatar at a time, but if an AI avatar with a personality that comprehends the user's behavior becomes possible, it will be possible for the user to act in multiple virtual spaces at the same time. Possessing an avatar, you can experience what happened in each virtual space later in fast-forward, and you will be able to use your limited time effectively.
Also, even if the user is offline, the avatar of the user's alter ego will perform the task or mission of the user.
Such possibilities also provide a solution to the problem of the Metaverse, in which the time in the real world and the time in the Metaverse progress in a one-to-one relationship, and the problem of the user's time being consumed more and more.
 ユーザが仮想空間構築情報学習部261の設定を操作できるようにすれば、AIアバタやAIイベントをカスタマイズできるようになる。ユーザはカスタマイズしたAIアバタやAIイベントを売買できる。
 たとえば、ユーザは対談イベントが行われる仮想空間や、専門知識を説明できるAIアバタを、NFT("Non-fungible token"、非代替性トークン)を利用してその唯一無二性を担保されたものとして扱い、売買することも可能になる。
If the user can operate the setting of the virtual space construction information learning unit 261, it becomes possible to customize AI avatars and AI events. Users can buy and sell customized AI avatars and AI events.
For example, users can use NFT ("Non-fungible token", non-fungible token) to secure the uniqueness of the virtual space where the dialogue event is held and the AI avatar that can explain their specialized knowledge. It will be possible to treat it as such and sell it.
 図3、図4はスマートフォンを使ったメタバースのマルチモーダルな入力手順の一例及び他の例を説明するための説明図である。以下、図3、図4を参照してユーザ端末101でのユーザ操作ならびにデータ形式変換部301でのデータ変換について説明する。  Figures 3 and 4 are explanatory diagrams for explaining one example and another example of a multimodal input procedure for the Metaverse using a smartphone. User operations at the user terminal 101 and data conversion at the data format conversion unit 301 will be described below with reference to FIGS. 3 and 4. FIG.
 ここではユーザ端末101としてスマートフォンを使った例としたが、ユーザ端末101がスマートフォンに限定されるものではない。ユーザ端末101にはカメラ102、マイク103が備えられており、実空間でのユーザ位置を取得するGPS機能やスマートフォンの傾きや動きを検出する各種の図示しないセンサが内蔵されている。説明のため、図3、図4ではスマートフォン画面を上下に分割し、上部にコンピュータグラフィックスによる仮想空間1001を表示し、下部に各入力に応じた画面を表示したが、これは一例であり、インターフェースデザインを限定するものではない。
 ユーザ端末101は、仮想空間を視覚的に表示せず、立体的に定位した音でユーザの位置や、周辺の環境、他のユーザの位置などを表現しても良い。
Although a smart phone is used as the user terminal 101 in this example, the user terminal 101 is not limited to a smart phone. The user terminal 101 is provided with a camera 102 and a microphone 103, and incorporates a GPS function for acquiring the user position in real space and various sensors (not shown) for detecting the tilt and movement of the smartphone. For the sake of explanation, in FIGS. 3 and 4, the smartphone screen is divided into upper and lower parts, the virtual space 1001 by computer graphics is displayed in the upper part, and the screen corresponding to each input is displayed in the lower part, but this is just an example. It does not limit interface design.
The user terminal 101 may express the position of the user, the surrounding environment, the positions of other users, and the like with stereoscopically localized sounds without visually displaying the virtual space.
 仮想空間1001には、当該ユーザのアバタ1002と、同じ空間にいる他のユーザのアバタ1003が表示されている。 In the virtual space 1001, an avatar 1002 of the user and avatars 1003 of other users in the same space are displayed.
 図3の(a)はテキスト入力による会話入力の例である。ユーザ端末101の画面下部にはテキスト入力画面1010が表示されており、ユーザはキーボード(タッチパネル)で文字入力し、送信ボタン1011で入力した一連の文字を送信する。
 この例では、”Hello”とテキストを送信し、続けてイラスト画像によるスタンプ(Sticker)を送信する。ユーザは「時間経過」3000にそって画面に一連の「タップ入力」3012を行なう。例では「H」、「e」、「l」、「l」、「o」、送信ボタン、スタンプを順にタップする。説明ではスタンプの操作は簡略化した。送信されるデータは「テキスト"Hello"」3013と「スタンプID」3014の2つのデータ群として処理される。
FIG. 3(a) is an example of conversation input by text input. A text input screen 1010 is displayed at the bottom of the screen of the user terminal 101 , and the user inputs characters using a keyboard (touch panel) and transmits a series of characters input using a send button 1011 .
In this example, the text "Hello" is sent, followed by a sticker with an illustration image. The user makes a series of 'tap inputs' 3012 on the screen along the 'time lapse' 3000 . In the example, "H", "e", "l", "l", "o", the send button, and the stamp are tapped in order. In the explanation, the operation of the stamp is simplified. The data to be transmitted is processed as two data groups of “text “Hello”” 3013 and “stamp ID” 3014 .
 「テキスト"Hello"」3013や「スタンプID」3014と併せて「共通テキストデータ」3001が暗号化されて送信される。「共通テキストデータ」3001としてはユーザIDやアバタID、ユーザ端末101のローカル時刻、実空間でのユーザ位置などが送信される。 "Common text data" 3001 is encrypted and transmitted together with "text "Hello"" 3013 and "stamp ID" 3014. As "common text data" 3001, a user ID, an avatar ID, the local time of the user terminal 101, the user's position in real space, and the like are transmitted.
 図3の(b)は音声による会話入力の例である。この例では「音声チャット画面」1021の下に「音声送信ボタン」1020を表示している。ユーザは「音声送信ボタン」1020を長押ししながら発話する。「音声送信ボタン」1020の「長押し開始から終了まで」3022の「音声」3023が一連の発話データとして処理される。この「音声」は、データ形式変換部301によって「音声テキスト変換処理」3024される。一連の「会話音声データ」3025は、音声テキスト変換されたテキストと音声データを指し示すID、元の「音声データ」3026、「共通テキストデータ」3001で構成される。 (b) of FIG. 3 is an example of voice conversation input. In this example, a “send voice button” 1020 is displayed below the “voice chat screen” 1021 . The user speaks while pressing the “send voice button” 1020 for a long time. "Voice" 3023 of "from start to end of long press" 3022 of "send voice button" 1020 is processed as a series of speech data. This “speech” is subjected to “speech text conversion processing” 3024 by the data format conversion unit 301 . A series of “conversation voice data” 3025 is composed of text converted into voice-to-text, ID indicating voice data, original “voice data” 3026 , and “common text data” 3001 .
 会話音声をテキスト送信することにより、仮想空間構築情報保存/更新部231は、「更新された仮想空間構築情報」において過去の会話をテキスト編集できる。データ形式変換部301は、編集したテキストをテキスト音声変換し、過去の会話音声と差し替えることもできる。AI合成音声でユーザの声に似せれば、データ形式変換部301は、過去の仮想空間での会話をその時にユーザが喋ったかのように差し替えられる。 By transmitting the conversation voice as text, the virtual space construction information storage/update unit 231 can text-edit the past conversation in the "updated virtual space construction information". The data format conversion unit 301 can also convert the edited text into text-to-speech and replace the past conversation speech. If AI-synthesized voice mimics the user's voice, the data format conversion unit 301 can replace the past conversation in the virtual space as if the user had spoken at that time.
 図3の(c)は動画による会話入力の例である。この例は音声チャットの例と類似している。ユーザはユーザ端末101の画面に表示された「ビデオチャット画面」1031の「動画送信ボタン」1030を長押ししながら会話する。ビデオチャットの場合はハンズフリーで操作したい場合も多いので、ユーザ端末101は、音量で発話の開始と終了を識別するようにしても良い。
 データ形式変換部301によって「音声」3033は「音声テキスト変換」3024され、「動画」3034は「動画圧縮」3035され、一連のデータ3036として変換されたテキスト、「音声データ」3023、「動画データ」3037と「共通テキストデータ」3001がユーザ端末101からサーバ110に送信される。
 「動画圧縮」3035では、データ形式変換部301は、データの転送負荷を低減する一般的な動画圧縮アルゴリズムのほかに、動画から発話の母音にあたる顔のフレーム画像や動きのある特徴的なフレーム画像を抽出し、それらのキーフレーム画像だけを送信しても良い。他のユーザのユーザ端末101は、キーフレーム画像を動画に補間してユーザがしゃべる動画を生成する。
 「更新された仮想空間構築情報」において発話の一部が編集された場合、データ形式変換部301は、キーフレーム画像を入れ替えて補間して動画化することにより、過去の仮想空間でのその時にユーザが喋ったかのように差し替えられる。
 一枚の顔の写真からアニメ―ション顔の動画を生成する技術があるが、これを応用して、データ形式変換部301は、ユーザ固有の特徴的な動きをしたフレーム画像だけを送信し、通常の会話を行なっている際の動画はユーザの静止画から生成しても良い。
(c) of FIG. 3 is an example of a conversation input using a moving image. This example is similar to the voice chat example. The user converses while pressing and holding the 'video transmission button' 1030 on the 'video chat screen' 1031 displayed on the screen of the user terminal 101 . In the case of video chat, since hands-free operation is often desired, the user terminal 101 may identify the start and end of speech by volume.
"Audio" 3033 is subjected to "speech text conversion" 3024 by the data format conversion unit 301, "moving image" 3034 is subjected to "moving image compression" 3035, and text converted as a series of data 3036, "sound data" 3023, "moving image data 3037 and “common text data” 3001 are transmitted from the user terminal 101 to the server 110 .
In the "moving image compression" 3035, the data format conversion unit 301 uses a general moving image compression algorithm that reduces the data transfer load, and also converts moving images into frame images of faces corresponding to vowels of speech and characteristic frame images with movement. and send only those keyframe images. The user terminal 101 of another user interpolates the keyframe image into a moving image to generate a moving image of the user speaking.
When part of the utterance is edited in the “updated virtual space construction information”, the data format conversion unit 301 replaces and interpolates the key frame images to create a video, thereby converting the speech in the past virtual space at that point in time. It is replaced as if the user had spoken.
There is a technique for generating an animation of an animated face from a photograph of a face. A moving image during normal conversation may be generated from a still image of the user.
 図4の(a)はアバタの空間内の移動に関するものである。ユーザ端末101の画面の下部には、アバタの仮想空間内の位置をわかりやすく表示する斜め上から見た視点(クオーター・ビュー)の「アバタ移動操作画面」1040が表示されている。地図アプリのように真上から見た表示(バーズアイ・ビュー)を用いても良いし、ヘッドマウントディスプレイの専用入力端末などでアバタの移動操作を行なっても良い。
 「アバタ移動操作画面」1040には「自分のアバタ」1043と「他のユーザのアバタ」1042が描かれている。また、「ユーザがいる仮想空間」1041と「隣接する仮想空間」1042が表示されている。仮想空間をウェブサイトのページのようにブロック分割し、それらのブロックを接続して大きな仮想空間を表す方式は、例えば特許文献2(特開2002-123841号公報)記載の技術を用いることができる。
FIG. 4(a) relates to movement of the avatar in space. At the bottom of the screen of the user terminal 101, an "avatar movement operation screen" 1040 is displayed obliquely from above (quarter view) to clearly display the position of the avatar in the virtual space. A display viewed from directly above (bird's eye view) like a map application may be used, or an avatar movement operation may be performed using a dedicated input terminal for a head-mounted display.
On the "avatar movement operation screen" 1040, "own avatar" 1043 and "another user's avatar" 1042 are drawn. Also, a “virtual space where the user is present” 1041 and an “adjacent virtual space” 1042 are displayed. A technique described in Patent Document 2 (Japanese Unexamined Patent Application Publication No. 2002-123841), for example, can be used as a method of dividing a virtual space into blocks like a website page and connecting those blocks to represent a large virtual space. .
 ユーザは「アバタ移動操作画面」1040をタップしてアバタを移動させる。画面の特定位置をタップすると、アバタがその場所まで歩くアニメーションが再生される。「時間経過」3000に応じて、ユーザは「アバタ移動操作画面」1040の複数の箇所a、b、c、dを連続してタップし(3044)、アバタを移動させる。bでは立ち止まるため2回タップしている。ユーザ端末101は、移動位置だけをサーバ110に送信することで通信負荷やサーバ110の処理負荷を低減させているが、ゲームのように細かい移動データを送信しても良い。 The user taps the "avatar movement operation screen" 1040 to move the avatar. When you tap a specific location on the screen, an animation of your avatar walking to that location is played. In accordance with the "passage of time" 3000, the user continuously taps a plurality of points a, b, c, and d on the "avatar movement operation screen" 1040 (3044) to move the avatar. In b, tap twice to stop. The user terminal 101 reduces the communication load and the processing load of the server 110 by transmitting only the movement position to the server 110, but may transmit detailed movement data like a game.
 アバタの位置情報はコンピュータグラフィクスでオブジェクトの位置を示す一般的な方法にならって、仮想空間の座標系で示された三次元的な位置座標(x,y,z)、アバタの向きを表す回転角度(x,y,z)、拡大縮小の倍率(x,y,z)を数値テキストデータの組で表しても良い。加えて、それぞれの位置と次の位置の間の移動アニメーションの補間方法を送信する。キーフレームアニメーションで一般的な指定方法として、指定位置間を直線的につなぐ「リニア」、滑らかな曲線でつなぐ「スプライン」などがある。
 位置情報は「共通テキストデータ」3001と併せて送信される。送信データの「P1」3045は同一の仮想空間内での移動、「P2」3046は異なる仮想空間への移動を伴う移動を表している。「P2」3046の場合は、移動先の仮想空間を示す情報、たとえばメタバースシステムが独自に保持するIDや仮想空間データのURLがユーザ端末101からサーバ110に送られる。
The avatar's position information follows the general method of indicating the position of an object in computer graphics. The angle (x, y, z) and the scaling factor (x, y, z) may be represented by a set of numeric text data. In addition, it sends how to interpolate the movement animation between each position and the next position. Common specification methods for keyframe animation include "linear", which connects specified positions in a straight line, and "spline", which connects specified positions with a smooth curve.
The location information is transmitted together with the “common text data” 3001 . "P1" 3045 of the transmission data represents movement within the same virtual space, and "P2" 3046 represents movement accompanied by movement to a different virtual space. In the case of “P2” 3046 , information indicating the destination virtual space, such as an ID uniquely held by the Metaverse system and the URL of the virtual space data, is sent from the user terminal 101 to the server 110 .
 図4の(b)は移動以外のアバタ操作に関するものである。アバタが手を振る、挨拶するなどの動作(モーション)、他のアバタの動作と連携する握手などの動作(インタラクション)、喜怒哀楽などの表情、アバタが仮想空間に出現するときや特定の動作に伴う視覚効果(エフェクト)、アバタが身に着けている服や帽子や靴や持ち物(アイテム)などの情報を含む。ここではモーションと表情を例に説明する。 (b) of FIG. 4 relates to avatar operations other than movement. An avatar waving or greeting (motion), a handshake that cooperates with other avatars (interaction), facial expressions such as emotions, avatars appearing in a virtual space, or specific actions It includes information such as visual effects (effects) associated with the avatar, clothes, hats, shoes, and belongings (items) worn by the avatar. Here, motion and facial expressions are explained as an example.
 ユーザ端末101の画面の下部にはアバタの「モーション指定画面」1050が表示されている。画面の上部にはいくつかの「ポーズのサムネイル」1051、画面の下部には「選択中のポーズ」1050が表示される。ユーザは「時間経過」3000に応じて「ポーズのサムネイル」1051をタップし、ポーズをつないでモーションを指定する(3052)。また同様に、ユーザ端末101は表情の選択肢を画面に表示して、ユーザはそれらをタップして表情を指定する(3053)。
 これは簡易的なモーションや表情の指定方法だが、ユーザが体にセンサ機器をつけて動作を記録し(モーションキャプチャ)リアルタイムに送信する方式もある。ユーザ端末101は、さまざまな方式で指定されたモーションや表情データに「共通テキストデータ」3001を併せて送信する。会話のスタンプにモーションや表情を連動させても良い。
An avatar “motion designation screen” 1050 is displayed at the bottom of the screen of the user terminal 101 . Several "thumbnails of poses" 1051 are displayed at the top of the screen, and "selected poses" 1050 are displayed at the bottom of the screen. The user taps the "thumbnail of pose" 1051 according to the "passage of time" 3000, connects poses, and designates a motion (3052). Similarly, the user terminal 101 displays options for facial expressions on the screen, and the user taps them to designate facial expressions (3053).
This is a simple method of specifying motion and facial expressions, but there is also a method in which the user attaches a sensor device to the body and records the movement (motion capture) and transmits it in real time. The user terminal 101 transmits "common text data" 3001 together with motion and facial expression data designated by various methods. A motion or facial expression may be linked to the conversation stamp.
 図4の(c)は連続した音声や動画を送信する場合である。このような場合にはユーザは画面操作を行なわずにユーザ端末101のカメラ102とマイク103に向かって話し続け、自分の「カメラ画像」1060を確認する。
 ユーザ端末101は、「時間経過」3000に伴う切れ間のない「音声」3061と「動画」3062をサーバ110で扱い易いように一定時間ごとに分割し、「音声テキスト変換」3024と「動画圧縮」3035などのデータ形式変換を行なった上で、変換されたテキストや「共通テキストデータ」3001とあわせて連続するデータの群としてサーバ110に送信する(3061)。
(c) of FIG. 4 is a case of transmitting continuous sound and moving images. In such a case, the user continues talking into the camera 102 and the microphone 103 of the user terminal 101 without operating the screen, and confirms his own "camera image" 1060 .
The user terminal 101 divides the continuous “speech” 3061 and “moving image” 3062 accompanying the “passage of time” 3000 at regular time intervals so that the server 110 can easily handle them, and divides them into “speech-to-text conversion” 3024 and “moving image compression” 3024 . After performing data format conversion such as 3035, it is transmitted to the server 110 as a continuous data group together with the converted text and "common text data" 3001 (3061).
 次に、図5、図6、図7を用いて、「ユーザ端末101」からのマルチモーダルな情報を仮想空間構築情報保存/更新部231がどのように一連の「仮想空間構築情報」として保存し、仮想空間構築制御情報履歴管理/選択部233が管理するかを説明する。 Next, referring to FIGS. 5, 6, and 7, how the virtual space construction information storage/update unit 231 stores the multimodal information from the “user terminal 101” as a series of “virtual space construction information”. and how the virtual space construction control information history management/selection unit 233 manages them.
 図5は「空間を基準に仮想空間構築情報を保存する」手順の一例を説明するための説明図である。図6は「ユーザまたはアバタを基準に仮想空間構築情報を保存する」手順の一例を説明するための説明図である。
 いずれの場合も、仮想空間構築情報保存/更新部231が「仮想空間構築情報」の保存を開始する時点を「Start Saving」2001、保存を終了する時点を「End Saving」2004として、この間を「仮想空間構築情報」のひとまとまりとして扱う。図5~図7において、紙面左側に示す符号「2000」、「2010」~「2019」は、仮想空間構築情報保存/更新部231に時間経過とともに送信される一連のデータ群を示す。例えば、「空間Aでの記録」2000は、空間Aで記録され、仮想空間構築情報保存/更新部231に送信されるデータ群である。符号「2100」~「2103」は、これらのデータ群が仮想空間構築情報保存/更新部231でまとめられた結果である「更新される仮想空間構築情報」を示す。また、図5~図7において、紙面右側に示す符号は、イベントがなされる時間(時点)又は実行される処理を示す。
FIG. 5 is an explanatory diagram for explaining an example of a procedure for "saving virtual space construction information based on space". FIG. 6 is an explanatory diagram for explaining an example of a procedure for "saving virtual space construction information based on a user or avatar".
In either case, "Start Saving" 2001 is the point at which the virtual space construction information storage/update unit 231 starts saving the "virtual space construction information", and "End Saving" 2004 is the point at which the saving ends. treated as a set of "virtual space construction information". 5 to 7, the symbols "2000", "2010" to "2019" shown on the left side of the page indicate a series of data groups sent to the virtual space construction information storage/update unit 231 over time. For example, “recorded in space A” 2000 is a data group recorded in space A and transmitted to the virtual space construction information storage/update unit 231 . Codes “2100” to “2103” indicate “updated virtual space construction information” which is the result of putting these data groups together by the virtual space construction information storage/update unit 231 . In FIGS. 5 to 7, the symbols shown on the right side of the paper indicate the time (point of time) at which an event occurs or the process to be executed.
 図5に示す「オリジナルの仮想空間構築情報」について説明する。「空間Aでの記録」2000の例では「Start Event」2002で何らかのイベントが開始され「End Event」2003でイベントが終了する。たとえば「Start Event」で仮想空間にギタリストのアバタが登場して演奏を行ない、「End Event」でその空間から退場する、といった例が考えられる。
 仮想イベントの開始前からユーザB、ユーザCは空間で会話してイベントを待ち、イベント開始直前にユーザAが空間に現れる。ユーザAはイベント終了後まもなく、その仮想空間を立ち去る。
The “original virtual space construction information” shown in FIG. 5 will be described. In the example of “recording in space A” 2000 , some event starts at “Start Event” 2002 and ends at “End Event” 2003 . For example, a guitarist's avatar appears in the virtual space at the "Start Event" and performs, and then exits the space at the "End Event".
Before the start of the virtual event, users B and C converse in the space and wait for the event, and user A appears in the space just before the start of the event. User A leaves the virtual space shortly after the event ends.
 「保存される仮想空間構築情報」2100には、「空間Aでの記録」と、「ユーザAの入力(行動)」2010に記した一連のマルチモーダルな情報、「ユーザBの入力(行動)」2011に記した一連のマルチモーダルな情報、「ユーザCの入力(行動)」2012に記した一連のマルチモーダルな情報が保存される。2010,2011,2012には図3、図4に示したような会話テキスト、会話音声、アバタ動作などが表示される。「ユーザAの入力(行動)」には、ユーザAが空間Aにはいってきた時点「Enter the Space」2005、ユーザAが空間Aから出た時点「Exit the Space」2006が示されている。
 この一連の「仮想空間構築情報」は「オリジナルの仮想空間構築情報」である。
In ``virtual space construction information to be saved'' 2100, ``recording in space A'', a series of multimodal information described in ``user A's input (behavior)'' 2010, and ``user B's input (behavior) 2011, and a series of multimodal information described in 'User C's input (behavior)' 2012 are saved. In 2010, 2011, and 2012, the conversation text, conversation voice, avatar action, etc. shown in FIGS. 3 and 4 are displayed. The "input (behavior) of user A" indicates "Enter the Space" 2005 when user A enters space A and "Exit the Space" 2006 when user A exits space A. FIG.
This series of "virtual space construction information" is "original virtual space construction information".
 図5に示す「更新1世代目+ユーザD」について説明する。ユーザDが「オリジナルの仮想空間構築情報」を更新し、その更新結果を「更新される仮想空間構築情報」2101として保存する。ユーザDは「オリジナルの仮想空間構築情報」の冒頭から再生を開始2007し、音声会話によりいくつかの会話を行ない、ギタリストに拍手するなどのアバタ動作を選択したのちに再生を終了2008する。
 イベントの主催者が自動的に「更新される仮想空間構築情報」を更新するように指定しており、「更新される仮想空間構築情報」2101が更新されている。これは更新1世代目にあたる。
The “updated first generation+user D” shown in FIG. 5 will be described. User D updates the “original virtual space construction information” and saves the update result as “updated virtual space construction information” 2101 . User D starts the reproduction from the beginning of the "original virtual space construction information" 2007, conducts several voice conversations, selects an avatar action such as clapping for the guitarist, and then ends the reproduction 2008.
The organizer of the event has specified to automatically update the 'updated virtual space construction information', and the 'updated virtual space construction information' 2101 has been updated. This is the first updated generation.
 図5に示す「更新1世代目+ユーザE、ユーザB(再訪)」について説明する。ユーザEはユーザBとグループを組んで「仮想空間構築情報」を再生し、その中で行動する。ユーザBのアバタは「オリジナルの仮想空間構築情報」にも参加しているが、ユーザEとともに再度イベントに参加する。ユーザBは遅れて参加し、ユーザEと会話したのち、早めに退出する。
 ユーザBによる更新により「オリジナルの仮想空間構築情報」の「ユーザBの入力」の一部が書き換えられる。更新した発話は先に会話した発話を上書きし、あるいは新たな発話として挿入されている。
 ユーザBのアバタの立ち位置や服装は「オリジナルの仮想空間構築情報」の時と変わっているかもしれない。このような場合、ユーザBが参加している間だけ、オリジナルのユーザBのアバタを差し替えても良い。
 あるいは、映画『バック・トゥ・ザ・フューチャー』のように、オリジナルのユーザBのアバタと更新されたユーザB1のアバタを同時に2体表示しても良い。
The “update first generation + user E, user B (revisit)” shown in FIG. 5 will be described. User E forms a group with user B, reproduces the "virtual space construction information", and acts in it. User B's avatar also participates in the "original virtual space construction information", but participates in the event again with user E. User B joins late, talks with user E, and leaves early.
Due to the update by user B, part of the "input of user B" in the "original virtual space construction information" is rewritten. The updated utterance overwrites the previously spoken utterance, or is inserted as a new utterance.
The standing position and clothes of user B's avatar may be different from those of the "original virtual space construction information". In such a case, the original user B's avatar may be replaced only while user B is participating.
Alternatively, like the movie "Back to the Future", the original avatar of user B and the updated avatar of user B1 may be displayed simultaneously.
 図5に示す「更新2世代目+ユーザD+ユーザF」について説明する。ユーザFは、ユーザDによる「更新された仮想空間構築情報」を再生すると、この再生時のユーザFの行動がさらに更新2世代目の「更新された仮想空間構築情報」2103として保存される。 "Updated second generation + user D + user F" shown in Fig. 5 will be explained. When user F reproduces the "updated virtual space construction information" by user D, user F's behavior at the time of this reproduction is further saved as the updated second generation "updated virtual space construction information" 2103 .
 図6は「ユーザやアバタを基準に仮想空間構築情報を保存する」手順の一例を説明するための説明図である。
 図6に示す「オリジナルの仮想空間構築情報」について説明する。ユーザAは、最初に空間Bを訪れ、そのあと空間Aに移動してイベントに参加し、空間Cに移動する。それぞれ異なる「仮想空間構築情報」が空間A、空間B、空間Cで処理されたユーザAの行動の結果として「保存される仮想空間構築情報」2100にまとめられる。
 空間AにはユーザB、ユーザCがおり、ユーザAが空間Aに滞在する間は彼らのアバタと会話できる。
 ユーザGは途中から空間Aを訪れ、ユーザAと出会って会話した後、ユーザAとともに空間Cに移動し、ユーザAとの会話を続けている。
FIG. 6 is an explanatory diagram for explaining an example of a procedure for "saving virtual space construction information based on users and avatars".
The “original virtual space construction information” shown in FIG. 6 will be described. User A first visits space B, then moves to space A to participate in an event, and moves to space C. Different "virtual space construction information" are summarized as "stored virtual space construction information" 2100 as a result of user A's actions processed in space A, space B, and space C. FIG.
User B and user C are in space A, and user A can talk with their avatars while staying in space A.
User G visits space A halfway through, meets user A and has a conversation with him, then moves to space C with user A, and continues the conversation with user A.
 図6に示す「更新1世代目+ユーザH、ユーザI」について説明する。更新1世代目では「オリジナルの仮想空間構築情報」をユーザH、ユーザIがグループで再生し、ユーザAに同伴するように空間B、空間A、空間Cを移動する。
 ユーザHはユーザAと離れずに空間を移動していくが、ユーザIは少し遅れて空間を移動する。このため、ユーザAを基準に保存された仮想空間情報の更新においては、ユーザHは一時的にユーザAがいる空間にいないことになり、更新情報から姿を消している。
 このような、参加するユーザが一時的に複数の空間にわかれたときに、同時に複数の空間に関する情報を「更新する仮想空間構築情報」に含めても良い。
 ただし、多数のユーザのアバタがまちまちに異なる仮想空間に分散した場合や、更新世代も重ねられて更新の都度新たな空間が追加されたりすると、大量の仮想空間に関する情報を仮想空間構築情報に含める必要が出てくるので、ユーザAと異なる空間に移動したユーザに関しては保存から除外するのが適切である。
The “updated first generation+user H, user I” shown in FIG. 6 will be described. In the first update generation, user H and user I reproduce the "original virtual space building information" as a group, and move in space B, space A, and space C as if accompanied by user A.
User H moves in the space without leaving user A, but user I moves in the space with a slight delay. Therefore, in updating the virtual space information stored with user A as a reference, user H is temporarily out of the space where user A is, and disappears from the update information.
When the participating users are temporarily divided into a plurality of spaces in this way, the information about the plurality of spaces may be included in the "updated virtual space construction information" at the same time.
However, if the avatars of a large number of users are scattered in different virtual spaces, or if update generations are superimposed and new spaces are added each time an update is performed, a large amount of information about the virtual space is included in the virtual space construction information. Since it will be necessary, it is appropriate to exclude users who have moved to a different space from user A from saving.
 図7は、仮想空間構築情報の系統樹を説明するための説明図である。この図7は、図5の例をもとに「オリジナルの仮想空間構築情報」3100を起点に更新1世代目の「更新される仮想空間構築情報」3101、3102,更新2世代目の「更新される仮想空間構築情報」3103などが更新され、系統樹的な分岐を形成していく様子を示している。 FIG. 7 is an explanatory diagram for explaining the phylogenetic tree of the virtual space construction information. Based on the example of FIG. 5, this FIG. 3103 is updated, and a phylogenetic branch is formed.
 このように系統樹的に「仮想空間構築情報」が分岐していく様子は、ユーザの用途によっては必ずしも使いやすいものではない可能性もある。しかしながら、対談の内容をさまざまな参加者が深めていき、アーカイブ化するようなものに関しては、こうした系統樹的な保持の仕方が有効である。
 なお、軽い内容のものに関しては、系統樹表示ではなく多くのユーザが閲覧した仮想空間構築情報のみを表示するようにしても良いし、ユーザにより評価システムを導入して人気の仮想空間構築情報のみをユーザに表示するようにしても良い。
Depending on the intended use of the user, there is a possibility that the way in which the "virtual space construction information" branches in a phylogenetic tree in this way is not necessarily easy to use. However, when various participants deepen the content of the dialogue and archive it, such a method of preserving it in a phylogenetic tree is effective.
In the case of light contents, it is possible to display only the virtual space construction information viewed by many users instead of displaying the phylogenetic tree, or to introduce only the popular virtual space construction information by introducing an evaluation system by users. may be displayed to the user.
 以上、本開示の実施形態を図面により詳述してきたが、上記各実施形態は本開示の例示にしか過ぎないものであり、本開示は上記各実施形態の構成にのみ限定されるものではない。本開示の要旨を逸脱しない範囲の設計の変更等があっても、本開示に含まれることは勿論である。 As described above, the embodiments of the present disclosure have been described in detail with reference to the drawings, but the above embodiments are merely examples of the present disclosure, and the present disclosure is not limited only to the configurations of the above embodiments. . Even if there is a change in design within the scope of the present disclosure, it is of course included in the present disclosure.
 本開示は、メタバースをはじめとする仮想空間を用いた体験型のサービスやコミュニティ・サービスに活用できる。 This disclosure can be used for experience-based services and community services using virtual spaces such as the Metaverse.
 メタバースにおいては、ユーザのオンライン状況が可視化されるため、他のアバタがいない空間=過疎化した空間と見えてしまい、これがユーザにサービス衰退を感じさせる大きな原因のひとつとなっている。また、サービス開始時で多くのユーザが獲得できていない時点では、ユーザが他のユーザに出会うことが稀になり、他のユーザに出会えなかったユーザはサービスを継続するモチベーションを失ってしまう。
 こうしたユーザがいなくなった空間においても、その空間が賑わっていた時間を蓄積し、ユーザがその時間に戻れるようにすることで、過去に蓄積された他のユーザによる「賑わい」を体験し、さらにその「賑わい」に参加できることで、サービス全体が過疎化したという印象を回避できる。このようにメタバースの根本的な課題のひとつである「過疎化した空間」のという課題を解決でき、サービスの収益化に貢献できる。
In the Metaverse, users' online statuses are visualized, so spaces without other avatars appear to be depopulated spaces. In addition, when many users have not been acquired at the start of the service, users rarely meet other users, and users who do not meet other users lose their motivation to continue the service.
Even in a space without such a user, by accumulating the time when the space was busy and allowing the user to return to that time, they can experience the "bustle" of other users accumulated in the past, and further By being able to participate in the "bustle", it is possible to avoid the impression that the service as a whole has depopulated. In this way, we can solve one of the fundamental problems of the Metaverse, the problem of "depopulated spaces," and contribute to the profitability of the service.
 また、本開示は、メタバース内での会話を活かした、対話型の教育コンテンツの作成にも活用できる。本開示のコミュニケーションシステムで提供されるコンテンツは、単なる対話型の教育コンテンツと異なり、たとえば、原子力発電所の危険な原子炉の中を、教師などの説明者と生徒などの受講者が数名のグループになってアバタで歩き回りながら、その仕組みの説明を受けるというようなコンテンツをあとから再生し、グループと一緒になって歩き回り、さらに質問をするなど過去のイベントにタイムトラベルしたような積極的参加が可能になる。図6の例で示したように、複数の仮想空間にまたがって移動するユーザの視点で空間構築情報を保存することで、たとえば単一空間の原子炉の中で説明を行なうのに留まらず、発電所全体の空間を移動しながら、あるいは発電所空間から電気が利用される都市空間といった広いエリアの仮想空間に移動しながら行われる説明を一緒に移動しながら体験できる。 This disclosure can also be used to create interactive educational content that makes use of conversations within the Metaverse. The content provided by the communication system of the present disclosure is different from mere interactive educational content. Active participation as if you traveled back in time to a past event, such as walking around in a group and listening to an explanation of how it works. becomes possible. As shown in the example of FIG. 6, by storing the space construction information from the viewpoint of the user who moves across multiple virtual spaces, for example, it is possible not only to explain in a nuclear reactor in a single space, but also to You can experience explanations while moving around the entire space of the power plant, or while moving from the power plant space to the virtual space of a wide area such as the urban space where electricity is used.
 スマートフォンの画面やスマートグラスなどの専用の機器を用い、現実の場所にアバタを投影するAR("Augmented Reality" 複合現実)技術を使えば、たとえば歴史的な合戦が行われた史跡について、ユーザが実際に史跡の様々な場所をまわりながら、説明員のアバタから説明を受け、質問をするなどの積極的参加が可能になる。説明員は後日、追加されたアバタの質問を確認して、さらに仮想空間を更新し、質問への回答を付け足す。
 この場合も前項の例と同様に、複数の広いエリアにまたがる史跡を説明者のアバタと参加者のアバタが一緒に歩き回りながる体験型の学習を実施できる。また、参加者は自分が興味ある地点でだけ参加することもできる。
By using AR ("Augmented Reality" mixed reality) technology that projects an avatar in a real place using dedicated devices such as smartphone screens and smart glasses, users can see historical sites where historical battles were fought, for example. While actually visiting various places in the historic site, you will be able to actively participate by listening to explanations from avatars and asking questions. At a later date, the facilitator checks the added avatar questions, further updates the virtual space, and adds answers to the questions.
In this case as well, as in the previous example, it is possible to carry out experiential learning in which the presenter's avatar and the participant's avatar walk together around historical sites spanning a plurality of large areas. Participants can also participate only at points of interest to them.
 本開示のコミュニケーションシステムは、たくさんの対話を蓄積し、複数の「仮想空間構築情報」を切り張りして編纂して、冗長な部分を省き、体験する対談集のような新しいコンテンツを作ることもできる。 The communication system of the present disclosure accumulates many dialogues, cuts and compiles multiple pieces of "virtual space construction information", eliminates redundant parts, and creates new content such as a collection of dialogues to experience. can.
 本開示のコミュニケーションシステムは、仮想空間内で独白し、あとから更新してコメントをつける使い方もできる。自らの考えを深化させるアイディア・プロセッサとしてメタバースを活用もできる。単自分の中の相反する意見を2つのアバタを使い分けて自問自答することも、現実の場所に紐付け独白もできる。たとえば、文化遺産についての自分の考えをその中を歩き回りながらまとめることができる。
 まとめた「仮想空間構築情報」をメタバースならではのコンテンツとして収益化もできる。
The communication system of the present disclosure can also be used to make a monologue in the virtual space and update it later to add a comment. You can also use the Metaverse as an idea processor to deepen your thoughts. You can use two different avatars to ask yourself about conflicting opinions within yourself, or you can make a monologue linked to a real place. For example, you can organize your thoughts about cultural heritage while walking around it.
Collected "virtual space construction information" can be monetized as content unique to the Metaverse.
 本開示のコミュニケーションシステムは、人工知能を搭載しても良く、人工知能技術が発展することで、アバタがさらに高度な人間的な振舞いをすることが可能になり、ユーザが要求に的確に応え、ユーザの知識や考え方の一部を受け継いで、ユーザのある特性を切り出した分身として行動する存在になれば、知識共有化の新たなサービスも実現可能になる。再生される仮想空間にこうしたAIアバタが常駐することで、新たな参加者に対して、当初のユーザがその場にいるかのような対応をとることができる。さらにユーザが、各自の専門知識や考え方を持つアバタを育てあげられるようになれば、その利用権を仮想空間対話イベントに付随させて、販売または貸与できる。このような専門知識を売買する「知の市場」が価値を産み出す。 The communication system of the present disclosure may be equipped with artificial intelligence, and with the development of artificial intelligence technology, it becomes possible for the avatar to behave more highly human-like, allowing the user to accurately respond to requests, If a part of the user's knowledge and way of thinking is inherited and it becomes an existence that acts as an alter ego that cuts out a certain characteristic of the user, it will be possible to realize a new service for sharing knowledge. Since such an AI avatar resides in the reproduced virtual space, it is possible to respond to new participants as if the original user were there. Furthermore, if users can nurture avatars with their own specialized knowledge and ways of thinking, the usage rights can be attached to virtual space dialogue events and sold or lent. The “knowledge market” that buys and sells such expertise creates value.
 本開示のコミュニケーションシステムは、アバタを仮想の世界に留めるのではなく、ロボットや自動運転車と組み合わせると、現実の世界で行動しながら、それぞれの場での対話に参加できるようになる。 The communication system of this disclosure does not keep avatars in the virtual world, but when combined with robots and self-driving cars, it will be possible to participate in dialogues in each place while acting in the real world.
 このように本開示は、教育や娯楽などの産業分野でのメタバースの新たな活用を創出する仕組みとして活用可能である。 In this way, the present disclosure can be used as a mechanism for creating new uses of the Metaverse in industrial fields such as education and entertainment.

Claims (3)

  1.  1又は複数のコンピュータから構成されるコミュニケーションシステムであって、第一の部分と、第二の部分と、第三の部分と、第四の部分とを備え、
     少なくとも1人以上のユーザが参加する仮想空間環境または実空間に対応する情報空間環境において、
     前記第一の部分は、ユーザ端末から入力されるテキスト、音声、静止画、動画、アバタ操作のコマンドの何れかを含むマルチモーダルな情報と、その入力情報をもとに取得した仮想空間情報、ユーザ情報ならびにアバタ情報とに加えて、前記第三の部分からの情報により複数の仮想空間を同時に構築制御し、その制御結果を前記ユーザ端末に配信し、
     前記第二の部分は、前記第一の部分が仮想空間を構築制御する際に使用した情報を時間経過と関連付けてメモリに保存または更新し、
     前記第三の部分は、前記第二の部分により保存または更新される情報を再生して前記第一の部分に送信し、
     前記第四の部分は、前記第二の部分が情報を保存または更新した履歴を系統樹的に管理し、ユーザの選択に応じて再生すべき保存または更新された情報を前記第三の部分に指示することを特徴とするコミュニケーションシステム。
    A communication system comprising one or more computers, comprising a first portion, a second portion, a third portion, and a fourth portion,
    In a virtual space environment or an information space environment corresponding to a real space in which at least one or more users participate,
    The first part includes multimodal information including any one of text, voice, still image, video, and avatar operation command input from the user terminal, and virtual space information obtained based on the input information; Simultaneously constructing and controlling a plurality of virtual spaces based on the information from the third part in addition to the user information and the avatar information, and delivering the control results to the user terminal;
    The second part saves or updates the information used when the first part constructs and controls the virtual space in a memory in association with the passage of time,
    said third portion reproduces and transmits to said first portion information stored or updated by said second portion;
    The fourth part systematically manages the history of information saved or updated by the second part, and stores the saved or updated information to be reproduced according to the user's selection in the third part. A communication system characterized by instructing.
  2.  前記第三の部分は、情報を再生する際に、その再生速度を変え、あるいは再生情報をスキップすることにより、ユーザの求めに応じて仮想空間の事象発生時間の早送り、巻き戻し、一時停止、スキップを行うように構成されていることを特徴とする請求項1に記載のコミュニケーションシステム。 The third part changes the playback speed or skips the playback information when playing back the information to fast-forward, rewind, pause, or pause the event occurrence time in the virtual space according to the user's request. 2. The communication system of claim 1, configured to skip.
  3.  前記第二の部分による情報の保存または更新と、前記第三の部分による情報の再生とは独立して行なわれるように構成され、前記第三の部分は、前記第二の部分による情報の保存または更新と同時に、若しくは前記第二の部分による情報の保存または更新の完了後に、情報の再生を開始可能に構成されていることを特徴とする請求項1または2に記載のコミュニケーションシステム。 The storage or update of information by the second part and the reproduction of information by the third part are configured to be performed independently, and the third part stores the information by the second part. 3. The communication system according to claim 1, wherein reproduction of information can be started simultaneously with updating or after completion of saving or updating of information by said second part.
PCT/JP2022/005691 2022-02-14 2022-02-14 Communication system WO2023152976A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2022/005691 WO2023152976A1 (en) 2022-02-14 2022-02-14 Communication system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2022/005691 WO2023152976A1 (en) 2022-02-14 2022-02-14 Communication system

Publications (1)

Publication Number Publication Date
WO2023152976A1 true WO2023152976A1 (en) 2023-08-17

Family

ID=87563975

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2022/005691 WO2023152976A1 (en) 2022-02-14 2022-02-14 Communication system

Country Status (1)

Country Link
WO (1) WO2023152976A1 (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000050226A (en) * 1998-07-28 2000-02-18 Fuji Xerox Co Ltd Method and system for multimodal and asynchronous conference with intervention of computer existing in virtual space

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000050226A (en) * 1998-07-28 2000-02-18 Fuji Xerox Co Ltd Method and system for multimodal and asynchronous conference with intervention of computer existing in virtual space

Similar Documents

Publication Publication Date Title
CN110850983B (en) Virtual object control method and device in video live broadcast and storage medium
Galloway et al. From Michael Moore to JFK Reloaded: Towards a working model of interactive documentary
CN102450032B (en) Avatar integrated shared media selection
Schroeder The social life of avatars: Presence and interaction in shared virtual environments
JP5122433B2 (en) Information communication system and information communication method
WO2012105318A1 (en) Input support device, input support method, and recording medium
Jin et al. A comparison of natural user interface and graphical user interface for narrative in HMD-based augmented reality
US11418848B2 (en) Device and method for interactive video presentation
Garau Selective fidelity: Investigating priorities for the creation of expressive avatars
Kirchner et al. A conference goes virtual: Lessons from creating a social event in the virtual reality
WO2023152976A1 (en) Communication system
Zagalo Narrative design of sadness in heavy rain
WO2023190344A1 (en) Information processing device, information processing method, and program
JP2005055846A (en) Remote educational communication system
Nakatsu Toward the creation of a new medium for the multimedia era
KR20140136288A (en) Thema park video making service pproviding system and method by user's participation
Rome Narrative virtual reality filmmaking: A communication conundrum
Haga Shifting Diegetic Boundaries
JP7445938B1 (en) Servers, methods and computer programs
Geigel et al. Adapting a virtual world for theatrical performance
US20240013488A1 (en) Groups and Social In Artificial Reality
Xue Redefining Audience Participation: The Immersive Interaction in “Wolf and Spices VR” and the Enlightenment of Journalism and Communication
Sermon et al. Liberate your avatar: The revolution will be socially networked
Freeman et al. " My Audience Gets to Know Me on a More Realistic Level": Exploring Social VR Streamers’ Unique Strategies to Engage with Their Audiences
Pan Mobile Cinema

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22925995

Country of ref document: EP

Kind code of ref document: A1