WO2024080135A1 - Display control device, display control method, and display control program - Google Patents

Display control device, display control method, and display control program Download PDF

Info

Publication number
WO2024080135A1
WO2024080135A1 PCT/JP2023/034925 JP2023034925W WO2024080135A1 WO 2024080135 A1 WO2024080135 A1 WO 2024080135A1 JP 2023034925 W JP2023034925 W JP 2023034925W WO 2024080135 A1 WO2024080135 A1 WO 2024080135A1
Authority
WO
WIPO (PCT)
Prior art keywords
display control
avatar
control device
user
virtual space
Prior art date
Application number
PCT/JP2023/034925
Other languages
French (fr)
Japanese (ja)
Inventor
悟郎 高木
智裕 石井
河野 温子
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2024080135A1 publication Critical patent/WO2024080135A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics

Definitions

  • the present disclosure relates to a display control device, a display control method, and a display control program for representing an avatar in virtual reality.
  • Metaverse Various technologies have been put into practical use that present users with a three-dimensional virtual space constructed within a computer or computer network, known as the Metaverse.
  • VR Virtual Reality
  • AR Augmented Reality
  • This disclosure therefore proposes a display control device, a display control method, and a display control program that can represent avatars in a virtual space in a variety of ways.
  • a display control device includes an acquisition unit that acquires user behavior in real space as input information, a determination unit that determines whether the input information includes an input corresponding to a specific action registered in advance, and a display control unit that reflects the user behavior corresponding to the input information in an avatar in a virtual space, and, if it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space.
  • FIG. 2 is a diagram showing an overview of a display control process according to the embodiment.
  • FIG. 11 is a diagram illustrating an example of a determination process according to the embodiment.
  • FIG. 1 is a diagram illustrating an example of the configuration of a display control device according to an embodiment.
  • FIG. 4 is a diagram illustrating an example of a gesture storage unit according to the embodiment.
  • 5 is a flowchart showing a flow of a display control process according to the embodiment.
  • FIG. 11 is a diagram (1) for explaining a first variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram (2) for explaining a first variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram (1) for explaining a second variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram illustrating an example of a determination process according to the embodiment.
  • FIG. 1 is a diagram illustrating an example of the configuration of a display control device according to an embodiment.
  • FIG. 4 is
  • FIG. 11 is a diagram (2) for explaining a second variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram (3) for explaining a second variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram for explaining a third variation of the display control process according to the embodiment;
  • FIG. 11 is a diagram (2) for explaining a third variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram (3) for explaining a third variation of the display control process according to the embodiment.
  • FIG. 4 is a diagram for explaining a third variation of the display control process according to the embodiment.
  • FIG. 5 is a diagram for explaining a third variation of the display control process according to the embodiment;
  • FIG. 6 is a diagram for explaining a third variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram for explaining a fourth variation of the display control process according to the embodiment;
  • FIG. 13 is a diagram (2) for explaining a fourth variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram (3) for explaining a fourth variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram for explaining a fifth variation of the display control process according to the embodiment;
  • FIG. 13 is a diagram (2) for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 11 is a diagram (3) for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 4 is a diagram for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 5 is a diagram (5) for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 2 is a hardware configuration diagram illustrating an example of a computer that realizes the functions of the display control device.
  • Embodiment 1-1 Overview of display control processing according to embodiment 1-2. Configuration of display control device according to embodiment 1-3. Processing procedure according to embodiment 1-4. Variations of processing according to embodiment 1-4-1. Expression of excitement between avatars 1-4-2. Expression of avatar emotions 1-4-3. Visualization of conversation 1-4-4. Proximity effect between avatars 1-4-5. Expression related to display and interaction of avatars 1-5. Modifications 2. Other embodiments 3. Effects of a display control device according to the present disclosure 4. Hardware configuration
  • FIG. 1 is a diagram showing an overview of the display control process according to the embodiment.
  • the display control process according to the embodiment is executed by a display control device 100, not shown in FIG. 1.
  • the display control device 100 is an information processing terminal for implementing VR and AR technology. That is, in the embodiment, the display control device 100 displays a three-dimensional virtual space (hereinafter simply referred to as "virtual space") constructed in a computer or computer network, such as a metaverse, on a display unit (display, etc.).
  • the display control device 100 also displays an avatar, which is a character that resembles a user, in the virtual space.
  • the display control device 100 is implemented by any information processing device, such as a PC (Personal Computer), a tablet terminal, or a smartphone.
  • the display control device 100 may be, for example, a head mounted display (HMD) or AR glasses.
  • HMD head mounted display
  • the display control device 100 captures an image of a user in real space with an image sensor (such as a camera) and reflects the user's gestures and facial expressions in an avatar in the virtual space. This allows the user to experience the sensation of having an alter-ego of himself or herself present in the virtual space.
  • an image sensor such as a camera
  • users can communicate via the avatars in the same way as in the real space. For example, users can deepen their interactions with other users by greeting them and conversing (voice chat or text chat) via their avatars.
  • virtual spaces can be expressed in a variety of ways, they can be constructed not only as realistic spaces based on real space, but also as spaces with cartoon-like expressions that users can relate to. In this case, it is also desirable for the provider of the virtual space to not only be able to make the avatar reproduce realistic movements based on reality, but also to be able to provide a variety of emotional expressions and diverse movements that users can relate to.
  • the display control device 100 realizes the above processing by the configuration shown below.
  • the display control device 100 acquires user behavior in real space as input information, and determines whether the input information includes an input corresponding to a specific action registered in advance.
  • the display control device 100 then reflects the user behavior corresponding to the input information in an avatar in a virtual space, and if it is determined that the input information includes an input corresponding to a specific action, it further reflects the specific action in the avatar in the virtual space.
  • the display control device 100 registers in advance comical movements including cartoon expressions, such as a thumbs-up, a signature pose of each avatar, and a high-five, as specific movements (hereinafter referred to as "gestures") to be performed by the avatar.
  • the display control device 100 receives an input from the user of an action for activating the gesture, the display control device 100 reflects the gesture, rather than the user's movement itself, in the avatar.
  • the display control device 100 may reflect the user's gesture itself in the avatar. In this way, the display control device 100 reflects both gestures associated with the user's conscious movements and the user's unconscious movements in the avatar.
  • the display control device 100 can display an avatar in a virtual space that not only imitates the user's movements, but also has expressive and comical movements. Through such expressions, the display control device 100 can build a virtual space that makes it easier for users to communicate with each other and to feel familiar with each other.
  • the display control device 100 displays an avatar 10A that exists in a virtual space on an example display screen 20.
  • the display control device 100 uses a camera to constantly capture images of the user corresponding to the avatar 10A (i.e., the user whose movements and expressions are reflected in the avatar 10A).
  • the display control device 100 acquires input information input by the user based on image data obtained from the camera.
  • input information is information acquired from various sensors including the camera, and is a general term for information used for displaying avatars, etc.
  • the display control device 100 acquires the user's facial expression 21 as input information using an image-based facial expression extraction technique.
  • the display control device 100 also acquires the user's skeleton 22 using an image-based skeleton estimation technique.
  • the display control device 100 also acquires gesture information 23 indicating the movement of the user's arms and hands using hand tracking technique.
  • the display control device 100 may acquire not only image data, but also voice or text information input by the user as input information.
  • the display control device 100 then combines this information to accept input information from the user to be reflected in the avatar 10A.
  • the display control device 100 acquires the avatar 10A nodding, shaking its head to indicate refusal, tilting its head in distress, etc. as unconscious actions of the user that are not specific actions.
  • the display control device 100 may also acquire various emotional expressions by the user, such as a laughing expression or expression, or a surprised expression or expression, as unconscious actions. In this disclosure, these are collectively referred to as "unconscious actions.”
  • the example on the left side of Figure 1 shows an example in which the unconscious action of the user raising his hand is reflected in the avatar 10A and displayed in the virtual space.
  • the display control device 100 can also determine, based on the acquired image data, whether or not the user is performing a specific action to trigger a gesture. That is, the display control device 100 determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance. For example, as shown in the right part of FIG. 1, the display control device 100 displays an avatar 10B that exists in a virtual space in a screen display example 25. At this time, the display control device 100 uses a camera to constantly capture images of the user corresponding to the avatar 10B (i.e., the user whose movements and expressions are reflected in the avatar 10B).
  • avatar 10A and avatar 10B are drawn separately to make the comparison easier to understand, but in reality, unconscious actions and conscious actions are expressed by the same avatar.
  • avatar 10 when there is no need to distinguish between avatar 10A, avatar 10B, etc., they will be collectively referred to simply as "avatar 10.”
  • the display control device 100 accepts a request to transition to an input mode for determining the specific action. For example, when the user desires that the avatar 10B make a distinctive gesture based on a specific action by the user, the user requests the display control device 100 to transition to an input mode for consciously making the gesture.
  • the display control device 100 acquires the user's behavior, such as the user's facial expression 26, the user's skeletal structure 27, and gesture information 28, based on the captured image. The display control device 100 then combines the acquired information to determine whether or not the characteristics of the user's behavior match the characteristics for activating a pre-registered gesture. Note that various known technologies may be used for acquiring the user's behavior and determining the gesture.
  • the display control device 100 determines that the behavior input by the user is a movement that has enough characteristics to trigger a gesture, it reflects the gesture in the avatar 10B. For example, as shown in the right part of FIG. 1, the avatar 10B makes a pose with its index finger and thumb. At this time, the display control device 100 may perform a comical expression that includes a dramatic effect such as displaying radial lines around the avatar 10B along with the pose. This expression allows the user to distinguish whether the movement of the avatar 10B is a specific gesture registered in advance or the user's movement itself. In other words, such specific gestures are consciously made to be performed by the avatar 10B by the user. For this reason, in this disclosure, these are collectively referred to as "conscious movements.”
  • the display control device 100 can reflect both these unconscious and conscious actions in the avatar 10 and express them as reactions of the avatar 10. This allows the display control device 100 to display the avatar 10 in the virtual space in a variety of expressions, including not only a direct reflection of the user's movements, but also cartoon-like expressions, etc.
  • the user can switch between the input modes of unconscious actions and conscious actions, and reflect input information from unconscious actions or conscious actions in the avatar 10. Such switching may be performed, for example, based on an explicit designation by the user (manual operation such as pressing a button, voice input, etc.).
  • the display control device 100 can set in advance a spatial range within which the user will perform conscious actions, and switch the input mode based on such a setting. This point will be explained using FIG. 2.
  • FIG. 2 is a diagram showing an example of the determination process according to the embodiment.
  • FIG. 2 shows an example in which a user inputs his/her own behavior by being imaged by the camera 141 of the display control device 100.
  • the user makes hand signs 30 in a pre-set space 31 (e.g., a space within a predetermined distance from the camera 141).
  • the display control device 100 determines whether the user's behavior was performed in the space 31, for example, using image analysis or a depth sensor. If the user's behavior was performed in the space 31, the display control device 100 considers the behavior to be one that ignites a conscious action and determines that a gesture has been ignited. On the other hand, the display control device 100 reflects the user's behavior observed in space 32, which is a range outside the space 31, in the avatar 10 as an unconscious action.
  • the display control device 100 when the display control device 100 is located at a predetermined distance from the camera 141, it does not determine that the input from the user is an action for igniting a gesture, but only determines whether the user's actions performed in the space 31 are actions for igniting a gesture. This allows the user to reflect unconscious and conscious actions in the avatar 10 at will, without having to explicitly specify the switching of input modes.
  • the switching of input modes is not limited to the example shown in FIG. 2.
  • voice chat i.e., voice input
  • text chat i.e., manual character input
  • the display control device 100 may automatically switch to an input mode for extracting conscious actions (gesture detection).
  • the display control device 100 may perform motion blending that reflects both movements while switching modes.
  • the display control device 100 may turn off the input mode itself while the user is performing a UI (User Interface) operation, so that movements of different intentions are not reflected in the avatar 10.
  • the display control device 100 may activate the gesture input mode only within a predetermined time after performing an arbitrary key operation.
  • the display control device 100 may prevent erroneous gesture input by requesting the user to input a predetermined word when performing a gesture.
  • the display control device 100 may activate the gesture input mode when a mute button that turns off voice input from the user is pressed, and may set the display control device 100 to the unconscious action input mode in other cases.
  • the display control device 100 may simultaneously turn off the voice input and transition to a privacy mode (wherein input from the user is not accepted). In this way, the display control device 100 may use different input modes in response to various aspects.
  • the display control process according to the embodiment can reflect unconscious and conscious actions in the avatar 10 as desired, allowing the avatar 10 to be displayed in a variety of ways.
  • Fig. 3 is a diagram showing an example of the configuration of the display control device 100 according to an embodiment.
  • the display control device 100 is one of the components of a display control system 1 according to the embodiment shown in Fig. 3.
  • the display control system 1 includes the display control device 100 and an analysis server 50.
  • the analysis server 50 performs various analysis processes described below and provides the analysis results to the display control device 100.
  • the analysis server 50 performs information processing with a relatively high processing load, such as text-based sentiment analysis and machine learning processing for image recognition, which are not performed by the display control device 100.
  • the analysis server 50 is a cloud server managed by a business operator that operates a service related to virtual space.
  • the display control device 100 has a communication unit 110, a memory unit 120, a control unit 130, a sensor unit 140, and a display unit 150.
  • the display control device 100 may also have an input unit (such as a keyboard or a touch panel) that accepts various operations from a user who operates the display control device 100.
  • the communication unit 110 is realized, for example, by a NIC (Network Interface Card) or a network interface controller.
  • the communication unit 110 is connected to the network N by wired or wireless means, and transmits and receives information to and from the analysis server 50, etc., via the network N.
  • the network N is realized, for example, by a wireless communication standard or method such as Bluetooth (registered trademark), the Internet, Wi-Fi (registered trademark), UWB (Ultra Wide Band), or LPWA (Low Power Wide Area).
  • the storage unit 120 is realized, for example, by a semiconductor memory element such as a random access memory (RAM) or a flash memory, or a storage device such as a hard disk or an optical disk.
  • a semiconductor memory element such as a random access memory (RAM) or a flash memory
  • a storage device such as a hard disk or an optical disk.
  • the storage unit 120 stores various information related to the display control process according to the embodiment.
  • the storage unit 120 includes a gesture storage unit 121.
  • FIG. 4 is a diagram showing an example of the gesture storage unit 121 according to an embodiment. As shown in FIG. 4, the gesture storage unit 121 has items such as "gesture ID,” "gesture,” and "determination item.”
  • Gesture ID is identification information for identifying a gesture.
  • Gesture indicates the specific content of the gesture.
  • Determination item indicates the item for extracting a feature amount used when determining whether or not the user's behavior will trigger the gesture.
  • the information on each gesture may include motion data (such as animation of the skeleton and facial expressions, dramatic effects, etc.) for moving the 3D model of the avatar 10 in accordance with the gesture.
  • the example shown in Figure 4 indicates that the gesture with gesture ID "A01" corresponds to a "thumbs up,” and whether or not to fire that gesture is determined by the user's "hand position” and "arm movement.”
  • the sensor unit 140 is a sensor that detects various environmental information.
  • the sensor unit 140 includes a camera 141, which is an imaging device that is composed of an outward-facing camera that captures images of the outside of the display control device 100, an inward-facing camera that captures images of the user, etc.
  • the camera 141 recognizes a subject (e.g., a user located in real space) located in front of the display control device 100.
  • the camera 141 acquires an image of the user who is the subject, and can calculate the distance from the display control device 100 (in other words, the camera 141) to the subject based on the parallax between images captured by the stereo camera and an analysis of the captured image data.
  • the sensor unit 140 may detect the distance in real space using a depth sensor capable of detecting the distance to any subject, such as a user.
  • the sensor unit 140 also includes a microphone 142 that can measure the user's voice and surrounding sounds. For example, when engaging in voice chat, the user can input voice by using the microphone 142 provided in the display control device 100.
  • the sensor unit 140 may have a function of detecting various information related to the user's motion, such as the orientation, inclination, motion, and moving speed of the user's body, in addition to the camera 141 and microphone 142. Specifically, the sensor unit 140 detects information related to the user's motion, such as information related to the user's head and posture, the motion of the user's head and body (acceleration and angular velocity), the direction of the field of view, and the speed of the viewpoint movement.
  • information related to the user's motion such as information related to the user's head and posture, the motion of the user's head and body (acceleration and angular velocity), the direction of the field of view, and the speed of the viewpoint movement.
  • the sensor unit 140 functions as various motion sensors such as a three-axis acceleration sensor, a gyro sensor, and a speed sensor, and detects information related to the user's motion. More specifically, the sensor unit 140 detects at least one change in the position and posture of the user's head by detecting the components of the yaw direction, pitch direction, and roll direction as the motion of the user's head.
  • the sensor unit 140 does not necessarily have to be provided in the display control device 100, and may be, for example, an external sensor connected to the display control device 100 by wire or wirelessly.
  • the camera 141 does not necessarily have to be provided inside the display control device 100, and may be an imaging device installed outside the display control device 100.
  • the display unit 150 displays various information output from the control unit 130.
  • the display unit 150 is a display that outputs video to the user.
  • the display unit 150 may also include an audio output unit (such as a speaker) that outputs audio.
  • the control unit 130 is realized, for example, by a CPU (Central Processing Unit), MPU (Micro Processing Unit), GPU, etc., executing a program stored inside the display control device 100 (for example, a display control program related to the present disclosure) using a RAM or the like as a working area.
  • the control unit 130 is also a controller, and may be realized, for example, by an integrated circuit such as an ASIC (Application Specific Integrated Circuit) or FPGA (Field Programmable Gate Array).
  • control unit 130 has an acquisition unit 131, a determination unit 132, and a display control unit 133.
  • the acquisition unit 131 acquires various types of information. For example, the acquisition unit 131 acquires input information to be reflected in the avatar 10 by sensing the user using the sensor unit 140. Specifically, the acquisition unit 131 acquires the user's behavior in real space as input information.
  • the acquisition unit 131 may also acquire various operations and setting requests from the user. For example, the acquisition unit 131 accepts designation from the user as to whether or not to reflect a specific action in the avatar 10. In other words, the acquisition unit 131 accepts designation from the user as to whether to set an input mode for inputting an unconscious action or an input mode for inputting a conscious action (gesture).
  • the acquisition unit 131 may acquire surrounding information, etc., using the sensor unit 140, etc. Specifically, as shown in FIG. 2, the acquisition unit 131 acquires the spatial position from the camera 141 when the user's behavior is captured. This allows the determination unit 132, which will be described later, to determine at what distance from the camera 141 the hand sign performed by the user is captured, and therefore the mode can be automatically switched to the gesture mode.
  • the acquisition unit 131 can acquire information not only about the user currently captured by the camera 141, but also about multiple users connected to the virtual space. In other words, the acquisition unit 131 can acquire the behavior of multiple users as input information. By using such input information from multiple users, the determination unit 132 can determine the level of interest in a conversation between multiple users, as described below.
  • the acquisition unit 131 may also receive various requests sent from the user via the UI. For example, the acquisition unit 131 receives requests to change the appearance of the avatar, and various operations for services related to the virtual space, such as purchasing, selecting, or exchanging an object (such as an item in the virtual space) linked to the avatar.
  • the determination unit 132 performs various determinations for controlling the display of the avatar 10. For example, the determination unit 132 determines whether or not the input information from the user includes an input corresponding to a specific action that has been registered in advance.
  • the input corresponding to a specific action that has been registered in advance is, in other words, a registered behavior that the user uses to trigger a gesture.
  • the determination unit 132 determines whether or not the input information includes an input corresponding to a specific action. In other words, the determination unit 132 determines whether or not the behavior performed by the user in front of the camera 141 is an intentional action that triggers a gesture.
  • the determination unit 132 does not determine whether or not the input information includes an input corresponding to a specific action if the position from the camera 141 is farther than a predetermined distance, but determines whether or not the input information includes an input corresponding to a specific action if the position from the camera 141 is within the predetermined distance.
  • the determination unit 132 performs gesture determination for behaviors such as hand signs made by the user in a position relatively close to the camera 141, and does not perform gesture determination for behaviors such as hand signs made by the user outside the set spatial range, determining them to be unconscious actions. This allows the determination unit 132 to freely switch input modes to express the avatar 10.
  • the display control unit 133 controls the display unit 150 to display information output from the control unit 130.
  • the display control unit 133 outputs the virtual space image rendered as video content to the display.
  • the output destination device is not limited to the built-in display, but may be other information devices such as a smartphone, or a video output device such as a television.
  • the display control unit 133 based on input information acquired from the user, such as an image of the user, a voice uttered by the user, or text entered by the user, the display control unit 133 reflects the user's behavior corresponding to the input information in the avatar 10 in the virtual space. Furthermore, when the display control unit 133 determines that the input information includes an input corresponding to a specific action, it reflects a gesture, which is a specific action, in the avatar 10 in the virtual space and displays the avatar 10 in the virtual space.
  • the display control unit 133 when the display control unit 133 is in a mode for inputting unconscious actions, it reflects the user's skeletal data, etc., acquired based on an image of the user, in the avatar 10.
  • the display control unit 133 when the display control unit 133 is in a gesture mode, if the determination unit 132 determines that the user's behavior will trigger a gesture, it reflects a gesture, including a specific animation or special effects, that has been registered in advance, in the avatar 10. That is, gestures made by the avatar 10 are expressed by animations that involve special movements or special effects, such as a thumbs-up, high-five, or clapping.
  • the display control unit 133 may control the sound output by the avatar (pre-registered laughter) in addition to the gestures.
  • Fig. 5 is a flowchart showing the flow of a display control process according to the embodiment.
  • the display control device 100 uses the sensor unit 140 to acquire an image of the user (step S11). As described above, the display control device 100 may also acquire the user's voice, input text, and the like along with the image.
  • the display control device 100 converts the user's image into skeletal data based on known posture estimation techniques, etc. (step S12).
  • the display control device 100 determines whether the current mode is the gesture mode (step S13). If the current mode is the gesture mode (step S13; Yes), the display control device 100 determines the feature point positions and movements in the skeletal data (step S14).
  • the display control device 100 determines whether or not the feature point position and movement correspond to a registered gesture (step S15). If they correspond to a registered gesture (step S15; Yes), the display control device 100 reads out the motion data of the gesture from the storage unit 120 (step S16). The display control device 100 then reflects the read out motion data in the avatar 10 (step S18).
  • step S13 If the mode is not the gesture mode (step S13; No), the display control device 100 does not perform gesture determination, and converts the skeletal data obtained from the user into skeletal data for the avatar 10 (step S17). Then, the display control device 100 reflects the skeletal data in the avatar 10 (step S18).
  • step S15 if the feature point positions and movements do not correspond to a gesture (step S15; No), the display control device 100 also converts the skeletal data obtained from the user into skeletal data for the avatar 10 (step S17). Then, the display control device 100 reflects the skeletal data in the avatar 10 (step S18).
  • the display control device 100 displays the avatar 10 in the virtual space based on the information reflected in the avatar 10 (step S19).
  • the display control device 100 repeats this display control process for each predetermined frame acquired by the sensor unit 140.
  • the display control device 100 displays various visual effects according to the excitement between multiple avatars.
  • the excitement between avatars means the excitement between users.
  • the display control device 100 can express in a way that other users can understand that active communication is taking place between multiple users in the virtual space.
  • the acquisition unit 131 acquires the behavior of multiple users as input information. Then, the determination unit 132 determines the level of excitement of the interaction between the multiple users in the virtual space based on the behavior of the multiple users. Furthermore, the display control unit 133 displays a presentation effect in the vicinity of the multiple users according to the level of excitement of the interaction.
  • the determination unit 132 determines the level of excitement in an exchange based on the number of users participating in the exchange, the frequency of conversations between users, and the degree of overlap in the voices of multiple users.
  • the display control unit 133 may also determine the expression or frequency of occurrence of the dramatic effect based on the level of excitement in the interaction, the body orientation of the avatar 10, and the content of the conversation between multiple users.
  • FIG. 6 is a diagram (1) for explaining the first variation of the display control process according to the embodiment.
  • the example shown in FIG. 6 shows that multiple avatars, avatar 10A, avatar 10B, and avatar 10C, are actively interacting by exchanging greetings and conversations.
  • the display control device 100 determines the level of excitement in the interaction between the multiple users, and displays an arc-shaped effect 200 and radial effect 201 around the avatars based on the determined level of excitement. In this way, the display control device 100 can perform a display that allows multiple avatars who are actively interacting to be distinguished simply by looking at the situation from a distance, even if they are not actually participating in the conversation.
  • the level of excitement between avatars may be derived from various parameters. For example, if the excitement between avatars is due to voice communication such as voice chat, parameters that may be used include the volume, speed, pitch, emotion, frequency of overlap, amount of filler, and the presence or absence of specific language expressions of the users' voices. In other words, the level of excitement in communication in a virtual space is not necessarily measured only by the volume of the voices or the amount of conversation.
  • the display control device 100 can use, for example, "overlap (simultaneous speech by multiple people)" as a parameter with high importance (weighted) for determining excitement.
  • the display control device 100 may also analyze the emotion of each voice by having the analysis server 50 or the like analyze the voices, and may calculate a high excitement level when a more emotional voice is included. Specifically, the display control device 100 may calculate a high excitement level for an exchange that includes shouts expressing admiration or strong emotion, such as "Yay (Japanese)" or “Wow! Mah! (English).” Alternatively, the display control device 100 may calculate a high excitement level based on the frequency of occurrence of greetings or words (positive words based on emotion analysis) that strongly imply an intention of celebration or joy, such as "Cheers!.
  • the display control device 100 can also change the presentation (visual) effects displayed depending on the level of excitement.
  • the display control device 100 may change the size or frequency of the dramatic effect represented by the arc depending on the level of excitement. Specifically, when the level of excitement is relatively high, the display control device 100 may display a large arc so as to cover all of the multiple users participating in the conversation. This allows the display control device 100 to accurately convey the excitement of the multiple users even to a user who is viewing the multiple users from a distance.
  • the display control device 100 may also display a large effect even if conversation is not necessarily active. For example, if multiple users are paying attention to some kind of event in the virtual space, conversation will be reduced, but it is determined that the users are actively interacting with each other. In this case, the display control device 100 may determine the level of excitement based on, for example, the intersection of the avatars' gazes or interactions other than conversation (such as the closeness of the users to each other). Then, depending on the level of excitement, the display control device 100 displays the arc-shaped or radial effect larger or brighter than usual. In this way, the display control device 100 can show other users that multiple users are actively interacting with each other, even without conversation.
  • FIG. 7 is a diagram (2) for explaining a first variation of the display control process according to the embodiment.
  • FIG. 7 shows a composite waveform 205 obtained from a conversation between multiple users.
  • range 206 shows a waveform in which the voices of multiple users are measured to be high, indicating that the conversation is taking place at a relatively loud volume.
  • Range 207 shows a waveform in which multiple voices from multiple users are synthesized, indicating that a lively exchange of conversation is taking place.
  • the display control device 100 determines the level of excitement among multiple users based on such waveform analysis. For example, the display control device 100 determines the level of excitement based on the characteristics of volume changes over a specified period of time. Alternatively, the display control device 100 may estimate the number of participants in a conversation based on waveform analysis, and determine the level of excitement to be higher the greater the estimated number of participants.
  • the display control device 100 may also change the representation of the dramatic effect based on the waveform.
  • the display control device 100 may change the color or density (frequency of occurrence) of an arc-shaped or radial display, the speed at which the arc-shaped or radial light spreads, etc., in accordance with the amplitude of the waveform.
  • the display control device 100 may not necessarily display an arc-shaped effect, but may display a speech bubble with a series of acute angles according to the voice, conversation content, or emotion. This allows the display control device 100 to express, for example, that multiple users are arguing with each other. In this case, the display control device 100 may display speech bubbles or other dramatic effects using a three-dimensional display (application of a rim shader) that is valid regardless of the viewing angle.
  • the display control device 100 may also change the speed at which the special effects are generated depending on the distance from the observer to the multiple users. Specifically, the display control device 100 may display the special effects slower the farther the distance is, and may display the special effects faster the closer the distance is. This allows the observing user to know at a glance how far away the actively interacting users are.
  • the determination unit 132 determines whether or not the input information from the user includes an input corresponding to a specific action linked to a specific emotion, such as laughing. If it is determined that the input information includes a specific action linked to a specific emotion, the display control unit 133 reflects the specific action linked to the specific emotion in the avatar 10.
  • the determination unit 132 may compare the input information with voiceprint information that is registered in advance by the user and is associated with a specific emotion, to determine whether the input information includes an input corresponding to a specific action associated with the specific emotion.
  • the display control unit 133 may also display an avatar 10 that reflects a specific action linked to a specific emotion, and may also display a dramatic effect linked to the specific emotion near the avatar 10.
  • FIG. 8 is a diagram (1) for explaining the second variation of the display control process according to the embodiment.
  • the avatar 10 is making a gesture associated with "laughing," which is an example of a specific emotion.
  • the display control device 100 may display a cartoon-like effect 210 associated with laughter around the avatar 10.
  • the display control device 100 can display an avatar 10 that is rich in emotional expression by having the avatar 10 perform a laughing gesture in response to the user laughing.
  • FIG. 9 is a diagram (2) for explaining a second variation of the display control process according to the embodiment.
  • FIG. 9 shows an example of an audio waveform 215 obtained from a user.
  • the measurement value obtained when the environmental sound (background noise) is measured while the user is engaged in voice chat or the like is set as measurement result 216.
  • the audio waveform obtained when the user is engaged in voice chat or the like is set as measurement result 217.
  • the peak of measurement result 217 is indicated by the line of result 218.
  • the difference between result 218 and measurement result 216 is indicated by difference 219.
  • the display control device 100 uses the microphone 142 to measure the environmental sound (background noise) when the user is engaged in voice chat, etc. Then, the display control device 100 measures the user's voice, estimates the peak volume as laughter, and measures the difference 219 with the environmental sound. In this way, the display control device 100 registers the peak value of the voice and the difference value with the environmental sound for each user.
  • the environmental sound background noise
  • the display control device 100 determines that the user has laughed. According to this method, the display control device 100 can determine whether or not the user has laughed based on the sound volume.
  • FIG. 10 is a diagram (3) for explaining a second variation of the display control process according to the embodiment.
  • FIG. 10 illustrates an example of a voice waveform 220 acquired from a user.
  • range 221 is the waveform when the user's laughter is measured.
  • the display control device 100 (or the analysis server 50) accumulates the waveforms when laughter is measured for each user, and learns the characteristics of the accumulated waveforms using a known voice learning method (e.g., deep learning that extracts characteristic patterns). This allows the display control device 100 to generate a learned voiceprint pattern that indicates the "laughing" state for each user.
  • a known voice learning method e.g., deep learning that extracts characteristic patterns
  • the display control device 100 can determine whether the voice contains laughter by comparing it with the learned voiceprint pattern. This allows the display control device 100 to determine with high accuracy whether the user has laughed.
  • human laughter is diverse, and the emotions contained in laughter include not only joy but also various other emotions. For this reason, in a machine learning model that determines laughter through learning, the display control device 100 may generate a model that categorizes emotions, etc., rather than a binary determination of whether or not someone is laughing.
  • the display control device 100 acquires a waveform indicating laughter based on the user's explicit designation. Specifically, the display control device 100 acquires the user's voice when the user presses an arbitrary button to cause the display control device 100 to acquire (record) laughter. At this time, the user laughs at will. This allows the display control device 100 to acquire a voiceprint pattern indicating the user's laughter (i.e., training data labeled "laughter"). Alternatively, the display control device 100 may acquire the user's voiceprint pattern so as to accumulate only the user's laughter using a general-purpose machine learning model for voice recognition that determines laughter. In general, the characteristics of laughter vary from person to person, so it is desirable for the display control device 100 to acquire and learn a voiceprint pattern for each individual user.
  • the display control device 100 may acquire voice based on various emotions of the user, not limited to laughter. Then, the display control device 100 can cause the avatar 10 to express the user's unique emotional expression by reflecting the expression based on the acquired voice in the avatar 10.
  • the display control device 100 may acquire the filler characteristics of each user and reflect the characteristics of the user who utters the filler in the avatar 10. Fillers such as “umm” and catchphrases such as "that's nice” uttered by a user can be said to indicate the characteristics of that user. For this reason, the display control device 100 learns the fillers and catchphrases of each user, and when a filler or catchphrase is observed, reflects it in the avatar 10, thereby reproducing the characteristics of that user in a virtual space. In this case, the display control device 100 can better reflect the characteristics of the user in the avatar 10 by emphasizing and cartoonish expressions, such as by expressing the sounds uttered as fillers or catchphrases in 3D using speech bubbles.
  • the display control unit 133 may display words uttered by the avatar 10 based on the input information, and may also display a history of words that transitions upward from the avatar.
  • the display control unit 133 may display the words uttered by the avatar 10 as icons that pictorially represent the meanings contained in the input information.
  • the determination unit 132 may also determine the distance between the position in the virtual space where the display of the words is observed (i.e., the position where the observer avatar is located) and the position of the avatar 10 that utters the words. In this case, the display control unit 133 displays the word history or icons as a dramatic effect that does not include the meaning of the words and icons, based on the determined distance. Note that, as will be described later, an example of a dramatic effect that does not include meaning is when no text or the like indicating the content of a conversation is displayed, and a history of lively conversation is simply displayed as a smoke signal or balloon.
  • the determination unit 132 may also determine the positional relationship between the position in the virtual space where the display of the words is observed and the position of the avatar 10 that speaks the words. In this case, the display control unit 133 may determine the angle and three-dimensional display of the display of the words observed in the virtual space based on the determined positional relationship.
  • the display control device 100 displays an icon or a conversation that includes visual effects, rather than a typical conversation displayed as text, etc. This allows the display control device 100 to display, in an easily observable manner, an active conversation taking place between avatars, etc.
  • FIG. 11 is a diagram (1) for explaining the third variation of the display control process according to the embodiment.
  • FIG. 11 shows that avatar 10A is having a conversation with the observer avatar, saying, "There's a movie I want to see!" Since avatar 10A is having the conversation in the vicinity of the observer, the conversation is displayed as text in speech bubble 230.
  • icon 231 the content of a conversation that avatar 10B is having, which is located farther away than avatar 10A, is shown by icon 231.
  • the display control device 100 performs a content and emotion analysis of the conversation of avatar 10B using the analysis server 50, and displays the analysis results as icon 231.
  • icon 231 indicates that avatar 10B had a conversation about going to the movies, then getting in a car, and then going out for dinner.
  • the display control device 100 displays the conversation content as icons (images, etc.) rather than displaying the text as is. This allows the observer to grasp the general content of the conversation and how lively it is, without having to decipher the detailed text information of distant conversations.
  • the display control device 100 can display conversations between avatars in an animation in which past history scrolls upward, as in icon 231. This allows the display control device 100 to communicate the flow of the conversation to other users in a visually easy-to-understand manner, and also allows users who are far away to understand how lively the conversation is.
  • FIG. 12 is a diagram (2) for explaining a third variation of the display control process according to the embodiment.
  • Figure 12 shows multiple avatars 235 having a conversation at a great distance from the observer.
  • the conversation between the multiple avatars 235 is displayed without displaying any semantic content such as text or icons, and only the conversation history 236 in the form of speech bubbles or circles (such displays are sometimes called "text balloons").
  • the conversation history 236 is animated and displayed upwards in the order of the most recent conversations. Therefore, in the example of FIG. 12, the conversations between multiple avatars 235 are expressed like smoke signals in the virtual space. This allows even a distant observer to know where in the virtual space an active conversation is taking place.
  • the display control device 100 can also change the expression method in various ways depending on the distance between the observer and the avatar 10. This will be explained using FIG. 13.
  • FIG. 13 is a diagram (3) for explaining a third variation of the display control process according to the embodiment.
  • Figure 13 shows how a conversation taking place near the observer (240), a conversation taking place at a medium distance from the observer (241), and a conversation taking place very far from the observer (242) are each expressed differently.
  • conversation 240 the content of the avatars' conversation is displayed in text, and the details of the conversation are displayed.
  • conversation 240 may also display image information shared between users, not just text, such as images sent by users during text chat. This allows an observer to get a rough idea of what the communication is about, without having to read the text.
  • conversation 241 the text display of the avatars' conversation is omitted, and icons and the like are inserted between the conversation so that the observer can grasp the general content of the conversation.
  • conversation 241 when two avatars are conversing, the observer can grasp the conversation content to the extent of which avatar is speaking.
  • older history in conversation 241 may be displayed in such a way that the lines of the avatars merge with each other as they move upward.
  • the display control device 100 may display the history more transparently the earlier it is (i.e., the conversation history further up), like a smoke signal in real space.
  • the display control device 100 may also display the display showing each line of dialogue smaller than conversations 240 and 241. In other words, the display control device 100 can change the display area, such as a speech bubble for displaying dialogue, depending on the distance.
  • the display control device 100 can provide the user with a UI that does not appear cluttered even when lively communication is taking place in the virtual space.
  • the display control device 100 can also change the expression method in various ways depending on the positional relationship between the observer and the avatar 10. This will be explained using FIG. 14.
  • FIG. 14 is a diagram (4) for explaining a third variation of the display control process according to the embodiment.
  • the left diagram in Figure 14 shows a conversation 250 taking place near the observer.
  • avatar 10A and avatar 10B are having a conversation.
  • conversation 250 includes a conversation history 251 of avatar 10A and a conversation history 252 of avatar 10B.
  • FIG. 14 shows an overhead view 255 of the conversation 250 as seen from above.
  • avatar 10A and avatar 10B are not facing each other as seen by an observer 256, but are standing at a slight angle.
  • the display control device 100 displays the conversation history 251 at an angle based on the positional relationship between the observer 256 and the avatar 10A (here, the angle between the line of sight of the observer 256 and the direction of the avatar 10A). Similarly, the display control device 100 displays the conversation history 252 at an angle based on the positional relationship between the observer 256 and the avatar 10B.
  • This representation allows the observer 256 to recognize the positional relationship with avatar 10A and avatar 10B simply by looking at the conversation history.
  • the display control device 100 may also perform display including dramatic effects not only on the angle but also on the content of the conversation itself, such as text. This point will be explained using FIG. 15.
  • FIG. 15 is a diagram (5) for explaining a third variation of the display control process according to the embodiment.
  • avatar 10A and avatar 10B are having a conversation, and text 260 indicating the content of the conversation, production effect 261, and text 262 are displayed.
  • the text 260 may include highlighting, such as being displayed in bold compared to other text.
  • the display control device 100 automatically highlights when the user utters a pre-registered word that strongly indicates an emotion, such as "yabai (crazy)."
  • the effect 261 is a character icon including a three-dimensional display.
  • the effect 261 is automatically displayed when the user utters a pre-registered word that strongly indicates an emotion, such as "yabai (crazy)."
  • the text 262 is also text that includes a three-dimensional display.
  • the text 262 is displayed based on a predetermined criterion, such as when the user utters a pre-registered word that strongly indicates an emotion, such as "surprise,” or when the user's voice volume is higher than normal.
  • These special text displays and special effects may be displayed based on, for example, sentiment analysis of user conversations.
  • the display control device 100 cooperates with the analysis server 50 to determine whether the content entered in the text chat is positive or negative using sentiment analysis.
  • the display control device 100 then automatically generates appropriate visual effects for the speech bubbles in which the text is displayed. Specifically, if the conversation is positive, the display control device 100 may display speech bubbles or text in bright colors or with animations of bouncing letters, or the like. Alternatively, if the conversation is negative, the display control device 100 may display speech bubbles in cool colors or with animations of crying text, or the like.
  • These speech bubbles and text may also be displayed in a history display (known as a word cloud, for example) in which the size of the text varies depending on the frequency of the words that appeared in the conversation.
  • a history display known as a word cloud, for example
  • FIG. 16 is a diagram (6) for explaining a third variation of the display control process according to the embodiment.
  • FIG. 16 shows examples of text 270 and text 275 with three-dimensional display.
  • Text 270 is configured so that the characters appear to be floating from the speech bubble. Therefore, in display 271 when text 270 is viewed from the side, the characters appear to be floating from the speech bubble.
  • the text 275 is also configured in such a way that the characters appear to be contained within a speech bubble. For this reason, when the text 275 is viewed from the side, the display 276 shows the characters as if they were contained within a cloud-like speech bubble. For this reason, to the observer, the characters appear to be displayed as two-dimensional plates, and the speech bubble appears to be displayed in a three-dimensional cloud or balloon-like shape.
  • the 3D display shown in Figure 16 etc. can also be displayed backwards. In this case, the images and text in the 3D display are displayed with the left and right reversed.
  • the display control device 100 can express text in conversations in a variety of ways, providing visual entertainment to the user.
  • the determination unit 132 may determine whether the first avatar is included within a range recognized by the second avatar. Then, when it is determined that the first avatar is included within a range recognized by the second avatar, the display control unit 133 may adjust the movement path of the first avatar so that the first avatar approaches the second avatar.
  • the determination unit 132 may also determine whether or not the first avatar has approached the second avatar in the virtual space. When it is determined that the first avatar has approached the second avatar, the display control unit 133 reflects a specific action corresponding to the approach in at least one of the first avatar and the second avatar.
  • the display control device 100 encourages interaction between users by using special displays and actions so that avatars can recognize each other's presence and communicate with each other in close proximity.
  • FIG. 17 is a diagram (1) for explaining the fourth variation of the display control process according to the embodiment.
  • the first diagram in FIG. 17 shows avatar 10D approaching a location where avatar 10A, avatar 10B, and avatar 10C are having a conversation.
  • avatar 10A recognizes that avatar 10D has entered a range where proximity is determined, either through initially set viewpoint recognition or by other means.
  • avatar 10A When avatar 10A recognizes the proximity of avatar 10D as shown in FIG. 17 (FIG. 2), it initiates a specific action. For example, avatar 10A performs a gesture of greeting by raising its hand. At this time, display control device 100 may automatically display a friendly message or the like to encourage avatar 10A to join the conversation with avatar 10D.
  • avatar 10D When avatar 10D responds to the greeting and indicates its intention to join the conversation with avatar 10A, etc., avatar 10D joins the conversation, as shown in FIG. 3 in FIG. 17.
  • the display control device 100 performs corrections so that the gaze of the avatar 10A and the avatar 10D meets, and causes the avatar 10A and the avatar 10D to perform a greeting gesture. In this way, the display control device 100 can allow the avatars to actively interact with each other.
  • the display control device 100 performs control processing based on several conditions. For example, the display control device 100 determines that one or more avatars are already in a conversation (chat) state. The display control device 100 also determines that another avatar has approached the area where the conversation is taking place. When these conditions are met, the display control device 100 corrects the display so that the avatars' viewpoints are more easily aligned, or causes them to perform gestures.
  • FIG. 18 is a diagram (2) for explaining a fourth variation of the display control process according to the embodiment.
  • the display control device 100 recognizes a proximity area 280 of avatar 10A and a proximity area 281 with avatar 10D.
  • the display control device 100 also recognizes a distance 282 between avatar 10A and avatar 10D.
  • the display control device 100 corrects the positions of avatar 10A and avatar 10D so that their lines of sight meet, for example by making a correction so that their bodies face each other.
  • the display control device 100 recognizes that the proximity area 280 of the avatar 10A and the proximity area 281 of the avatar 10D have come into contact. In this case, the display control device 100 causes the avatar 10A and the avatar 10D to make gestures to greet each other.
  • display control device 100 starts a conversation such as a voice chat between the two. At this time, display control device 100 newly recognizes conversation area 283 in which the conversation between avatar 10A and avatar 10D takes place.
  • the display control device 100 may correct the positions of their avatars so that they are not close to each other.
  • the display control device 100 may also perform control to prioritize voice chat when avatars are physically close to each other. This allows the display control device 100 to create an environment in which a newly joining avatar can easily call out to the unspecified number of avatars already participating in the conversation.
  • the display control device 100 may also prioritize text chat when the user has muted voice input.
  • FIG. 19 is a diagram (3) for explaining a fourth variation of the display control process according to the embodiment.
  • FIG. 19 shows a situation in which some kind of event is taking place in a virtual space.
  • the event involves a specific avatar appearing on stage 290.
  • the display control device 100 sets conversation-prohibited areas 292 and conversation-prohibited areas 293 in the passageways leading to stage 290 or in locations away from stage 290, so that the avatar does not remain in the passageways leading to stage 290 or in locations away from stage 290.
  • the display control device 100 sets conversation areas 294 around stage 290, etc., to actively encourage conversation around stage 290 and around monument 291.
  • the display control device 100 does not make any corrections to allow avatars to start a conversation in the conversation prohibited area 292 or the conversation prohibited area 293.
  • the display control device 100 makes corrections to allow avatars to start a conversation in the conversation area 294. This allows the display control device 100 to actively guide conversations between avatars in areas in the virtual space where conversation is encouraged.
  • the display control device 100 may correct the movement path of the avatar 10D so that the avatar 10D can easily participate in the conversation. For example, the display control device 100 may correct the walking path or walking speed so that the path along which the avatar 10D is walking automatically approaches an area close to the avatar 10A or an area where the avatars 10A and the like are having a conversation. The display control device 100 may also correct the body orientation so that the avatar 10D can easily make eye contact with the avatar 10A while walking.
  • the acquisition unit 131 acquires a request from the user to change the appearance of the avatar 10.
  • the display control unit 133 then reflects the change based on the request in the avatar 10, changing the appearance of the avatar 10 in the virtual space along with a dramatic effect.
  • the acquisition unit 131 may also acquire from the user a request to exchange any one of a plurality of objects linked to the first avatar with a second avatar. In this case, when the request is acquired, the display control unit 133 displays a list of the plurality of objects linked to the first avatar near the first avatar.
  • the acquisition unit 131 may also acquire an instruction from the user to select one of the objects displayed in a list to be exchanged for the second avatar.
  • the display control unit 133 acquires the instruction, it displays one of the objects so that it floats from the list display, and moves the floating object in the virtual space so that it overlaps with the second avatar in accordance with the operation received from the user.
  • the display control device 100 performs display control such as changing the appearance of the avatar 10 and actively encouraging the exchange of items between avatars 10. This allows the display control device 100 to provide visual entertainment for the user and actively encourage interaction between users.
  • FIG. 20 is a diagram (1) for explaining the fifth variation of the display control process according to the embodiment.
  • a UI displays candidate options 300 for changing the appearance of avatar 10A.
  • the second diagram in FIG. 20 shows the display control device 100 changing the appearance of the avatar 10A.
  • the display control device 100 may display a dramatic effect 302 in which the avatar 10A appears to be covered in smoke.
  • FIG. 3 in FIG. 20 shows the state in which the display control device 100 has changed the appearance of the avatar 10A.
  • the display control device 100 displays the changed appearance of the avatar 10A in the virtual space. In this way, the user can arbitrarily change the appearance of the avatar 10A.
  • FIG. 21 is a diagram (2) for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 21 shows the UI display when avatar 10A and an observer exchange items.
  • Avatar 10A presents exchange candidate items 310 to the observer.
  • the observer also presents exchange candidate items 311.
  • exchange button 312 When the observer selects exchange button 312, the items are exchanged.
  • the display control device 100 may perform internal processing so that such ownership is also exchanged when the items are handed over.
  • FIG. 22 is a diagram (3) for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 22 shows a state in which avatar 10A and avatar 10B are about to exchange item 320.
  • the user corresponding to avatar 10A selects item 320 on the screen display using a pointing device such as a mouse.
  • the display control device 100 enlarges the display around the item 320.
  • the user can move the item 320 closer to the avatar 10B by, for example, moving their arm in real space.
  • the display control device 100 displays a bone display 321 of the avatar 10A that corresponds to the user's arm, and moves the item 320 displayed near the arm of the avatar 10A toward the avatar 10B.
  • the display control device 100 determines that the item 320 has been transferred. The display control device 100 then displays the item 320 near the bone display 322 and transfers ownership of the item 320 to the avatar 10B.
  • the display control device 100 can complete the item exchange in a way that makes it appear as if avatars are actually handing over items to each other, rather than displaying a UI that indicates the item exchange.
  • FIG. 23 is a diagram (4) for explaining a fifth variation of the display control process according to the embodiment.
  • the avatar 10A has a head 330, a face 332, a body 334, and feet 338 set as parts of the appearance that can be changed.
  • the user can display a list of items that can be changed in each part.
  • the display control device 100 displays an item candidate 331 that corresponds to the head 330.
  • the display control device 100 displays an item candidate 333 that corresponds to the face 332, an item candidate 335 that corresponds to the body 334, and an item candidate 337 that corresponds to the feet 338.
  • the display control device 100 can also display an item candidate 339 for changing the overall appearance of the avatar 10A.
  • the display control device 100 displays a pop-up display of target item candidates for each part of the avatar 10A where a change in appearance is requested. Then, the display control device 100 selects an item for changing the appearance of the avatar 10A in accordance with the user's operation.
  • the display control device 100 may, for example, superimpose items present in a lower hierarchical level on a selected item.
  • the display control device 100 may further display multiple candidate items included in the glasses category. In this case, the user can select one item from the candidate items 333, for example, in a single stroke, and then select an item in the lower hierarchical level.
  • the display control device 100 may also set an appropriate shortcut depending on the type of device used by the user when selecting an item. For example, when selecting an item, the display control device 100 may accept input by key operation on a keyboard or the like, in addition to input by a pointing device such as a mouse or by the user's gestures.
  • the display control device 100 may also control the delivery of items using a UI display. This will be described with reference to FIG. 24.
  • FIG. 24 is a diagram (5) for explaining a fifth variation of the display control process according to the embodiment.
  • FIG. 24 shows a state in which the user selects item 341 from among item candidates 340 held by avatar 10A.
  • the user selects item 341 by pressing and holding a finger or a pointing device on the screen.
  • the display control device 100 transitions item 341 to a display in which it appears to be floating above item candidates 340 (floating display).
  • the user drags the floating item 341 to the avatar 10B.
  • the display control device 100 determines that the item 341 has been handed over to the avatar 10B.
  • the display control device 100 can perform processes such as item exchange between users using various UIs and screen displays.
  • the things exchanged between avatars in the virtual space are described as items, but the form that the items take can be arbitrarily set by the service side providing the virtual space.
  • the items may be data such as NFT (Non-Fungible Token).
  • each device in Fig. 3 conceptually shows a function in the display control system 1, and may take various forms depending on the embodiment.
  • the display control device 100 performs processing in cooperation with the analysis server 50.
  • the display control device 100 may also perform the processing performed by the analysis server 50 on its own device.
  • the display control device 100 may also be composed of multiple devices.
  • the display control device 100 may be divided into a first information processing device that acquires the user's behavior as an image and processes the acquired image, and a second information processing device that processes input of the user's voice, text, etc.
  • the first information processing device acquires the user's behavior as an image, performs posture estimation based on the image, and performs hand-raising determination, head direction determination, hand sign determination, etc.
  • the second information processing device may also perform laughter determination based on voice and emotion analysis based on text.
  • the display control device 100 may also be a device in which the display unit and the information processing unit are configured separately.
  • the information processing unit of the display control device 100 may be any information processing device such as a server or a PC.
  • each component of each device shown in the figure is a functional concept, and does not necessarily have to be physically configured as shown in the figure.
  • the specific form of distribution and integration of each device is not limited to that shown in the figure, and all or part of them can be functionally or physically distributed and integrated in any unit depending on various loads, usage conditions, etc.
  • the display control device according to the present disclosure includes an acquisition unit (the acquisition unit 131 in the embodiment), a determination unit (the determination unit 132 in the embodiment), and a display control unit (the display control unit 133 in the embodiment).
  • the acquisition unit acquires a user's behavior in the real space as input information.
  • the determination unit determines whether or not the input information includes an input corresponding to a specific action registered in advance.
  • the display control unit reflects the user's behavior corresponding to the input information in an avatar in a virtual space, and when it is determined that the input information includes an input corresponding to a specific action, the display control unit further reflects the specific action in the avatar in the virtual space and displays the avatar in the virtual space.
  • the display control device displays an avatar by combining unconscious actions that accompany the user's natural movements with conscious actions to activate gestures.
  • This allows the display control device to display an avatar in a virtual space not only in a way that directly reflects the user's movements, but also in a variety of expressions, including cartoon-like expressions, etc.
  • the acquisition unit also receives a designation from the user as to whether or not a specific action should be reflected in the avatar. If there is a designation from the user, the determination unit determines whether or not the input information includes an input corresponding to the specific action.
  • the acquisition unit also acquires the position from the imaging device when the user's behavior is captured. If the position from the imaging device is farther than a predetermined distance, the determination unit does not determine whether the input information includes an input corresponding to a specific action, and if the position from the imaging device is within the predetermined distance, the determination unit determines whether the input information includes an input corresponding to the specific action.
  • the display control device reflects the conscious action in the avatar when the user specifies that a conscious action should be performed or when the user makes a gesture in a specific area. This allows the display control device to display the avatar in accordance with the user's intention.
  • the acquisition unit also acquires the behavior of the multiple users as input information.
  • the determination unit determines the level of excitement of the interaction between the multiple users in the virtual space based on the behavior of the multiple users.
  • the display control unit displays a presentation effect in the vicinity of the multiple users according to the level of excitement of the interaction.
  • the determination unit also determines the level of excitement in the exchange based on the number of users participating in the exchange, the frequency of conversation between users, and the degree of overlap in the voices of multiple users.
  • the display control unit also determines the expression or frequency of occurrence of the dramatic effect based on the level of excitement in the interaction, the orientation of the avatar's body, and the content of the conversation between multiple users.
  • the display control device can visually display the excitement of a conversation between avatars, effectively showing even users who are not participating in the conversation how the avatars are actively interacting with each other.
  • the determination unit also determines whether or not the input information includes an input corresponding to a specific action linked to a specific emotion. If it is determined that the input information includes a specific action linked to a specific emotion, the display control unit reflects the specific action linked to the specific emotion in the avatar and displays the avatar in the virtual space.
  • the determination unit also compares the voiceprint information registered in advance by the user, which is associated with a specific emotion, with the input information to determine whether the input information includes an input corresponding to a specific action associated with the specific emotion.
  • the display control unit also displays an avatar that reflects a specific action linked to a specific emotion, and displays a dramatic effect linked to the specific emotion near the avatar.
  • the display control device can reflect gestures associated with specific emotions, such as laughter, in an avatar. This allows the display control device to express a wide variety of emotions in virtual space, just as it does in real space.
  • the display control unit also displays the words spoken by the avatar based on the input information, and displays the history of those words in an upward transition starting from the avatar.
  • the display control unit also displays the words uttered by the avatar as icons that pictorially represent the meaning contained in the input information.
  • the determination unit also determines the distance between the position in the virtual space where the display of the words is observed and the position of the avatar uttering the words. Based on the determined distance, the display control unit displays the word history or an icon as a dramatic effect that does not include the meaning of the words and icons.
  • the determination unit also determines the positional relationship between the position in the virtual space where the display of the words is observed and the position of the avatar uttering the words.
  • the display control unit determines the angle and three-dimensional display of the display of the words observed in the virtual space based on the determined positional relationship.
  • the display control device can present the conversation history in various ways, without cluttering the screen display, and allows the user to see at a glance where active communication is taking place.
  • the determination unit also determines whether the first avatar is included within a range recognized by the second avatar in the virtual space. If it is determined that the first avatar is included within a range recognized by the second avatar, the display control unit adjusts the movement path of the first avatar so that the first avatar approaches the second avatar.
  • the determination unit also determines whether or not the first avatar has approached the second avatar in the virtual space.
  • the display control unit displays the first avatar and the second avatar while reflecting a specific action corresponding to the approach of at least one of the first avatar and the second avatar.
  • the display control device can correct the behavior of avatars, thereby encouraging the avatars to actively interact with each other. This allows the display control device to increase opportunities for conversations between users in the virtual space, and stimulate interaction in the virtual space.
  • the acquisition unit also acquires a request from the user to change the appearance of the avatar.
  • the display control unit reflects the changes based on the request in the avatar, and changes the appearance of the avatar in the virtual space along with the dramatic effects.
  • the acquisition unit also acquires a request from the user to exchange any one of the multiple objects linked to the first avatar for the second avatar.
  • the display control unit displays a list of the multiple objects linked to the first avatar near the first avatar.
  • the acquisition unit also acquires an instruction from the user to select one of the objects displayed in the list to be exchanged for the second avatar.
  • the display control unit acquires the instruction, it displays one of the objects in question so that it floats from the list display, and moves the floating object in the virtual space so that it overlaps with the second avatar in accordance with the operation received from the user.
  • the display control device provides a UI with excellent operability, thereby reducing stress for users when interacting with each other, such as exchanging items. This allows the display control device to support active interaction between users.
  • Fig. 25 is a hardware configuration diagram showing an example of a computer 1000 that realizes the functions of the display control device 100.
  • the computer 1000 has a CPU 1100, a RAM 1200, a ROM (Read Only Memory) 1300, a HDD (Hard Disk Drive) 1400, a communication interface 1500, and an input/output interface 1600.
  • Each unit of the computer 1000 is connected by a bus 1050.
  • the CPU 1100 operates based on the programs stored in the ROM 1300 or the HDD 1400 and controls each component. For example, the CPU 1100 loads the programs stored in the ROM 1300 or the HDD 1400 into the RAM 1200 and executes processes corresponding to the various programs.
  • the ROM 1300 stores boot programs such as the Basic Input Output System (BIOS) that is executed by the CPU 1100 when the computer 1000 starts up, as well as programs that depend on the hardware of the computer 1000.
  • BIOS Basic Input Output System
  • HDD 1400 is a computer-readable recording medium that non-temporarily records programs executed by CPU 1100 and data used by such programs.
  • HDD 1400 is a recording medium that records a display control program related to the present disclosure, which is an example of program data 1450.
  • the communication interface 1500 is an interface for connecting the computer 1000 to an external network 1550 (e.g., the Internet).
  • the CPU 1100 receives data from other devices and transmits data generated by the CPU 1100 to other devices via the communication interface 1500.
  • the input/output interface 1600 is an interface for connecting the input/output device 1650 and the computer 1000.
  • the CPU 1100 receives data from an input device such as a keyboard or a mouse via the input/output interface 1600.
  • the CPU 1100 also transmits data to an output device such as a display, edger, or printer via the input/output interface 1600.
  • the input/output interface 1600 may also function as a media interface that reads programs and the like recorded on a specific recording medium. Examples of media include optical recording media such as DVDs (Digital Versatile Discs) and PDs (Phase change rewritable Disks), magneto-optical recording media such as MOs (Magneto-Optical Disks), tape media, magnetic recording media, and semiconductor memories.
  • the CPU 1100 of the computer 1000 executes a display control program loaded onto the RAM 1200 to realize the functions of the control unit 130, etc.
  • the display control program according to the present disclosure and data in the storage unit 120 are stored in the HDD 1400.
  • the CPU 1100 reads and executes the program data 1450 from the HDD 1400, but as another example, the CPU 1100 may obtain these programs from other devices via the external network 1550.
  • the present technology can also be configured as follows.
  • an acquisition unit that acquires user behavior in a real space as input information; a determination unit that determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance; a display control unit that reflects a behavior of the user corresponding to the input information in an avatar in a virtual space, and, when it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space;
  • a display control device comprising: (2) The acquisition unit is accepting a designation from the user as to whether or not the specific action is to be reflected in the avatar; The determination unit is determining whether or not an input corresponding to the specific action is included in the input information when the specific action is designated by the user; The display control device according to (1).
  • the acquisition unit is Acquire a position from an imaging device when the behavior of the user is imaged;
  • the determination unit is When the position from the imaging device is farther than a predetermined distance, it is not determined whether or not the input corresponding to the specific action is included in the input information, and when the position from the imaging device is within a predetermined distance, it is determined whether or not the input corresponding to the specific action is included in the input information.
  • the display control device according to (1) or (2).
  • the acquisition unit is Obtaining the behavior of multiple users as input information,
  • the determination unit is determining a degree of excitement of an interaction between the plurality of users in the virtual space based on the behavior of the plurality of users;
  • the display control unit is displaying a performance effect in the vicinity of the plurality of users according to the level of excitement of the exchange;
  • the display control device according to any one of (1) to (3).
  • the determination unit is determining a level of excitement in the exchange based on the number of users participating in the exchange, the frequency of conversations between the users, and the degree of overlap of voices between the multiple users;
  • the display control device according to (4).
  • the display control unit is determining an expression or occurrence frequency of the performance effect based on the level of excitement of the exchange, the body orientation of the avatar, and the content of the conversation between the plurality of users;
  • the display control device according to (5) above.
  • the determination unit is determining whether an input corresponding to the specific action associated with a specific emotion is included in the input information;
  • the display control unit is when it is determined that the input information includes a specific action associated with the specific emotion, reflecting the specific action associated with the specific emotion in the avatar and displaying the avatar in the virtual space;
  • a display control device according to any one of (1) to (6).
  • the determination unit is comparing the input information with voiceprint information registered in advance by the user, the voiceprint information being associated with the specific emotion, to determine whether or not the input information includes an input corresponding to a specific action associated with the specific emotion;
  • the display control device according to (7) above.
  • the display control unit is displaying the avatar reflecting a specific action associated with the specific emotion, and displaying a performance effect associated with the specific emotion in the vicinity of the avatar;
  • the display control unit is displaying words uttered by the avatar based on the input information, and displaying a history of the words in an upward transition manner starting from the avatar;
  • a display control device according to any one of (1) to (9).
  • the display control unit is displaying the words uttered by the avatar as icons pictorially representing the meanings contained in the input information;
  • the display control device according to (10).
  • the determination unit is determining a distance between a position in a virtual space where the display of the words is observed and a position of an avatar uttering the words;
  • the display control unit is displaying the history of the words or the icons as a dramatic effect not including the meaning of the words and the icons based on the determined distance;
  • the display control device according to (11).
  • the determination unit is determining a positional relationship between a position in a virtual space where the display of the words is observed and a position of an avatar uttering the words;
  • the display control unit is determining an angle and a three-dimensional representation of the representation of the words observed in a virtual space based on the determined positional relationship;
  • the display control device according to (11) or (12).
  • the determination unit is determining whether a first avatar is included within a range recognized by a second avatar in the virtual space;
  • the display control unit is When it is determined that the first avatar is included in a range recognized by the second avatar, a movement path of the first avatar is adjusted so that the first avatar approaches the second avatar;
  • the display control device according to any one of (1) to (13).
  • the determination unit is determining whether the first avatar has come close to the second avatar in the virtual space;
  • the display control unit is when it is determined that the first avatar is in proximity to the second avatar, displaying the first avatar and the second avatar by reflecting a specific action corresponding to the proximity in at least one of the first avatar and the second avatar;
  • the display control device according to (14).
  • the acquisition unit is receiving a request from the user to change an appearance of the avatar;
  • the display control unit is reflecting the change based on the request in the avatar, thereby changing the appearance of the avatar in the virtual space together with a dramatic effect;
  • the display control device according to any one of (1) to (15).
  • the acquisition unit is receiving from the user a request to exchange any one of a plurality of objects associated with a first avatar with a second avatar;
  • the display control unit is When the request is received, a list of a plurality of objects associated with the first avatar is displayed in the vicinity of the first avatar.
  • the display control device according to any one of (1) to (16).
  • the acquisition unit is obtaining, from the user, an instruction to select one of the plurality of objects displayed in the list to be exchanged for the second avatar;
  • the display control unit is When the instruction is acquired, the display device displays any one of the objects in a floating manner from the list display, and moves the floating object in a virtual space so as to overlap the second avatar in accordance with an operation received from the user.
  • the display control device according to (17).
  • the computer The user's behavior in the real world is acquired as input information, determining whether the input information includes an input corresponding to a specific action registered in advance; reflecting the behavior of the user corresponding to the input information in an avatar in a virtual space, and when it is determined that the input information includes an input corresponding to the specific action, further reflecting the specific action in an avatar in the virtual space and displaying the avatar in the virtual space;
  • a display control method comprising: (20) Computer, an acquisition unit that acquires user behavior in a real space as input information; a determination unit that determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance; a display control unit that reflects a behavior of the user corresponding to the input information in an avatar in a virtual space, and, when it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space;
  • a display control program for causing the display control device to function as a display control device having the display
  • REFERENCE SIGNS LIST 10 Avatar 50 Analysis server 100
  • Display control device 110 Communication unit 120
  • Control unit 131 Acquisition unit 132 Determination unit 133
  • Sensor unit 141 Camera 142
  • Microphone 150 Display unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Software Systems (AREA)
  • Computer Hardware Design (AREA)
  • Computer Graphics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

A display control device according to one aspect of the present disclosure comprises: an acquisition unit that acquires, as input information, a motion of a user in a real space; a determination unit that determines whether or not an input corresponding to a preregistered specific action is included in the input information; and a display control unit that reflects, in an avatar in a virtual space, the motion of the user corresponding to the input information, and that, if the input corresponding to the specific action is determined to be included in the input information, reflects the specific action in the avatar in the virtual space and displays the avatar in the virtual space.

Description

表示制御装置、表示制御方法および表示制御プログラムDISPLAY CONTROL DEVICE, DISPLAY CONTROL METHOD, AND DISPLAY CONTROL PROGRAM
 本開示は、仮想現実におけるアバター(Avatar)を表現するための表示制御装置、表示制御方法および表示制御プログラムに関する。 The present disclosure relates to a display control device, a display control method, and a display control program for representing an avatar in virtual reality.
 メタバース(Meta-verse)等と称される、コンピュータやコンピュータネットワークの中に構築された3次元仮想空間をユーザに呈示する様々な技術が実用化されている。 Various technologies have been put into practical use that present users with a three-dimensional virtual space constructed within a computer or computer network, known as the Metaverse.
 例えば、仮想空間における表現にVR(Virtual Reality)やAR(Augmented Reality)に関する技術を応用することで、操作性に優れたアバターをユーザに提供したり、現実に即した背景表示を可能としたりする技術が提案されている(例えば、特許文献1)。 For example, technology has been proposed that applies VR (Virtual Reality) and AR (Augmented Reality) technologies to expressions in virtual space, providing users with avatars that are easy to operate and enabling backgrounds to be displayed that are in line with reality (e.g., Patent Document 1).
国際公開第2014/119098号International Publication No. 2014/119098
 しかし、仮想空間においてアバターを表現する手法については、さらに改善の余地がある。 However, there is still room for improvement in the methods used to represent avatars in virtual spaces.
 例えば、仮想空間では、アバターを用いてユーザ同士がコミュニケーションを親密にとることが可能である。このため、仮想空間を提供するサービス側は、現実に即したリアルな動きをアバターに再現させるだけでなく、ユーザが親しみを抱くような、様々な感情表現や多彩な動きを表現するアバターを提供することが望ましい。 For example, in a virtual space, users can communicate intimately with each other using avatars. For this reason, it is desirable for the service provider of the virtual space to not only have the avatar reproduce realistic movements that correspond to reality, but also to provide avatars that can express a variety of emotions and movements that users can feel close to.
 そこで、本開示では、仮想空間のアバターを多様に表現することのできる表示制御装置、表示制御方法および表示制御プログラムを提案する。 This disclosure therefore proposes a display control device, a display control method, and a display control program that can represent avatars in a virtual space in a variety of ways.
 上記の課題を解決するために、本開示に係る一形態の表示制御装置は、実空間におけるユーザの挙動を入力情報として取得する取得部と、予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する判定部と、前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する表示制御部と、を備える。 In order to solve the above problems, a display control device according to one embodiment of the present disclosure includes an acquisition unit that acquires user behavior in real space as input information, a determination unit that determines whether the input information includes an input corresponding to a specific action registered in advance, and a display control unit that reflects the user behavior corresponding to the input information in an avatar in a virtual space, and, if it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space.
実施形態に係る表示制御処理の概要を示す図である。FIG. 2 is a diagram showing an overview of a display control process according to the embodiment. 実施形態に係る判定処理の一例を示す図である。FIG. 11 is a diagram illustrating an example of a determination process according to the embodiment. 実施形態に係る表示制御装置の構成例を示す図である。FIG. 1 is a diagram illustrating an example of the configuration of a display control device according to an embodiment. 実施形態に係るジェスチャー記憶部の一例を示す図である。FIG. 4 is a diagram illustrating an example of a gesture storage unit according to the embodiment. 実施形態に係る表示制御処理の流れを示すフローチャートである。5 is a flowchart showing a flow of a display control process according to the embodiment. 実施形態に係る表示制御処理の第1のバリエーションを説明するための図(1)である。FIG. 11 is a diagram (1) for explaining a first variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第1のバリエーションを説明するための図(2)である。FIG. 11 is a diagram (2) for explaining a first variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第2のバリエーションを説明するための図(1)である。FIG. 11 is a diagram (1) for explaining a second variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第2のバリエーションを説明するための図(2)である。FIG. 11 is a diagram (2) for explaining a second variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第2のバリエーションを説明するための図(3)である。FIG. 11 is a diagram (3) for explaining a second variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第3のバリエーションを説明するための図(1)である。FIG. 11 is a diagram for explaining a third variation of the display control process according to the embodiment; 実施形態に係る表示制御処理の第3のバリエーションを説明するための図(2)である。FIG. 11 is a diagram (2) for explaining a third variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第3のバリエーションを説明するための図(3)である。FIG. 11 is a diagram (3) for explaining a third variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第3のバリエーションを説明するための図(4)である。FIG. 4 is a diagram for explaining a third variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第3のバリエーションを説明するための図(5)である。FIG. 5 is a diagram for explaining a third variation of the display control process according to the embodiment; 実施形態に係る表示制御処理の第3のバリエーションを説明するための図(6)である。FIG. 6 is a diagram for explaining a third variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第4のバリエーションを説明するための図(1)である。FIG. 11 is a diagram for explaining a fourth variation of the display control process according to the embodiment; 実施形態に係る表示制御処理の第4のバリエーションを説明するための図(2)である。FIG. 13 is a diagram (2) for explaining a fourth variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第4のバリエーションを説明するための図(3)である。FIG. 11 is a diagram (3) for explaining a fourth variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第5のバリエーションを説明するための図(1)である。FIG. 11 is a diagram for explaining a fifth variation of the display control process according to the embodiment; 実施形態に係る表示制御処理の第5のバリエーションを説明するための図(2)である。FIG. 13 is a diagram (2) for explaining a fifth variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第5のバリエーションを説明するための図(3)である。FIG. 11 is a diagram (3) for explaining a fifth variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第5のバリエーションを説明するための図(4)である。FIG. 4 is a diagram for explaining a fifth variation of the display control process according to the embodiment. 実施形態に係る表示制御処理の第5のバリエーションを説明するための図(5)である。FIG. 5 is a diagram (5) for explaining a fifth variation of the display control process according to the embodiment. 表示制御装置の機能を実現するコンピュータの一例を示すハードウェア構成図である。FIG. 2 is a hardware configuration diagram illustrating an example of a computer that realizes the functions of the display control device.
 以下に、実施形態について図面に基づいて詳細に説明する。なお、以下の各実施形態において、同一の部位には同一の符号を付することにより重複する説明を省略する。 The following describes the embodiments in detail with reference to the drawings. Note that in each of the following embodiments, the same parts are designated by the same reference numerals, and duplicate descriptions will be omitted.
 以下に示す項目順序に従って本開示を説明する。
  1.実施形態
   1-1.実施形態に係る表示制御処理の概要
   1-2.実施形態に係る表示制御装置の構成
   1-3.実施形態に係る処理の手順
   1-4.実施形態に係る処理のバリエーション
    1-4-1.アバター同士の盛り上がり表現
    1-4-2.アバターの感情表現
    1-4-3.会話の可視化表現
    1-4-4.アバター同士の近接効果
    1-4-5.アバターの表示や交流に関する表現
   1-5.変形例
  2.その他の実施形態
  3.本開示に係る表示制御装置の効果
  4.ハードウェア構成
The present disclosure will be described in the following order.
1. Embodiment 1-1. Overview of display control processing according to embodiment 1-2. Configuration of display control device according to embodiment 1-3. Processing procedure according to embodiment 1-4. Variations of processing according to embodiment 1-4-1. Expression of excitement between avatars 1-4-2. Expression of avatar emotions 1-4-3. Visualization of conversation 1-4-4. Proximity effect between avatars 1-4-5. Expression related to display and interaction of avatars 1-5. Modifications 2. Other embodiments 3. Effects of a display control device according to the present disclosure 4. Hardware configuration
(1.実施形態)
(1-1.実施形態に係る表示制御処理の概要)
 図1を用いて、実施形態に係る表示制御処理の一例を説明する。図1は、実施形態に係る表示制御処理の概要を示す図である。
1. EMBODIMENTS
(1-1. Overview of Display Control Process According to the Embodiment)
An example of a display control process according to the embodiment will be described with reference to Fig. 1. Fig. 1 is a diagram showing an overview of the display control process according to the embodiment.
 実施形態に係る表示制御処理は、図1に図示しない表示制御装置100によって実行される。表示制御装置100は、VRやAR技術を実現するための情報処理端末である。すなわち、実施形態では、表示制御装置100は、メタバース(Meta-verse)等と称される、コンピュータやコンピュータネットワークの中に構築された3次元仮想空間(以下、単に「仮想空間」と称する)を表示部(ディスプレイ等)に表示する。また、表示制御装置100は、仮想空間において、ユーザを模したキャラクターであるアバターを表示する。例えば、表示制御装置100は、PC(Personal Computer)やタブレット端末、スマートフォンなど、任意の情報処理装置によって実現される。なお、表示制御装置100は、例えばヘッドマウントディスプレイ(HMD, Head Mounted Display)やARグラス等であってもよい。 The display control process according to the embodiment is executed by a display control device 100, not shown in FIG. 1. The display control device 100 is an information processing terminal for implementing VR and AR technology. That is, in the embodiment, the display control device 100 displays a three-dimensional virtual space (hereinafter simply referred to as "virtual space") constructed in a computer or computer network, such as a metaverse, on a display unit (display, etc.). The display control device 100 also displays an avatar, which is a character that resembles a user, in the virtual space. For example, the display control device 100 is implemented by any information processing device, such as a PC (Personal Computer), a tablet terminal, or a smartphone. The display control device 100 may be, for example, a head mounted display (HMD) or AR glasses.
 例えば、表示制御装置100は、イメージセンサ(カメラ等)で実空間のユーザを撮像し、当該ユーザの身振りや表情を仮想空間のアバターに反映する。これにより、ユーザは、仮想空間にあたかも自身の分身が存在しているかのような感覚を体験することができる。また、仮想空間には複数のアバター(言い換えれば、複数のユーザ)が同時に存在することができるため、ユーザは、アバターを介して、実空間と同じようなコミュニケーションをとることができる。例えば、ユーザは、アバターを介して、他のユーザに挨拶をしたり、会話(ボイスチャットあるいはテキストチャット)したりして、他のユーザとの交流を深めることができる。 For example, the display control device 100 captures an image of a user in real space with an image sensor (such as a camera) and reflects the user's gestures and facial expressions in an avatar in the virtual space. This allows the user to experience the sensation of having an alter-ego of himself or herself present in the virtual space. Furthermore, since multiple avatars (in other words, multiple users) can exist simultaneously in the virtual space, users can communicate via the avatars in the same way as in the real space. For example, users can deepen their interactions with other users by greeting them and conversing (voice chat or text chat) via their avatars.
 仮想空間は様々な態様で表現することができるため、実空間に即したリアルな空間のみならず、ユーザが親しみを抱くような、漫画的表現を伴う空間として構築可能である。この場合、仮想空間の提供側も、アバターの表現として、現実に即したリアルな動きをアバターに再現させるだけでなく、ユーザが親しみを抱くような、様々な感情表現や多彩な動きを提供できることが望ましい。 Because virtual spaces can be expressed in a variety of ways, they can be constructed not only as realistic spaces based on real space, but also as spaces with cartoon-like expressions that users can relate to. In this case, it is also desirable for the provider of the virtual space to not only be able to make the avatar reproduce realistic movements based on reality, but also to be able to provide a variety of emotional expressions and diverse movements that users can relate to.
 そこで、本開示に係る表示制御装置100は、以下に示す構成により、上記処理を実現する。例えば、表示制御装置100は、実空間におけるユーザの挙動を入力情報として取得するとともに、予め登録された特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する。そして、表示制御装置100は、入力情報に対応するユーザの挙動を仮想空間のアバターに反映するとともに、入力情報に特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映する。 The display control device 100 according to the present disclosure realizes the above processing by the configuration shown below. For example, the display control device 100 acquires user behavior in real space as input information, and determines whether the input information includes an input corresponding to a specific action registered in advance. The display control device 100 then reflects the user behavior corresponding to the input information in an avatar in a virtual space, and if it is determined that the input information includes an input corresponding to a specific action, it further reflects the specific action in the avatar in the virtual space.
 具体的には、表示制御装置100は、アバターが行う特定の動作(以下、「ジェスチャー」と称する)として、サムアップや、アバター各自の決めポーズや、ハイタッチなど、漫画的表現を含むコミカルな動きを予め登録する。そして、表示制御装置100は、ユーザから当該ジェスチャーを発火させるための動作を入力として受け付けた場合、ユーザの動きそのものではなく、当該ジェスチャーをアバターに反映する。なお、表示制御装置100は、ジェスチャーをユーザから受け付けない場合、ユーザの身振りそのものをアバターに反映してもよい。このように、表示制御装置100は、ユーザの意識的な動作に伴うジェスチャーと、ユーザの無意識的な動きとのいずれをもアバターに反映する。これにより、表示制御装置100は、単にユーザの動きを模しただけでなく、表現豊かな、コミカルな動きを伴うアバターを仮想空間上で表示することができる。このような表現を通じて、表示制御装置100は、ユーザ同士の意思の疎通を図りやすくしたり、互いのユーザが親しみを覚えやすくさせたりする仮想空間を構築できる。 Specifically, the display control device 100 registers in advance comical movements including cartoon expressions, such as a thumbs-up, a signature pose of each avatar, and a high-five, as specific movements (hereinafter referred to as "gestures") to be performed by the avatar. When the display control device 100 receives an input from the user of an action for activating the gesture, the display control device 100 reflects the gesture, rather than the user's movement itself, in the avatar. Note that when the display control device 100 does not receive a gesture from the user, the display control device 100 may reflect the user's gesture itself in the avatar. In this way, the display control device 100 reflects both gestures associated with the user's conscious movements and the user's unconscious movements in the avatar. This allows the display control device 100 to display an avatar in a virtual space that not only imitates the user's movements, but also has expressive and comical movements. Through such expressions, the display control device 100 can build a virtual space that makes it easier for users to communicate with each other and to feel familiar with each other.
 以下、図1を用いて、上記の表示制御処理について説明する。図1の左部に示すように、表示制御装置100は、表示画面例20において、仮想空間に存在するアバター10Aを表示している。このとき、表示制御装置100は、カメラを用いて、アバター10Aに対応するユーザ(すなわち、アバター10Aに動きや表情が反映されるユーザ)を常時撮像する。 The above display control process will be explained below with reference to FIG. 1. As shown in the left part of FIG. 1, the display control device 100 displays an avatar 10A that exists in a virtual space on an example display screen 20. At this time, the display control device 100 uses a camera to constantly capture images of the user corresponding to the avatar 10A (i.e., the user whose movements and expressions are reflected in the avatar 10A).
 そして、表示制御装置100は、カメラから得られる画像データに基づいて、ユーザから入力される入力情報を取得する。なお、入力情報とは、カメラを含む様々なセンサから取得される情報であり、アバターの表示等に利用される情報を総称するものである。 Then, the display control device 100 acquires input information input by the user based on image data obtained from the camera. Note that input information is information acquired from various sensors including the camera, and is a general term for information used for displaying avatars, etc.
 一例として、表示制御装置100は、入力情報として、画像に基づく表情抽出の技術を用いてユーザの表情21を取得する。また、表示制御装置100は、画像に基づく骨格推定の技術を用いて、ユーザの骨格22を取得する。また、表示制御装置100は、ハンドトラッキング技術を用いて、ユーザの腕や手の動きを示した身振り情報23を取得する。なお、表示制御装置100は、画像データのみならず、ユーザから入力される音声やテキストの情報を入力情報として取得してもよい。そして、表示制御装置100は、これらの情報を総合して、アバター10Aに反映するユーザからの入力情報を受け付ける。 As an example, the display control device 100 acquires the user's facial expression 21 as input information using an image-based facial expression extraction technique. The display control device 100 also acquires the user's skeleton 22 using an image-based skeleton estimation technique. The display control device 100 also acquires gesture information 23 indicating the movement of the user's arms and hands using hand tracking technique. Note that the display control device 100 may acquire not only image data, but also voice or text information input by the user as input information. The display control device 100 then combines this information to accept input information from the user to be reflected in the avatar 10A.
 例えば、表示制御装置100は、ユーザの入力情報に基づいて、アバター10Aがうなずいたり、首を振って拒絶の意図を示したり、首をかしげて悩むしぐさ等を、特定の動作でない、ユーザの無意識的な動作として取得する。なお、表示制御装置100は、ユーザの笑う表情や表現、驚く表情や表現など、ユーザによる様々な感動表現を無意識的な動作として取得してもよい。本開示では、これらを総称して、「無意識的動作」と称する。図1の左部の例では、ユーザが手を挙げた無意識的動作が、アバター10Aに反映されて仮想空間に表示されている例を示している。 For example, based on the user's input information, the display control device 100 acquires the avatar 10A nodding, shaking its head to indicate refusal, tilting its head in distress, etc. as unconscious actions of the user that are not specific actions. The display control device 100 may also acquire various emotional expressions by the user, such as a laughing expression or expression, or a surprised expression or expression, as unconscious actions. In this disclosure, these are collectively referred to as "unconscious actions." The example on the left side of Figure 1 shows an example in which the unconscious action of the user raising his hand is reflected in the avatar 10A and displayed in the virtual space.
 一方、表示制御装置100は、取得した画像データに基づいて、ユーザがジェスチャーを発火させるための特定の動作を行っているか否かを判定することもできる。すなわち、表示制御装置100は、予め登録された特定の動作に対応する入力が入力情報に含まれるか否かを判定する。例えば、図1の右部に示すように、表示制御装置100は、画面表示例25において、仮想空間に存在するアバター10Bを表示している。このとき、表示制御装置100は、カメラを用いて、アバター10Bに対応するユーザ(すなわち、アバター10Bに動きや表情が反映されるユーザ)を常時撮像する。 On the other hand, the display control device 100 can also determine, based on the acquired image data, whether or not the user is performing a specific action to trigger a gesture. That is, the display control device 100 determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance. For example, as shown in the right part of FIG. 1, the display control device 100 displays an avatar 10B that exists in a virtual space in a screen display example 25. At this time, the display control device 100 uses a camera to constantly capture images of the user corresponding to the avatar 10B (i.e., the user whose movements and expressions are reflected in the avatar 10B).
 なお、図1の例では、対比をわかりやすくするため、アバター10Aおよびアバター10Bを別々に描いているが、実際には、無意識的動作および意識的動作は、同一のアバターにおいて表現される。以下、アバター10Aやアバター10B等を区別する必要のない場合、単に「アバター10」と総称する。 In the example of Figure 1, avatar 10A and avatar 10B are drawn separately to make the comparison easier to understand, but in reality, unconscious actions and conscious actions are expressed by the same avatar. Hereinafter, when there is no need to distinguish between avatar 10A, avatar 10B, etc., they will be collectively referred to simply as "avatar 10."
 表示制御装置100は、ユーザが特定の動作を行う場合、その特定の動作を判定するための入力モードへの移行要求を受け付ける。例えば、ユーザによる特定の動作に基づき、アバター10Bが特徴的なジェスチャーを発することを所望する場合、ユーザは、意識的にジェスチャーを行わせるための入力モードへの移行を表示制御装置100に要求する。 When the user performs a specific action, the display control device 100 accepts a request to transition to an input mode for determining the specific action. For example, when the user desires that the avatar 10B make a distinctive gesture based on a specific action by the user, the user requests the display control device 100 to transition to an input mode for consciously making the gesture.
 その後、ユーザは、「カメラの前で人差し指と親指で予め登録しておいたポーズを決める」など、特徴的な動きを行う。表示制御装置100は、図1の左部と同様、撮像された画像に基づいて、ユーザの表情26や、ユーザの骨格27や、身振り情報28など、ユーザの挙動を取得する。そして、表示制御装置100は、取得した情報を総合して、かかるユーザの挙動の特徴が、予め登録されていたジェスチャーを発火させるための特徴と一致するか否かを判定する。なお、ユーザの挙動の取得やジェスチャーの判定処理に関しては、様々な既知の技術が利用されてもよい。 Then, the user makes a characteristic movement, such as "taking a pre-registered pose with his index finger and thumb in front of the camera." Similar to the left part of FIG. 1, the display control device 100 acquires the user's behavior, such as the user's facial expression 26, the user's skeletal structure 27, and gesture information 28, based on the captured image. The display control device 100 then combines the acquired information to determine whether or not the characteristics of the user's behavior match the characteristics for activating a pre-registered gesture. Note that various known technologies may be used for acquiring the user's behavior and determining the gesture.
 表示制御装置100は、ユーザから入力された挙動が、ジェスチャーを発火させるだけの特徴を有する動きと判定すると、当該ジェスチャーをアバター10Bに反映させる。例えば、図1の右部に示すように、アバター10Bは、人差し指と親指でポーズを作る。このとき、表示制御装置100は、ポーズとともに、アバター10Bの周囲に放射線状の線が表示されるような演出効果を含む、コミカルな表現を行ってもよい。かかる表現により、アバター10Bの動作が、事前に登録された特定のジェスチャーであるか、ユーザの動きそのものであるかをユーザが区別することができる。すなわち、このような特定のジェスチャーは、ユーザが意識的にアバター10Bに実行させるものである。このため、本開示では、これを総称して「意識的動作」と称する。 When the display control device 100 determines that the behavior input by the user is a movement that has enough characteristics to trigger a gesture, it reflects the gesture in the avatar 10B. For example, as shown in the right part of FIG. 1, the avatar 10B makes a pose with its index finger and thumb. At this time, the display control device 100 may perform a comical expression that includes a dramatic effect such as displaying radial lines around the avatar 10B along with the pose. This expression allows the user to distinguish whether the movement of the avatar 10B is a specific gesture registered in advance or the user's movement itself. In other words, such specific gestures are consciously made to be performed by the avatar 10B by the user. For this reason, in this disclosure, these are collectively referred to as "conscious movements."
 そして、表示制御装置100は、これらの無意識的動作および意識的動作のいずれをもアバター10に反映させ、アバター10のリアクションとして表現することができる。これにより、表示制御装置100は、ユーザの動きをそのまま反映させるだけでなく、漫画的表現等も伴った、多様な表現で仮想空間のアバター10を表示することができる。 Then, the display control device 100 can reflect both these unconscious and conscious actions in the avatar 10 and express them as reactions of the avatar 10. This allows the display control device 100 to display the avatar 10 in the virtual space in a variety of expressions, including not only a direct reflection of the user's movements, but also cartoon-like expressions, etc.
 ところで、上述のように、ユーザは、無意識的動作および意識的動作の入力モードの切り替えを行い、アバター10に無意識的動作もしくは意識的動作による入力情報を反映させることができる。かかる切り替えは、例えばユーザによる明示の指定(ボタン押下等の手動操作や音声入力等)に基づき行われてもよい。さらに、表示制御装置100は、ユーザが意識的動作を行う空間上の範囲を予め設定しておき、かかる設定に基づいて、入力モードの切り替えを行うこともできる。この点について、図2を用いて説明する。図2は、実施形態に係る判定処理の一例を示す図である。 As described above, the user can switch between the input modes of unconscious actions and conscious actions, and reflect input information from unconscious actions or conscious actions in the avatar 10. Such switching may be performed, for example, based on an explicit designation by the user (manual operation such as pressing a button, voice input, etc.). Furthermore, the display control device 100 can set in advance a spatial range within which the user will perform conscious actions, and switch the input mode based on such a setting. This point will be explained using FIG. 2. FIG. 2 is a diagram showing an example of the determination process according to the embodiment.
 図2では、ユーザが、表示制御装置100のカメラ141から撮像されることで、自身の挙動を入力している例を示している。このとき、ユーザは、ハンドサイン30を、予め設定された空間31(例えば、カメラ141から所定距離内の空間)で行う。表示制御装置100は、例えば、画像解析や、デプスセンサ等を用いて、ユーザの挙動が空間31で行われたものか否かを判定する。そして、表示制御装置100は、ユーザの挙動が空間31で行われた場合、かかる挙動は、意識的動作を発火させるものとみなして、ジェスチャーの発火を判定する。一方、表示制御装置100は、空間31以外の範囲である空間32において観測されるユーザの挙動は、無意識的動作としてアバター10に反映させる。 FIG. 2 shows an example in which a user inputs his/her own behavior by being imaged by the camera 141 of the display control device 100. At this time, the user makes hand signs 30 in a pre-set space 31 (e.g., a space within a predetermined distance from the camera 141). The display control device 100 determines whether the user's behavior was performed in the space 31, for example, using image analysis or a depth sensor. If the user's behavior was performed in the space 31, the display control device 100 considers the behavior to be one that ignites a conscious action and determines that a gesture has been ignited. On the other hand, the display control device 100 reflects the user's behavior observed in space 32, which is a range outside the space 31, in the avatar 10 as an unconscious action.
 すなわち、表示制御装置100は、カメラ141からの位置が所定距離より離れている場合には、ユーザからの入力を、ジェスチャーを発火させるための動作と判定せず、空間31で行われたユーザの動作のみについて、ジェスチャーを発火させるための動作か否か判定する。これにより、ユーザは、入力モードの切り替えについて、明示の指定を行わなくても、無意識的動作および意識的動作を任意にアバター10に反映させることができる。 In other words, when the display control device 100 is located at a predetermined distance from the camera 141, it does not determine that the input from the user is an action for igniting a gesture, but only determines whether the user's actions performed in the space 31 are actions for igniting a gesture. This allows the user to reflect unconscious and conscious actions in the avatar 10 at will, without having to explicitly specify the switching of input modes.
 なお、入力モードの切り替えは、図2で示した例に限られない。例えば、表示制御装置100は、ユーザがボイスチャット(すなわち音声入力)を行っている場合、自動的に無意識的動作の抽出(ボーン検出)の入力モードに切り替えてもよい。あるいは、表示制御装置100は、ユーザがテキストチャット(すなわち手動による文字入力)を行っている場合、自動的に意識的動作の抽出(ジェスチャー検出)の入力モードに切り替えてもよい。 Note that the switching of input modes is not limited to the example shown in FIG. 2. For example, when a user is engaged in voice chat (i.e., voice input), the display control device 100 may automatically switch to an input mode for extracting unconscious actions (bone detection). Alternatively, when a user is engaged in text chat (i.e., manual character input), the display control device 100 may automatically switch to an input mode for extracting conscious actions (gesture detection).
 また、表示制御装置100は、モード切替を行っている途中においては、双方の動きを反映させたモーションブレンドを行ってもよい。また、表示制御装置100は、ユーザがUI(User Interface)操作を行っている途中は、入力モードそのものをオフにして、異なる意図の動きがアバター10に反映されないようにしてもよい。あるいは、表示制御装置100は、任意のキー操作を行ってから所定時間内のみ、ジェスチャー入力モードを発動させてもよい。また、表示制御装置100は、ジェスチャーを行う際にはユーザから予め決められた言葉の入力を要求することで、誤ったジェスチャー入力が行われないようにしてもよい。また、表示制御装置100は、ユーザからの音声入力をオフにするミュートボタンが押下された際にジェスチャー入力モードを発動し、その他の場合には無意識的動作の入力モードに設定しておいてもよい。また、表示制御装置100は、無意識的動作の入力モードがオフになっているときには、同時に音声入力をオフにして、プライバシーモード(ユーザから入力を受け付けない)に移行してもよい。このように、表示制御装置100は、様々な態様に対応して、入力モードを使い分けてもよい。 In addition, the display control device 100 may perform motion blending that reflects both movements while switching modes. In addition, the display control device 100 may turn off the input mode itself while the user is performing a UI (User Interface) operation, so that movements of different intentions are not reflected in the avatar 10. Alternatively, the display control device 100 may activate the gesture input mode only within a predetermined time after performing an arbitrary key operation. In addition, the display control device 100 may prevent erroneous gesture input by requesting the user to input a predetermined word when performing a gesture. In addition, the display control device 100 may activate the gesture input mode when a mute button that turns off voice input from the user is pressed, and may set the display control device 100 to the unconscious action input mode in other cases. In addition, when the unconscious action input mode is turned off, the display control device 100 may simultaneously turn off the voice input and transition to a privacy mode (wherein input from the user is not accepted). In this way, the display control device 100 may use different input modes in response to various aspects.
 以上、図1乃至図2を用いて説明したように、実施形態に係る表示制御処理によれば、無意識的動作および意識的動作を任意にアバター10に反映させることができるので、多様な表現でアバター10を表示することができる。 As described above with reference to Figures 1 and 2, the display control process according to the embodiment can reflect unconscious and conscious actions in the avatar 10 as desired, allowing the avatar 10 to be displayed in a variety of ways.
(1-2.実施形態に係る表示制御装置の構成)
 次に、表示制御装置100の構成について説明する。図3は、実施形態に係る表示制御装置100の構成例を示す図である。なお、表示制御装置100は、図3に示す実施形態に係る表示制御システム1の構成要素の1つである。表示制御システム1は、表示制御装置100と、分析サーバ50とを含む。
(1-2. Configuration of the display control device according to the embodiment)
Next, a configuration of the display control device 100 will be described. Fig. 3 is a diagram showing an example of the configuration of the display control device 100 according to an embodiment. The display control device 100 is one of the components of a display control system 1 according to the embodiment shown in Fig. 3. The display control system 1 includes the display control device 100 and an analysis server 50.
 分析サーバ50は、後述する各種分析処理を行い、分析結果を表示制御装置100に提供する。分析サーバ50は、例えば、表示制御装置100が実行しない、テキストに基づく感情分析や、画像認識に対する機械学習処理等、比較的処理負荷の高い情報処理を行う。例えば、分析サーバ50は、仮想空間に係るサービスを運営する事業者等によって管理されるクラウドサーバ等である。 The analysis server 50 performs various analysis processes described below and provides the analysis results to the display control device 100. The analysis server 50 performs information processing with a relatively high processing load, such as text-based sentiment analysis and machine learning processing for image recognition, which are not performed by the display control device 100. For example, the analysis server 50 is a cloud server managed by a business operator that operates a service related to virtual space.
 図3に示すように、表示制御装置100は、通信部110と、記憶部120と、制御部130と、センサ部140と、表示部150を有する。なお、表示制御装置100は、表示制御装置100を操作するユーザから各種操作を受け付ける入力部(キーボードやタッチパネル等)を有してもよい。 As shown in FIG. 3, the display control device 100 has a communication unit 110, a memory unit 120, a control unit 130, a sensor unit 140, and a display unit 150. The display control device 100 may also have an input unit (such as a keyboard or a touch panel) that accepts various operations from a user who operates the display control device 100.
 通信部110は、例えば、NIC(Network Interface Card)やネットワークインタフェイスコントローラ(Network Interface Controller)等によって実現される。通信部110は、ネットワークNと有線または無線で接続され、ネットワークNを介して、分析サーバ50等と情報の送受信を行う。ネットワークNは、例えば、Bluetooth(登録商標)、インターネット、Wi-Fi(登録商標)、UWB(Ultra Wide Band)、LPWA(Low Power Wide Area)等の無線通信規格もしくは方式で実現される。 The communication unit 110 is realized, for example, by a NIC (Network Interface Card) or a network interface controller. The communication unit 110 is connected to the network N by wired or wireless means, and transmits and receives information to and from the analysis server 50, etc., via the network N. The network N is realized, for example, by a wireless communication standard or method such as Bluetooth (registered trademark), the Internet, Wi-Fi (registered trademark), UWB (Ultra Wide Band), or LPWA (Low Power Wide Area).
 記憶部120は、例えば、RAM(Random Access Memory)、フラッシュメモリ(Flash Memory)等の半導体メモリ素子、または、ハードディスク、光ディスク等の記憶装置によって実現される。 The storage unit 120 is realized, for example, by a semiconductor memory element such as a random access memory (RAM) or a flash memory, or a storage device such as a hard disk or an optical disk.
 記憶部120は、実施形態に係る表示制御処理に関する種々の情報を記憶する。実施形態では、記憶部120は、ジェスチャー記憶部121を含む。 The storage unit 120 stores various information related to the display control process according to the embodiment. In the embodiment, the storage unit 120 includes a gesture storage unit 121.
 図4を用いて、ジェスチャー記憶部121について説明する。図4は、実施形態に係るジェスチャー記憶部121の一例を示す図である。図4に示すように、ジェスチャー記憶部121は、「ジェスチャーID」、「ジェスチャー」、「判定項目」といった項目を有する。 The gesture storage unit 121 will be described with reference to FIG. 4. FIG. 4 is a diagram showing an example of the gesture storage unit 121 according to an embodiment. As shown in FIG. 4, the gesture storage unit 121 has items such as "gesture ID," "gesture," and "determination item."
 「ジェスチャーID」は、ジェスチャーを識別するための識別情報である。「ジェスチャー」は、具体的なジェスチャーの内容を示す。「判定項目」は、ユーザの挙動が当該ジェスチャーを発火させるものであるか否かを判定する際に用いられる特徴量を抽出する事項を示す。なお、図4での図示は省略するが、各ジェスチャーの情報には、アバター10の3Dモデルをジェスチャーに即して動かすためのモーションデータ(骨格や表情のアニメーションや演出効果等)を含んでもよい。 "Gesture ID" is identification information for identifying a gesture. "Gesture" indicates the specific content of the gesture. "Determination item" indicates the item for extracting a feature amount used when determining whether or not the user's behavior will trigger the gesture. Although not shown in FIG. 4, the information on each gesture may include motion data (such as animation of the skeleton and facial expressions, dramatic effects, etc.) for moving the 3D model of the avatar 10 in accordance with the gesture.
 例えば、図4に示した一例は、ジェスチャーIDが「A01」のジェスチャーが「サムアップ」に対応し、そのジェスチャーを発火させるか否かは、ユーザの「手の位置」や「腕の動き」で判定される、ということを示している。 For example, the example shown in Figure 4 indicates that the gesture with gesture ID "A01" corresponds to a "thumbs up," and whether or not to fire that gesture is determined by the user's "hand position" and "arm movement."
 図3に戻って説明を続ける。センサ部140は、種々の環境情報を検知するセンサである。例えば、センサ部140は、表示制御装置100の外部を撮影する外向きカメラ、ユーザ側を撮影する内向きカメラ等から構成される、撮像装置であるカメラ141を含む。 Referring back to FIG. 3, the explanation will continue. The sensor unit 140 is a sensor that detects various environmental information. For example, the sensor unit 140 includes a camera 141, which is an imaging device that is composed of an outward-facing camera that captures images of the outside of the display control device 100, an inward-facing camera that captures images of the user, etc.
 例えば、カメラ141は、表示制御装置100の前方に位置する被写体(例えば、実空間に所在するユーザ)を認識する。この場合、カメラ141は、被写体であるユーザの画像を取得するとともに、ステレオカメラで撮像された画像間の視差や、撮像された画像データの解析に基づいて、表示制御装置100(言い換えればカメラ141)から被写体までの距離を算出できる。あるいは、センサ部140は、ユーザ等、任意の被写体までの距離を検知可能なデプスセンサを用いて、実空間における距離を検出してもよい。 For example, the camera 141 recognizes a subject (e.g., a user located in real space) located in front of the display control device 100. In this case, the camera 141 acquires an image of the user who is the subject, and can calculate the distance from the display control device 100 (in other words, the camera 141) to the subject based on the parallax between images captured by the stereo camera and an analysis of the captured image data. Alternatively, the sensor unit 140 may detect the distance in real space using a depth sensor capable of detecting the distance to any subject, such as a user.
 また、センサ部140は、ユーザの声や周囲の音を測定可能なマイクロホン142を含む。例えば、ユーザは、ボイスチャットを行う場合、表示制御装置100が備えたマイクロホン142を利用することで、音声入力を行うことができる。 The sensor unit 140 also includes a microphone 142 that can measure the user's voice and surrounding sounds. For example, when engaging in voice chat, the user can input voice by using the microphone 142 provided in the display control device 100.
 また、例えば表示制御装置100がHMD等のウェアラブルデバイスである場合、センサ部140は、カメラ141やマイクロホン142のほか、ユーザの身体の向き、傾き、動きや移動速度等、ユーザの動作に関する各種情報を検知する機能を有してもよい。具体的には、センサ部140は、ユーザの動作に関する情報として、ユーザの頭部や姿勢に関する情報、ユーザの頭部や身体の動き(加速度や角速度)、視野の方向や視点移動の速度等を検知する。例えば、センサ部140は、3軸加速度センサや、ジャイロセンサや、速度センサ等の各種モーションセンサとして機能し、ユーザの動作に関する情報を検知する。より具体的には、センサ部140は、ユーザの頭部の動きとして、ヨー(yaw)方向、ピッチ(pitch)方向、及びロール(roll)方向それぞれの成分を検出することで、ユーザの頭部の位置及び姿勢のうち少なくともいずれかの変化を検知する。 Furthermore, for example, when the display control device 100 is a wearable device such as an HMD, the sensor unit 140 may have a function of detecting various information related to the user's motion, such as the orientation, inclination, motion, and moving speed of the user's body, in addition to the camera 141 and microphone 142. Specifically, the sensor unit 140 detects information related to the user's motion, such as information related to the user's head and posture, the motion of the user's head and body (acceleration and angular velocity), the direction of the field of view, and the speed of the viewpoint movement. For example, the sensor unit 140 functions as various motion sensors such as a three-axis acceleration sensor, a gyro sensor, and a speed sensor, and detects information related to the user's motion. More specifically, the sensor unit 140 detects at least one change in the position and posture of the user's head by detecting the components of the yaw direction, pitch direction, and roll direction as the motion of the user's head.
 なお、センサ部140は、必ずしも表示制御装置100に備えられることを要せず、例えば、表示制御装置100と有線もしくは無線で接続される外部センサであってもよい。例えば、カメラ141は、必ずしも表示制御装置100内部に備えられるものではなく、表示制御装置100外部に設置された撮像装置であってもよい。 The sensor unit 140 does not necessarily have to be provided in the display control device 100, and may be, for example, an external sensor connected to the display control device 100 by wire or wirelessly. For example, the camera 141 does not necessarily have to be provided inside the display control device 100, and may be an imaging device installed outside the display control device 100.
 表示部150は、制御部130から出力された各種情報を表示する。例えば、表示部150は、ユーザに向けて映像を出力するディスプレイである。なお、表示部150は、音声を出力する音声出力部(スピーカー等)を含んでもよい。 The display unit 150 displays various information output from the control unit 130. For example, the display unit 150 is a display that outputs video to the user. The display unit 150 may also include an audio output unit (such as a speaker) that outputs audio.
 制御部130は、例えば、CPU(Central Processing Unit)やMPU(Micro Processing Unit)、GPU等によって、表示制御装置100内部に記憶されたプログラム(例えば、本開示に係る表示制御プログラム)がRAM等を作業領域として実行されることにより実現される。また、制御部130は、コントローラ(controller)であり、例えば、ASIC(Application Specific Integrated Circuit)やFPGA(Field Programmable Gate Array)等の集積回路により実現されてもよい。 The control unit 130 is realized, for example, by a CPU (Central Processing Unit), MPU (Micro Processing Unit), GPU, etc., executing a program stored inside the display control device 100 (for example, a display control program related to the present disclosure) using a RAM or the like as a working area. The control unit 130 is also a controller, and may be realized, for example, by an integrated circuit such as an ASIC (Application Specific Integrated Circuit) or FPGA (Field Programmable Gate Array).
 図3に示すように、制御部130は、取得部131と、判定部132と、表示制御部133とを有する。 As shown in FIG. 3, the control unit 130 has an acquisition unit 131, a determination unit 132, and a display control unit 133.
 取得部131は、各種情報を取得する。例えば、取得部131は、センサ部140を用いてユーザをセンシングすることによって、アバター10に反映させる入力情報を取得する。具体的には、取得部131は、実空間におけるユーザの挙動を入力情報として取得する。 The acquisition unit 131 acquires various types of information. For example, the acquisition unit 131 acquires input information to be reflected in the avatar 10 by sensing the user using the sensor unit 140. Specifically, the acquisition unit 131 acquires the user's behavior in real space as input information.
 また、取得部131は、ユーザからの各種操作や設定要求を取得してもよい。例えば、取得部131は、ユーザから特定の動作をアバター10に反映するか否かの指定を受け付ける。すなわち、取得部131は、ユーザから、無意識的動作を入力する入力モードとするか、意識的動作(ジェスチャー)を入力する入力モードとするかの指定を受け付ける。 The acquisition unit 131 may also acquire various operations and setting requests from the user. For example, the acquisition unit 131 accepts designation from the user as to whether or not to reflect a specific action in the avatar 10. In other words, the acquisition unit 131 accepts designation from the user as to whether to set an input mode for inputting an unconscious action or an input mode for inputting a conscious action (gesture).
 また、取得部131は、入力情報が取得される際、センサ部140等を用いて周辺情報等を取得してもよい。具体的には、取得部131は、図2で示したように、ユーザの挙動が撮像される際のカメラ141からの空間的な位置を取得する。これにより、後述する判定部132は、ユーザが実行するハンドサインがカメラ141からどのくらいの距離で撮像されているかを判定できるので、モードをジェスチャーモードに自動的に切り替えることができる。 Furthermore, when the input information is acquired, the acquisition unit 131 may acquire surrounding information, etc., using the sensor unit 140, etc. Specifically, as shown in FIG. 2, the acquisition unit 131 acquires the spatial position from the camera 141 when the user's behavior is captured. This allows the determination unit 132, which will be described later, to determine at what distance from the camera 141 the hand sign performed by the user is captured, and therefore the mode can be automatically switched to the gesture mode.
 なお、取得部131は、カメラ141によって現に撮像されたユーザのみならず、仮想空間に接続される複数のユーザに関する情報を取得することもできる。すなわち、取得部131は、複数のユーザの挙動を入力情報として取得することができる。このような複数のユーザの入力情報を用いることで、判定部132は、後述するような複数のユーザによる会話の盛り上がり等の判定を行うことができる。 The acquisition unit 131 can acquire information not only about the user currently captured by the camera 141, but also about multiple users connected to the virtual space. In other words, the acquisition unit 131 can acquire the behavior of multiple users as input information. By using such input information from multiple users, the determination unit 132 can determine the level of interest in a conversation between multiple users, as described below.
 また、取得部131は、UIを介して、ユーザから送信される各種要求を受け付けてもよい。例えば、取得部131は、アバターの外観を変更する旨の要求や、アバターに紐付けられている物体(仮想空間上のアイテム等)の購入や選択、交換等、仮想空間に係るサービスに対する各種操作等を受け付ける。 The acquisition unit 131 may also receive various requests sent from the user via the UI. For example, the acquisition unit 131 receives requests to change the appearance of the avatar, and various operations for services related to the virtual space, such as purchasing, selecting, or exchanging an object (such as an item in the virtual space) linked to the avatar.
 判定部132は、アバター10を表示制御するための各種判定を行う。例えば、判定部132は、予め登録された特定の動作に対応する入力がユーザからの入力情報に含まれるか否かを判定する。予め登録された特定の動作に対応する入力とは、すなわち、ユーザがジェスチャーを発火させるための登録済みの挙動である。 The determination unit 132 performs various determinations for controlling the display of the avatar 10. For example, the determination unit 132 determines whether or not the input information from the user includes an input corresponding to a specific action that has been registered in advance. The input corresponding to a specific action that has been registered in advance is, in other words, a registered behavior that the user uses to trigger a gesture.
 例えば、判定部132は、ユーザから入力モードの指定(ジェスチャーモードの指定)がある場合には、特定の動作に対応する入力が入力情報に含まれるか否かを判定する。すなわち、判定部132は、ユーザがカメラ141の前で行った挙動が、ジェスチャーを発火させる意識的動作か否かを判定する。 For example, when the user specifies an input mode (gesture mode), the determination unit 132 determines whether or not the input information includes an input corresponding to a specific action. In other words, the determination unit 132 determines whether or not the behavior performed by the user in front of the camera 141 is an intentional action that triggers a gesture.
 また、判定部132は、距離に応じてモードが変更される場合、カメラ141からの位置が所定距離より離れている場合には、特定の動作に対応する入力が入力情報に含まれるか否かを判定せず、カメラ141からの位置が所定距離以内である場合に、特定の動作に対応する入力が入力情報に含まれるか否かを判定する。すなわち、判定部132は、カメラ141に比較的近い位置でユーザから出されたハンドサイン等の挙動に対してはジェスチャー判定を行い、設定された空間範囲から外れた範囲でユーザから出されたハンドサイン等の挙動に対してはジェスチャー判定を行わず、無意識的動作と判定する。これにより、判定部132は、入力モードを自在に切り替えてアバター10を表現することができる。 Furthermore, when the mode is changed according to the distance, the determination unit 132 does not determine whether or not the input information includes an input corresponding to a specific action if the position from the camera 141 is farther than a predetermined distance, but determines whether or not the input information includes an input corresponding to a specific action if the position from the camera 141 is within the predetermined distance. In other words, the determination unit 132 performs gesture determination for behaviors such as hand signs made by the user in a position relatively close to the camera 141, and does not perform gesture determination for behaviors such as hand signs made by the user outside the set spatial range, determining them to be unconscious actions. This allows the determination unit 132 to freely switch input modes to express the avatar 10.
 表示制御部133は、制御部130から出力される情報を表示部150に表示するよう制御する。すなわち、表示制御部133は、映像コンテンツとしてレンダリングされた仮想空間映像をディスプレイに出力する。なお、出力先デバイスとは、内蔵ディスプレイに限らず、スマートフォン等の他の情報機器や、テレビなどの映像出力装置でもよい。 The display control unit 133 controls the display unit 150 to display information output from the control unit 130. In other words, the display control unit 133 outputs the virtual space image rendered as video content to the display. Note that the output destination device is not limited to the built-in display, but may be other information devices such as a smartphone, or a video output device such as a television.
 例えば、表示制御部133は、ユーザを撮像した画像や、ユーザが発した音声や、ユーザが入力したテキスト等、ユーザから取得された入力情報に基づき、当該入力情報に対応するユーザの挙動を仮想空間のアバター10に反映する。さらに、表示制御部133は、入力情報に特定の動作に対応する入力が含まれると判定された場合には、特定の動作であるジェスチャーを仮想空間のアバター10に反映して、仮想空間にアバター10を表示する。 For example, based on input information acquired from the user, such as an image of the user, a voice uttered by the user, or text entered by the user, the display control unit 133 reflects the user's behavior corresponding to the input information in the avatar 10 in the virtual space. Furthermore, when the display control unit 133 determines that the input information includes an input corresponding to a specific action, it reflects a gesture, which is a specific action, in the avatar 10 in the virtual space and displays the avatar 10 in the virtual space.
 例えば、表示制御部133は、無意識的動作を入力するモードである場合、ユーザを撮像した画像に基づいて取得されるユーザの骨格データ等をアバター10に反映する。一方、表示制御部133は、ジェスチャーモードである場合、ユーザの挙動がジェスチャーを発火させるものと判定部132により判定された場合、予め登録されていた、特定のアニメーションや演出効果を含むジェスチャーをアバター10に反映させる。すなわち、アバター10によるジェスチャーは、サムアップやハイタッチ、拍手等、特殊な動きや演出を伴うアニメーションにより表現される。なお、表示制御部133は、ジェスチャーとともに、アバターによる音声出力(予め登録されていた笑い声)等を制御してもよい。 For example, when the display control unit 133 is in a mode for inputting unconscious actions, it reflects the user's skeletal data, etc., acquired based on an image of the user, in the avatar 10. On the other hand, when the display control unit 133 is in a gesture mode, if the determination unit 132 determines that the user's behavior will trigger a gesture, it reflects a gesture, including a specific animation or special effects, that has been registered in advance, in the avatar 10. That is, gestures made by the avatar 10 are expressed by animations that involve special movements or special effects, such as a thumbs-up, high-five, or clapping. Note that the display control unit 133 may control the sound output by the avatar (pre-registered laughter) in addition to the gestures.
(1-3.実施形態に係る処理の手順)
 次に、図5を用いて、実施形態に係る処理の手順について説明する。図5は、実施形態に係る表示制御処理の流れを示すフローチャートである。
(1-3. Processing Procedure According to the Embodiment)
Next, a process procedure according to the embodiment will be described with reference to Fig. 5. Fig. 5 is a flowchart showing the flow of a display control process according to the embodiment.
 図5に示すように、表示制御装置100は、センサ部140を用いて、ユーザを撮像した画像を取得する(ステップS11)。なお、表示制御装置100は、上述のように、ユーザの音声や入力されたテキスト等を画像とともに取得してもよい。 As shown in FIG. 5, the display control device 100 uses the sensor unit 140 to acquire an image of the user (step S11). As described above, the display control device 100 may also acquire the user's voice, input text, and the like along with the image.
 その後、表示制御装置100は、既知の姿勢推定技術等に基づいて、ユーザの画像を骨格データに変換する(ステップS12)。 Then, the display control device 100 converts the user's image into skeletal data based on known posture estimation techniques, etc. (step S12).
 続いて、表示制御装置100は、現在のモードがジェスチャーモードであるか否かを判定する(ステップS13)。ジェスチャーモードである場合(ステップS13;Yes)、表示制御装置100は、骨格データにおいて特徴点位置と動きを判定する(ステップS14)。 Then, the display control device 100 determines whether the current mode is the gesture mode (step S13). If the current mode is the gesture mode (step S13; Yes), the display control device 100 determines the feature point positions and movements in the skeletal data (step S14).
 そして、表示制御装置100は、特徴点位置と動きが、登録済みのジェスチャーに対応するものか否かを判定する(ステップS15)。登録済みのジェスチャーに対応するものである場合(ステップS15;Yes)、表示制御装置100は、記憶部120からジェスチャーのモーションデータを読み出す(ステップS16)。そして、表示制御装置100は、読み出したモーションデータをアバター10に反映する(ステップS18)。 Then, the display control device 100 determines whether or not the feature point position and movement correspond to a registered gesture (step S15). If they correspond to a registered gesture (step S15; Yes), the display control device 100 reads out the motion data of the gesture from the storage unit 120 (step S16). The display control device 100 then reflects the read out motion data in the avatar 10 (step S18).
 なお、モードがジェスチャーモードでない場合(ステップS13;No)、表示制御装置100は、ジェスチャー判定を行わず、ユーザから得られた骨格データをアバター10の骨格データに変換する(ステップS17)。そして、表示制御装置100は、骨格データをアバター10に反映する(ステップS18)。 If the mode is not the gesture mode (step S13; No), the display control device 100 does not perform gesture determination, and converts the skeletal data obtained from the user into skeletal data for the avatar 10 (step S17). Then, the display control device 100 reflects the skeletal data in the avatar 10 (step S18).
 また、ステップS15において、特徴点位置と動きがジェスチャーに対応するものでなかった場合も(ステップS15;No)、表示制御装置100は、ユーザから得られた骨格データをアバター10の骨格データに変換する(ステップS17)。そして、表示制御装置100は、骨格データをアバター10に反映する(ステップS18)。 Also, in step S15, if the feature point positions and movements do not correspond to a gesture (step S15; No), the display control device 100 also converts the skeletal data obtained from the user into skeletal data for the avatar 10 (step S17). Then, the display control device 100 reflects the skeletal data in the avatar 10 (step S18).
 その後、表示制御装置100は、アバター10に反映された情報に基づいて、アバター10を仮想空間に表示する(ステップS19)。表示制御装置100は、センサ部140によって取得される所定フレームごとに、かかる表示制御処理を繰り返す。 Then, the display control device 100 displays the avatar 10 in the virtual space based on the information reflected in the avatar 10 (step S19). The display control device 100 repeats this display control process for each predetermined frame acquired by the sensor unit 140.
(1-4.実施形態に係る処理のバリエーション)
(1-4-1.アバター同士の盛り上がり表現)
 上述のような無意識的動作および意識的動作によるアバター10の表現は、様々な他の表現と組み合わされてもよい。以下では、実施形態に係る表示制御処理のバリエーションを説明する。
(1-4. Variations of Processing According to the Embodiment)
(1-4-1. Expressing excitement between avatars)
The above-described expressions of the avatar 10 based on unconscious and conscious actions may be combined with various other expressions. Variations of the display control process according to the embodiment will be described below.
 まず、第1のバリエーションについて説明する。第1のバリエーションでは、表示制御装置100は、複数のアバター同士の盛り上がりに応じて様々な視覚効果を表示する。なお、アバター同士の盛り上がりとは、すなわちユーザ同士の盛り上がりを意味する。これにより、表示制御装置100は、仮想空間において、複数のユーザ間で活発なコミュニケーションが行われていることを他のユーザも理解できるような形で表現することができる。 First, the first variation will be described. In the first variation, the display control device 100 displays various visual effects according to the excitement between multiple avatars. Note that the excitement between avatars means the excitement between users. In this way, the display control device 100 can express in a way that other users can understand that active communication is taking place between multiple users in the virtual space.
 すなわち、第1のバリエーションにおいて、取得部131は、複数のユーザの挙動を入力情報として取得する。そして、判定部132は、複数のユーザの挙動に基づいて、仮想空間における複数のユーザ同士の交流の盛り上がり度を判定する。さらに、表示制御部133は、複数のユーザ同士の近傍に、交流の盛り上がり度に応じた演出効果を表示する。 That is, in the first variation, the acquisition unit 131 acquires the behavior of multiple users as input information. Then, the determination unit 132 determines the level of excitement of the interaction between the multiple users in the virtual space based on the behavior of the multiple users. Furthermore, the display control unit 133 displays a presentation effect in the vicinity of the multiple users according to the level of excitement of the interaction.
 具体的には、判定部132は、交流に参加するユーザの人数、ユーザ同士の会話の頻度、複数のユーザ同士の音声の重なり具合に基づいて、交流の盛り上がり度を判定する。 Specifically, the determination unit 132 determines the level of excitement in an exchange based on the number of users participating in the exchange, the frequency of conversations between users, and the degree of overlap in the voices of multiple users.
 また、表示制御部133は、交流の盛り上がり度、アバター10の身体の向き、複数のユーザの会話の内容に基づいて、演出効果の表現もしくは発生頻度を決定してもよい。 The display control unit 133 may also determine the expression or frequency of occurrence of the dramatic effect based on the level of excitement in the interaction, the body orientation of the avatar 10, and the content of the conversation between multiple users.
 以下、図6を用いて、第1のバリエーションについて具体的に説明する。図6は、実施形態に係る表示制御処理の第1のバリエーションを説明するための図(1)である。 The first variation will be specifically described below with reference to FIG. 6. FIG. 6 is a diagram (1) for explaining the first variation of the display control process according to the embodiment.
 図6に示す例では、複数のアバターであるアバター10Aとアバター10Bとアバター10Cとが、挨拶や会話をすることで盛んに交流していることを示す。このとき、表示制御装置100は、複数のユーザ同士の交流の盛り上がり度を判定し、判定した盛り上がり度に基づいて、円弧状の演出効果200や、アバター周囲の放射線状の演出効果201を表示する。このように、表示制御装置100は、実際の会話に参加していなくとも、遠くから様子を見るだけで、盛んに交流している複数のアバターを見分けることができるような表示を行うことができる。 The example shown in FIG. 6 shows that multiple avatars, avatar 10A, avatar 10B, and avatar 10C, are actively interacting by exchanging greetings and conversations. At this time, the display control device 100 determines the level of excitement in the interaction between the multiple users, and displays an arc-shaped effect 200 and radial effect 201 around the avatars based on the determined level of excitement. In this way, the display control device 100 can perform a display that allows multiple avatars who are actively interacting to be distinguished simply by looking at the situation from a distance, even if they are not actually participating in the conversation.
 なお、アバター同士の盛り上がり度は、様々なパラメータから導出されてよい。例えば、アバター同士の盛り上がりが、ボイスチャット等の音声コミュニケーションによるものである場合、パラメータとしては、ユーザの声の大きさや速さ、高さ、感情、オーバーラップの頻度、フィラーの量、特定の言語表現の有無等が用いられる。すなわち、仮想空間におけるコミュニケーションの盛り上がりは、必ずしも音声の大きさや会話量だけで測定されるものではない。 The level of excitement between avatars may be derived from various parameters. For example, if the excitement between avatars is due to voice communication such as voice chat, parameters that may be used include the volume, speed, pitch, emotion, frequency of overlap, amount of filler, and the presence or absence of specific language expressions of the users' voices. In other words, the level of excitement in communication in a virtual space is not necessarily measured only by the volume of the voices or the amount of conversation.
 例えば、仮想的な3次元空間のような立体音響下においては、実空間と同じく同時発話が成立する。このため、表示制御装置100は、例えば、「オーバーラップ(複数人の同時発話)」を盛り上がり判定の重要度の高い(重み付けがされた)パラメータとして用いることができる。 For example, in a stereophonic environment such as a virtual three-dimensional space, simultaneous speech is possible just as in a real space. For this reason, the display control device 100 can use, for example, "overlap (simultaneous speech by multiple people)" as a parameter with high importance (weighted) for determining excitement.
 また、表示制御装置100は、分析サーバ50等に音声を分析させることにより、各音声の感情を分析し、より感情的である音声が含まれる場合に、盛り上がり度を高く算出してもよい。具体的には、表示制御装置100は、「イェーイ(日本語)」や「Wow!Yeah!(英語)」など、感嘆や強い感情を表す掛け声が含まれる交流について、盛り上がり度を高く算出してもよい。あるいは、表示制御装置100は、「乾杯!」など、祝いや喜びの意図を強く含む挨拶やワード(感情分析に基づくポジティブワード)の発生頻度等に基づいて、盛り上がり度を高く算出してもよい。 The display control device 100 may also analyze the emotion of each voice by having the analysis server 50 or the like analyze the voices, and may calculate a high excitement level when a more emotional voice is included. Specifically, the display control device 100 may calculate a high excitement level for an exchange that includes shouts expressing admiration or strong emotion, such as "Yay (Japanese)" or "Wow! Yeah! (English)." Alternatively, the display control device 100 may calculate a high excitement level based on the frequency of occurrence of greetings or words (positive words based on emotion analysis) that strongly imply an intention of celebration or joy, such as "Cheers!".
 また、表示制御装置100は、盛り上がり度に応じて、表示する演出(視覚)効果を変化させることができる。 The display control device 100 can also change the presentation (visual) effects displayed depending on the level of excitement.
 例えば、表示制御装置100は、円弧で示される演出効果について、盛り上がり度に応じて大きさや頻度を変化させてもよい。具体的には、表示制御装置100は、盛り上がり度が比較的高い場合には、会話に参加している複数のユーザ全てを覆うように、大きく円弧を表示してもよい。これにより、表示制御装置100は、複数のユーザを遠目から見るユーザに対しても、複数のユーザの盛り上がりを正確に伝えることができる。 For example, the display control device 100 may change the size or frequency of the dramatic effect represented by the arc depending on the level of excitement. Specifically, when the level of excitement is relatively high, the display control device 100 may display a large arc so as to cover all of the multiple users participating in the conversation. This allows the display control device 100 to accurately convey the excitement of the multiple users even to a user who is viewing the multiple users from a distance.
 また、表示制御装置100は、必ずしも会話が活発でなくても、演出効果を大きく表示してもよい。例えば、複数のユーザが何らかの仮想空間上のイベントに注目しているような場合、会話は少なくなるものの、複数のユーザにとっての交流は盛んであると判定される。この場合、表示制御装置100は、例えば、互いのアバターの視線の交錯や、会話以外の交流(ユーザ同士の距離の近さなど)に基づく盛り上がり度を判定してもよい。そして、表示制御装置100は、盛り上がり度に応じて、円弧状や放射線上の演出効果を通常よりも大きく、あるいは、明るく表示する。これにより、表示制御装置100は、会話によらずとも、複数のユーザが盛んに交流していることを他のユーザに示すことができる。 The display control device 100 may also display a large effect even if conversation is not necessarily active. For example, if multiple users are paying attention to some kind of event in the virtual space, conversation will be reduced, but it is determined that the users are actively interacting with each other. In this case, the display control device 100 may determine the level of excitement based on, for example, the intersection of the avatars' gazes or interactions other than conversation (such as the closeness of the users to each other). Then, depending on the level of excitement, the display control device 100 displays the arc-shaped or radial effect larger or brighter than usual. In this way, the display control device 100 can show other users that multiple users are actively interacting with each other, even without conversation.
 次に、図7を用いて、表示制御装置100が盛り上がり度を判定する処理の一例を示す。図7は、実施形態に係る表示制御処理の第1のバリエーションを説明するための図(2)である。 Next, an example of a process in which the display control device 100 determines the level of excitement will be shown with reference to FIG. 7. FIG. 7 is a diagram (2) for explaining a first variation of the display control process according to the embodiment.
 図7には、複数のユーザの会話から取得された合成波形205を示す。図7において、範囲206は、複数のユーザの音声の音量が高く測定された波形を示しており、比較的大きな音声での会話がなされていることを示す。また、範囲207は、複数のユーザの多数の音声が合成された波形を示しており、活発な会話の応酬がなされていることを示す。 FIG. 7 shows a composite waveform 205 obtained from a conversation between multiple users. In FIG. 7, range 206 shows a waveform in which the voices of multiple users are measured to be high, indicating that the conversation is taking place at a relatively loud volume. Range 207 shows a waveform in which multiple voices from multiple users are synthesized, indicating that a lively exchange of conversation is taking place.
 表示制御装置100は、このような波形解析に基づき、複数のユーザの盛り上がり度を判定する。例えば、表示制御装置100は、所定時間における音量変化の特徴に基づいて、盛り上がり度を判定する。あるいは、表示制御装置100は、波形解析に基づいて会話の参加人数を推定し、推定した人数が多いほど、盛り上がり度を高く判定してもよい。 The display control device 100 determines the level of excitement among multiple users based on such waveform analysis. For example, the display control device 100 determines the level of excitement based on the characteristics of volume changes over a specified period of time. Alternatively, the display control device 100 may estimate the number of participants in a conversation based on waveform analysis, and determine the level of excitement to be higher the greater the estimated number of participants.
 また、表示制御装置100は、波形に基づいて、演出効果の表現を変化させてもよい。例えば、表示制御装置100は、波形の振幅に即して、円弧状や放射状の表示の色や密度(発生頻度)、円弧や放射状の光が広がっていく速度等を変更してもよい。 The display control device 100 may also change the representation of the dramatic effect based on the waveform. For example, the display control device 100 may change the color or density (frequency of occurrence) of an arc-shaped or radial display, the speed at which the arc-shaped or radial light spreads, etc., in accordance with the amplitude of the waveform.
 また、表示制御装置100は、演出効果として、必ずしも円弧状の表現を行うのではなく、音声や会話内容や感情に応じて、連続した鋭角を有する吹き出し上の表示を行ってもよい。これにより、表示制御装置100は、例えば、複数のユーザ同士が言い争いをしていること等を表現することができる。この場合、表示制御装置100は、吹き出し等の演出効果を、見る角度によらずに成立する3次元表示(リムシェーダーの応用)などで表示してもよい。 In addition, the display control device 100 may not necessarily display an arc-shaped effect, but may display a speech bubble with a series of acute angles according to the voice, conversation content, or emotion. This allows the display control device 100 to express, for example, that multiple users are arguing with each other. In this case, the display control device 100 may display speech bubbles or other dramatic effects using a three-dimensional display (application of a rim shader) that is valid regardless of the viewing angle.
 また、表示制御装置100は、観測者から複数のユーザまでの距離に応じて、演出効果の発生する速度を変えてもよい。具体的には、表示制御装置100は、距離が遠いほど演出効果の表示を遅くし、距離が近付くほど演出効果の表示を速くしてもよい。これにより、観測するユーザは、活発に交流するユーザがどれくらいの距離にいるかを一目で把握することができる。 The display control device 100 may also change the speed at which the special effects are generated depending on the distance from the observer to the multiple users. Specifically, the display control device 100 may display the special effects slower the farther the distance is, and may display the special effects faster the closer the distance is. This allows the observing user to know at a glance how far away the actively interacting users are.
(1-4-2.アバターの感情表現)
 続いて、第2のバリエーションについて説明する。第2のバリエーションでは、表示制御装置100がユーザの笑い声等、感情が強く示された入力を取得し、ユーザが笑っていると判定される場合に、アバター10に笑いのジェスチャーを反映させる。
(1-4-2. Avatar Emotion Expression)
Next, a second variation will be described. In the second variation, when the display control device 100 acquires an input that strongly indicates an emotion, such as a laughing voice of the user, and determines that the user is laughing, the display control device 100 causes the avatar 10 to make a laughing gesture.
 すなわち、第2のバリエーションにおいて、判定部132は、笑い等の特定の感情に紐付いた特定の動作に対応する入力が、ユーザからの入力情報に含まれるか否かを判定する。表示制御部133は、入力情報に特定の感情に紐付いた特定の動作が含まれると判定された場合には、特定の感情に紐付いた特定の動作をアバター10に反映する。 In other words, in the second variation, the determination unit 132 determines whether or not the input information from the user includes an input corresponding to a specific action linked to a specific emotion, such as laughing. If it is determined that the input information includes a specific action linked to a specific emotion, the display control unit 133 reflects the specific action linked to the specific emotion in the avatar 10.
 このとき、判定部132は、予めユーザから登録された声紋情報であって、特定の感情に紐付く声紋情報と入力情報とを比較することで、特定の感情に紐付いた特定の動作に対応する入力が入力情報に含まれるか否かを判定してもよい。 At this time, the determination unit 132 may compare the input information with voiceprint information that is registered in advance by the user and is associated with a specific emotion, to determine whether the input information includes an input corresponding to a specific action associated with the specific emotion.
 また、表示制御部133は、特定の感情に紐付いた特定の動作を反映させたアバター10を表示するとともに、特定の感情に紐付いた演出効果をアバター10の近傍に表示してもよい。 The display control unit 133 may also display an avatar 10 that reflects a specific action linked to a specific emotion, and may also display a dramatic effect linked to the specific emotion near the avatar 10.
 第2のバリエーションについて、図8以下を用いて具体的に説明する。図8は、実施形態に係る表示制御処理の第2のバリエーションを説明するための図(1)である。 The second variation will be described in detail with reference to FIG. 8 and subsequent figures. FIG. 8 is a diagram (1) for explaining the second variation of the display control process according to the embodiment.
 図8に示す例では、アバター10が、特定の感情の一例である「笑い」に紐付いてジェスチャーを行っていることを示す。このとき、表示制御装置100は、アバター10の周囲に、笑いに紐付く漫画的表現の演出効果210を表示してもよい。すなわち、表示制御装置100は、ユーザが笑ったことに対応してアバター10に笑いのジェスチャーを実行させることで、感情表現が豊かなアバター10を表示することができる。 In the example shown in FIG. 8, the avatar 10 is making a gesture associated with "laughing," which is an example of a specific emotion. At this time, the display control device 100 may display a cartoon-like effect 210 associated with laughter around the avatar 10. In other words, the display control device 100 can display an avatar 10 that is rich in emotional expression by having the avatar 10 perform a laughing gesture in response to the user laughing.
 表示制御装置100は、いくつかの手法により、ユーザが笑ったことを判定することができる。この点について、図9を用いて説明する。図9は、実施形態に係る表示制御処理の第2のバリエーションを説明するための図(2)である。 The display control device 100 can determine that the user has laughed using several techniques. This will be explained with reference to FIG. 9. FIG. 9 is a diagram (2) for explaining a second variation of the display control process according to the embodiment.
 図9には、ユーザから取得される音声波形215を例示する。音声波形215において、ユーザがボイスチャット等を行う際の環境音(暗騒音)を測定した際の測定値を測定結果216とする。また、音声波形215において、ユーザがボイスチャット等を行った際の音声波形を測定結果217とする。このとき、測定結果217のピークは、結果218の線で示される。そして、結果218と測定結果216との差は、差分219で示される。 FIG. 9 shows an example of an audio waveform 215 obtained from a user. In the audio waveform 215, the measurement value obtained when the environmental sound (background noise) is measured while the user is engaged in voice chat or the like is set as measurement result 216. In addition, in the audio waveform 215, the audio waveform obtained when the user is engaged in voice chat or the like is set as measurement result 217. At this time, the peak of measurement result 217 is indicated by the line of result 218. And the difference between result 218 and measurement result 216 is indicated by difference 219.
 すなわち、表示制御装置100は、マイクロホン142を用いて、ユーザがボイスチャット等を行う際の環境音(暗騒音)を測定する。その後、表示制御装置100は、ユーザの音声を測定し、その音量のピークを笑い声と推定し、環境音との差分219を測定する。このように、表示制御装置100は、ユーザごとに、音声のピーク値と、環境音との差分値を登録しておく。 In other words, the display control device 100 uses the microphone 142 to measure the environmental sound (background noise) when the user is engaged in voice chat, etc. Then, the display control device 100 measures the user's voice, estimates the peak volume as laughter, and measures the difference 219 with the environmental sound. In this way, the display control device 100 registers the peak value of the voice and the difference value with the environmental sound for each user.
 そして、表示制御装置100は、次にユーザが音声を発し、差分219を超える(もしくは同等の)音声が観測された際に、ユーザが笑ったと判定する。かかる手法によれば、表示制御装置100は、音量に基づいて、ユーザが笑ったか否かを判定することができる。 Then, when the user next makes a sound and a sound exceeding (or equal to) the difference 219 is observed, the display control device 100 determines that the user has laughed. According to this method, the display control device 100 can determine whether or not the user has laughed based on the sound volume.
 なお、表示制御装置100は、他の手法として、ユーザが笑った際の声紋情報を学習し、かかる学習結果に基づく判定を行うこともできる。この点について、図10を用いて説明する。図10は、実施形態に係る表示制御処理の第2のバリエーションを説明するための図(3)である。 As another method, the display control device 100 can learn voiceprint information when the user laughs and make a determination based on the learning results. This will be explained using FIG. 10. FIG. 10 is a diagram (3) for explaining a second variation of the display control process according to the embodiment.
 図10には、ユーザから取得された音声波形220を例示する。音声波形220のうち、範囲221は、ユーザの笑い声を測定した際の波形である。表示制御装置100(もしくは分析サーバ50)は、ユーザごとに笑い声を測定した際の波形を集積し、集積した波形の特徴を既知の音声学習手法(例えば特徴的なパターンを抽出する深層学習等)により学習する。これにより、表示制御装置100は、ユーザごとに、「笑い」の状態を示した学習済み声紋パターンを生成することができる。 FIG. 10 illustrates an example of a voice waveform 220 acquired from a user. Within the voice waveform 220, range 221 is the waveform when the user's laughter is measured. The display control device 100 (or the analysis server 50) accumulates the waveforms when laughter is measured for each user, and learns the characteristics of the accumulated waveforms using a known voice learning method (e.g., deep learning that extracts characteristic patterns). This allows the display control device 100 to generate a learned voiceprint pattern that indicates the "laughing" state for each user.
 そして、表示制御装置100は、次にユーザが音声を発し、音声波形が取得された際には、学習済み声紋パターンと比較することで、当該音声に笑い声が含まれているかを判定できる。これにより、表示制御装置100は、高い精度で、ユーザが笑ったか否かを判定することができる。 Then, the next time the user speaks and the voice waveform is acquired, the display control device 100 can determine whether the voice contains laughter by comparing it with the learned voiceprint pattern. This allows the display control device 100 to determine with high accuracy whether the user has laughed.
 なお、人の笑い声は多様であり、また、笑い声に含まれる感情も、単に喜びだけでなく、様々な感情が含まれる。このため、表示制御装置100は、学習により笑い声を判定する機械学習モデルにおいて、笑ったか笑っていないかの2値判定ではなく、感情等をカテゴライズするようモデルを生成してもよい。 It should be noted that human laughter is diverse, and the emotions contained in laughter include not only joy but also various other emotions. For this reason, in a machine learning model that determines laughter through learning, the display control device 100 may generate a model that categorizes emotions, etc., rather than a binary determination of whether or not someone is laughing.
 また、学習データとしての笑い声を取得する手法としても、いくつかの手法が考えられる。例えば、表示制御装置100は、ユーザの明示の指定に基づいて、笑いを示す波形を取得する。具体的には、表示制御装置100は、ユーザが笑い声を表示制御装置100に取得(録音)させるための任意のボタンを押下したことを契機として、ユーザの音声を取得する。このとき、ユーザは、任意に笑い声を発する。これにより、表示制御装置100は、ユーザの笑い声を示す声紋パターン(すなわち、「笑い」というラベルの付与された教師データ)を取得することができる。あるいは、表示制御装置100は、笑い声を判定する汎用の音声認識用の機械学習モデルを用いて、ユーザの笑い声のみを集積するように、ユーザの声紋パターンを取得してもよい。一般に、笑い声の特徴は個々人でばらつきがあるため、表示制御装置100は、個々のユーザごとに声紋パターンを取得し、学習することが望ましい。 Also, there are several possible methods for acquiring laughter as training data. For example, the display control device 100 acquires a waveform indicating laughter based on the user's explicit designation. Specifically, the display control device 100 acquires the user's voice when the user presses an arbitrary button to cause the display control device 100 to acquire (record) laughter. At this time, the user laughs at will. This allows the display control device 100 to acquire a voiceprint pattern indicating the user's laughter (i.e., training data labeled "laughter"). Alternatively, the display control device 100 may acquire the user's voiceprint pattern so as to accumulate only the user's laughter using a general-purpose machine learning model for voice recognition that determines laughter. In general, the characteristics of laughter vary from person to person, so it is desirable for the display control device 100 to acquire and learn a voiceprint pattern for each individual user.
 なお、表示制御装置100は、笑い声に限らず、様々なユーザの感情に基づく音声を取得してもよい。そして、表示制御装置100は、取得した音声に基づく表現をアバター10に反映することで、ユーザ独自の感情表現をアバター10でも表現することができる。 The display control device 100 may acquire voice based on various emotions of the user, not limited to laughter. Then, the display control device 100 can cause the avatar 10 to express the user's unique emotional expression by reflecting the expression based on the acquired voice in the avatar 10.
 一例として、表示制御装置100は、ユーザごとのフィラーの特徴を取得し、フィラーを発するユーザの特徴をアバター10に反映してもよい。ユーザが発する「えーっと」などのフィラーや、「いいですね」などの口癖は、そのユーザの特徴を示すものであるといえる。このため、表示制御装置100は、ユーザごとのフィラーや口癖を学習し、フィラーや口癖が観測された場合に、それをアバター10に反映することで、そのユーザの特徴を仮想空間上で再現できる。この場合、表示制御装置100は、フィラーや口癖として発せられた音声を吹き出しで3D表現するなど、強調的、漫画的表現を伴うことで、より当該ユーザの特徴をアバター10において反映させることができる。 As an example, the display control device 100 may acquire the filler characteristics of each user and reflect the characteristics of the user who utters the filler in the avatar 10. Fillers such as "umm" and catchphrases such as "that's nice" uttered by a user can be said to indicate the characteristics of that user. For this reason, the display control device 100 learns the fillers and catchphrases of each user, and when a filler or catchphrase is observed, reflects it in the avatar 10, thereby reproducing the characteristics of that user in a virtual space. In this case, the display control device 100 can better reflect the characteristics of the user in the avatar 10 by emphasizing and cartoonish expressions, such as by expressing the sounds uttered as fillers or catchphrases in 3D using speech bubbles.
(1-4-3.会話の可視化表現)
 続いて、第3のバリエーションについて説明する。第3のバリエーションでは、表示制御装置100がアバター同士の会話を仮想空間上で表示する際に、一般的な吹き出しに会話内容を示すテキストを表示するのみならず、多様な表現で会話内容を表現する手法を説明する。
(1-4-3. Visualization of conversation)
Next, a third variation will be described. In the third variation, a method will be described in which, when the display control device 100 displays a conversation between avatars in a virtual space, not only is text indicating the conversation content displayed in a typical speech bubble, but the conversation content is expressed in a variety of ways.
 例えば、第3のバリエーションにおいて、表示制御部133は、入力情報に基づいてアバター10が発する言葉を表示するとともに、言葉の履歴がアバターを起点として上空方向に遷移するよう表示してもよい。 For example, in a third variation, the display control unit 133 may display words uttered by the avatar 10 based on the input information, and may also display a history of words that transitions upward from the avatar.
 また、表示制御部133は、入力情報に含まれる意味内容に基づいて、アバター10が発する言葉を、意味内容を絵で表したアイコンで表示してもよい。 In addition, the display control unit 133 may display the words uttered by the avatar 10 as icons that pictorially represent the meanings contained in the input information.
 また、判定部132は、言葉の表示を観測する仮想空間の位置(すなわち、観測者であるアバターが所在する位置)と、言葉を発するアバター10の位置との距離を判定してもよい。この場合、表示制御部133は、判定された距離に基づいて、言葉の履歴もしくはアイコンを、言葉およびアイコンの意味内容を含まない演出効果として表示する。なお、意味内容を含まない演出効果の一例としては、後述するように、会話内容を示すテキスト等が表示されず、単に盛んに会話が交わされた履歴が狼煙やバルーンのように表現されることが該当する。 The determination unit 132 may also determine the distance between the position in the virtual space where the display of the words is observed (i.e., the position where the observer avatar is located) and the position of the avatar 10 that utters the words. In this case, the display control unit 133 displays the word history or icons as a dramatic effect that does not include the meaning of the words and icons, based on the determined distance. Note that, as will be described later, an example of a dramatic effect that does not include meaning is when no text or the like indicating the content of a conversation is displayed, and a history of lively conversation is simply displayed as a smoke signal or balloon.
 また、判定部132は、言葉の表示を観測する仮想空間の位置と、言葉を発するアバター10の位置との位置関係を判定してもよい。この場合、表示制御部133は、判定された位置関係に基づいて、仮想空間で観測される言葉の表示の角度および3次元表示を決定してもよい。 The determination unit 132 may also determine the positional relationship between the position in the virtual space where the display of the words is observed and the position of the avatar 10 that speaks the words. In this case, the display control unit 133 may determine the angle and three-dimensional display of the display of the words observed in the virtual space based on the determined positional relationship.
 上記のように、表示制御装置100は、第3のバリエーションにおいて、テキスト等で表示される一般的な会話でなく、アイコン表示を行ったり、視覚効果を含む会話表示をしたりする。これにより、表示制御装置100は、アバター間で盛んに会話が行われていること等を一目で観測可能に表現できる。 As described above, in the third variation, the display control device 100 displays an icon or a conversation that includes visual effects, rather than a typical conversation displayed as text, etc. This allows the display control device 100 to display, in an easily observable manner, an active conversation taking place between avatars, etc.
 第3のバリエーションについて、図11以下を用いて具体的に説明する。図11は、実施形態に係る表示制御処理の第3のバリエーションを説明するための図(1)である。 The third variation will be described in detail using FIG. 11 and subsequent figures. FIG. 11 is a diagram (1) for explaining the third variation of the display control process according to the embodiment.
 図11に示す例では、アバター10Aが、観測者であるアバターに対して、「見たい映画があるんだ!」という会話を行っていることを示す。アバター10Aは、観測者の近傍で会話していることから、かかる会話は、吹き出し230にテキストとして表示される。 The example shown in FIG. 11 shows that avatar 10A is having a conversation with the observer avatar, saying, "There's a movie I want to see!" Since avatar 10A is having the conversation in the vicinity of the observer, the conversation is displayed as text in speech bubble 230.
 一方で、アバター10Aよりも遠方で、アバター10Bが会話をしている内容については、アイコン231で示される。具体的には、表示制御装置100は、分析サーバ50によってアバター10Bの会話について内容および感情分析を行い、分析結果をアイコン231として表示する。図11の例では、アイコン231は、アバター10Bが、映画を見に行き、その後で車に乗り、食事に出かけたといった内容の会話を行ったことを示している。 On the other hand, the content of a conversation that avatar 10B is having, which is located farther away than avatar 10A, is shown by icon 231. Specifically, the display control device 100 performs a content and emotion analysis of the conversation of avatar 10B using the analysis server 50, and displays the analysis results as icon 231. In the example of FIG. 11, icon 231 indicates that avatar 10B had a conversation about going to the movies, then getting in a car, and then going out for dinner.
 このように、表示制御装置100は、観測者から所定距離を超えて離れているアバター10Bの会話については、テキストをそのまま表示するのではなく、会話内容を示したアイコン(画像等)で表示する。これにより、観測者は、遠くの会話について、詳細に文字情報を解読しなくても、大まかな会話の内容や、会話がどれくらい活発に行われているかといった状況を把握することができる。 In this way, for conversations between avatars 10B that are farther away from the observer than a predetermined distance, the display control device 100 displays the conversation content as icons (images, etc.) rather than displaying the text as is. This allows the observer to grasp the general content of the conversation and how lively it is, without having to decipher the detailed text information of distant conversations.
 また、表示制御装置100は、アバター間で行われた会話については、アイコン231のように、過去の履歴が上方に流れていくようなアニメーションで表示することができる。これにより、表示制御装置100は、会話の流れを視覚的にわかりやすく他のユーザに伝達することができるとともに、会話がどれくらい活発に行われているかを遠方のユーザからも把握させることができる。 Furthermore, the display control device 100 can display conversations between avatars in an animation in which past history scrolls upward, as in icon 231. This allows the display control device 100 to communicate the flow of the conversation to other users in a visually easy-to-understand manner, and also allows users who are far away to understand how lively the conversation is.
 この点について、図12を用いて詳細に説明する。図12は、実施形態に係る表示制御処理の第3のバリエーションを説明するための図(2)である。 This point will be described in detail with reference to FIG. 12. FIG. 12 is a diagram (2) for explaining a third variation of the display control process according to the embodiment.
 図12には、観測者から極めて遠方で、複数のアバター235が会話をしている様子を示す。このとき、複数のアバター235による会話は、テキストやアイコン等、意味内容を含む表示が省略され、吹き出しや円の形(このような表示は「テキストバルーン」と称される場合がある)をした会話の履歴236のみが表示される。 Figure 12 shows multiple avatars 235 having a conversation at a great distance from the observer. At this time, the conversation between the multiple avatars 235 is displayed without displaying any semantic content such as text or icons, and only the conversation history 236 in the form of speech bubbles or circles (such displays are sometimes called "text balloons").
 上述のように、会話の履歴236は、過去の履歴順に上方にアニメーション表示される。このため、図12の例では、複数のアバター235の会話が、仮想空間上での狼煙のように表現される。これにより、遠方の観測者であっても、仮想空間上のどこで活発に会話が行われているかを知ることができる。 As described above, the conversation history 236 is animated and displayed upwards in the order of the most recent conversations. Therefore, in the example of FIG. 12, the conversations between multiple avatars 235 are expressed like smoke signals in the virtual space. This allows even a distant observer to know where in the virtual space an active conversation is taking place.
 また、表示制御装置100は、観測者とアバター10との距離に応じて、様々に表現手法を変えることができる。この点について、図13を用いて説明する。図13は、実施形態に係る表示制御処理の第3のバリエーションを説明するための図(3)である。 The display control device 100 can also change the expression method in various ways depending on the distance between the observer and the avatar 10. This will be explained using FIG. 13. FIG. 13 is a diagram (3) for explaining a third variation of the display control process according to the embodiment.
 図13には、観測者の近傍での会話240、観測者から中間距離だけ離れて行われる会話241、観測者から極めて遠方で行われる会話242について、それぞれ異なる表現がされている様子を示す。 Figure 13 shows how a conversation taking place near the observer (240), a conversation taking place at a medium distance from the observer (241), and a conversation taking place very far from the observer (242) are each expressed differently.
 例えば、会話240では、アバターの会話内容がテキストで表示され、会話の詳細が表示される。なお、会話240では、例えば、テキストチャット時にユーザが送信した画像など、テキストに限らず、ユーザ間で共有したイメージ情報も表示されてもよい。これにより、観測者は、テキストを読まずとも、どのような話題でコミュニケーションが行われているかを大まかに把握することができる。 For example, in conversation 240, the content of the avatars' conversation is displayed in text, and the details of the conversation are displayed. Note that conversation 240 may also display image information shared between users, not just text, such as images sent by users during text chat. This allows an observer to get a rough idea of what the communication is about, without having to read the text.
 例えば、会話241では、アバターの会話内容についてテキスト表示を省略し、合間にアイコン等を挟むことで、おおまかな会話内容が観測者から把握できるよう表示される。例えば、会話241では、2人のアバターが会話している場合、観測者は、いずれのアバターが発言しているか程度の粒度での会話内容の把握が可能である。なお、会話241のうち古い履歴については、上方にいくほど、互いの台詞が合流するような態様で表示されてもよい。 For example, in conversation 241, the text display of the avatars' conversation is omitted, and icons and the like are inserted between the conversation so that the observer can grasp the general content of the conversation. For example, in conversation 241, when two avatars are conversing, the observer can grasp the conversation content to the extent of which avatar is speaking. Note that older history in conversation 241 may be displayed in such a way that the lines of the avatars merge with each other as they move upward.
 例えば、会話242では、アバターの会話内容についてテキストやアイコン等の意味内容の表示を省略し、狼煙のように会話の履歴のみが表示される。このとき、表示制御装置100は、実空間での狼煙のように、過去の履歴(すなわち、より上方の会話履歴)ほど、透明になるよう表示してもよい。また、表示制御装置100は、個々の台詞を示す表示を、会話240や会話241と比較して小さく表示するようにしてもよい。すなわち、表示制御装置100は、距離に応じて、台詞を表示するための吹き出し等の表示欄を可変することができる。 For example, in conversation 242, the display of the meaning of the avatar's conversation, such as text and icons, is omitted, and only the conversation history is displayed like a smoke signal. At this time, the display control device 100 may display the history more transparently the earlier it is (i.e., the conversation history further up), like a smoke signal in real space. The display control device 100 may also display the display showing each line of dialogue smaller than conversations 240 and 241. In other words, the display control device 100 can change the display area, such as a speech bubble for displaying dialogue, depending on the distance.
 このように、表示制御装置100は、距離に応じて会話の表現を変えることにより、仮想空間で活発なコミュニケーションが行われていても、煩雑な表示にならないようなUIをユーザに提供することができる。 In this way, by changing the way conversation is expressed depending on the distance, the display control device 100 can provide the user with a UI that does not appear cluttered even when lively communication is taking place in the virtual space.
 また、表示制御装置100は、観測者とアバター10との位置関係に応じて、様々に表現手法を変えることができる。この点について、図14を用いて説明する。図14は、実施形態に係る表示制御処理の第3のバリエーションを説明するための図(4)である。 The display control device 100 can also change the expression method in various ways depending on the positional relationship between the observer and the avatar 10. This will be explained using FIG. 14. FIG. 14 is a diagram (4) for explaining a third variation of the display control process according to the embodiment.
 図14の左図には、観測者の近傍で会話250が行われている様子を示す。具体的には、会話250では、アバター10Aとアバター10Bとが会話を行っている。また、会話250は、アバター10Aの会話履歴251や、アバター10Bの会話履歴252を含む。 The left diagram in Figure 14 shows a conversation 250 taking place near the observer. Specifically, in conversation 250, avatar 10A and avatar 10B are having a conversation. Furthermore, conversation 250 includes a conversation history 251 of avatar 10A and a conversation history 252 of avatar 10B.
 図14の右図には、会話250を上空から見た俯瞰表示255を示す。俯瞰表示255に示すように、アバター10Aとアバター10Bは、観測者256から見て正対しておらず、やや斜めの角度で起立しているものとする。 The right diagram in FIG. 14 shows an overhead view 255 of the conversation 250 as seen from above. As shown in the overhead view 255, avatar 10A and avatar 10B are not facing each other as seen by an observer 256, but are standing at a slight angle.
 このとき、表示制御装置100は、図14の左図に示すように、観測者256とアバター10Aとの位置関係(ここでは観測者256の視線とアバター10Aの向きとが成す角度)に基づいて、会話履歴251に角度を付けて表示する。同様に、表示制御装置100は、観測者256とアバター10Bとの位置関係に基づいて、会話履歴252に角度を付けて表示する。 At this time, as shown in the left diagram of FIG. 14, the display control device 100 displays the conversation history 251 at an angle based on the positional relationship between the observer 256 and the avatar 10A (here, the angle between the line of sight of the observer 256 and the direction of the avatar 10A). Similarly, the display control device 100 displays the conversation history 252 at an angle based on the positional relationship between the observer 256 and the avatar 10B.
 かかる表現により、観測者256は、会話履歴を見ただけで、アバター10Aやアバター10Bとの位置関係を認識することができる。 This representation allows the observer 256 to recognize the positional relationship with avatar 10A and avatar 10B simply by looking at the conversation history.
 また、表示制御装置100は、角度のみならず、テキスト等、会話内容そのものに演出効果を含む表示を行ってもよい。この点について、図15を用いて説明する。図15は、実施形態に係る表示制御処理の第3のバリエーションを説明するための図(5)である。 The display control device 100 may also perform display including dramatic effects not only on the angle but also on the content of the conversation itself, such as text. This point will be explained using FIG. 15. FIG. 15 is a diagram (5) for explaining a third variation of the display control process according to the embodiment.
 図15では、アバター10Aとアバター10Bとが会話を行っており、その会話内容を示すテキスト260、演出効果261、テキスト262が表示されている様子を示す。 In FIG. 15, avatar 10A and avatar 10B are having a conversation, and text 260 indicating the content of the conversation, production effect 261, and text 262 are displayed.
 例えば、テキスト260は、他のテキストと比較して太字で表示されるなど、強調表示を含んでもよい。例えば、表示制御装置100は、「ヤバい」など、予め登録されている、感情を強く示す言葉をユーザが発したときに、自動的に強調表示を行う。 For example, the text 260 may include highlighting, such as being displayed in bold compared to other text. For example, the display control device 100 automatically highlights when the user utters a pre-registered word that strongly indicates an emotion, such as "yabai (crazy)."
 また、演出効果261は、3次元表示を含む文字アイコンである。例えば、演出効果261は、「ヤバい」など、予め登録されている、感情を強く示す言葉をユーザが発したときに、自動的に表示される。 The effect 261 is a character icon including a three-dimensional display. For example, the effect 261 is automatically displayed when the user utters a pre-registered word that strongly indicates an emotion, such as "yabai (crazy)."
 また、テキスト262は、3次元表示を含むテキストである。例えば、テキスト262は、「びっくり」など、予め登録されている、感情を強く示す言葉をユーザが発したときや、ユーザの音量が通常時より大きいとき等、所定の判定基準に基づいて表示される。 The text 262 is also text that includes a three-dimensional display. For example, the text 262 is displayed based on a predetermined criterion, such as when the user utters a pre-registered word that strongly indicates an emotion, such as "surprise," or when the user's voice volume is higher than normal.
 なお、これらの特殊なテキスト表示や演出効果は、ユーザの会話における感情分析等に基づいて表示されてもよい。例えば、表示制御装置100は、分析サーバ50と協働して、テキストチャットで入力された内容を、センチメント分析を用いたポジティブ・ネガティブ判定を行う。そして、表示制御装置100は、テキストが表示される吹き出しに対して、適切な視覚効果を自動発生させる。具体的には、表示制御装置100は、会話がポジティブな場合、明るい色や、文字が弾むアニメーション等を伴う吹き出しやテキストを表示したりしてもよい。あるいは、表示制御装置100は、会話がネガティブな場合、吹き出しを冷感のある色で表示したり、テキストが泣いているようなアニメーション等を伴う吹き出しやテキストを表示したりしてもよい。 These special text displays and special effects may be displayed based on, for example, sentiment analysis of user conversations. For example, the display control device 100 cooperates with the analysis server 50 to determine whether the content entered in the text chat is positive or negative using sentiment analysis. The display control device 100 then automatically generates appropriate visual effects for the speech bubbles in which the text is displayed. Specifically, if the conversation is positive, the display control device 100 may display speech bubbles or text in bright colors or with animations of bouncing letters, or the like. Alternatively, if the conversation is negative, the display control device 100 may display speech bubbles in cool colors or with animations of crying text, or the like.
 また、これらの吹き出しやテキストは、会話において登場した単語の頻度等に応じてテキストの大小が異なるような履歴表示(ワードクラウド等と称される)で表現されてもよい。 These speech bubbles and text may also be displayed in a history display (known as a word cloud, for example) in which the size of the text varies depending on the frequency of the words that appeared in the conversation.
 また、これらのテキスト表示等は、3次元の物理特性を伴ってもよい。この点について、図16を用いて説明する。図16は、実施形態に係る表示制御処理の第3のバリエーションを説明するための図(6)である。 These text displays and the like may also have three-dimensional physical characteristics. This will be explained using FIG. 16. FIG. 16 is a diagram (6) for explaining a third variation of the display control process according to the embodiment.
 図16には、3次元表示を伴うテキスト270と、テキスト275とを例示する。テキスト270は、文字が吹き出しから浮いたような形状で構成されている。このため、テキスト270を横から見た際の表示271では、文字部分が吹き出しから浮いたように表現される。 FIG. 16 shows examples of text 270 and text 275 with three-dimensional display. Text 270 is configured so that the characters appear to be floating from the speech bubble. Therefore, in display 271 when text 270 is viewed from the side, the characters appear to be floating from the speech bubble.
 また、テキスト275は、文字が吹き出しの中に取り込まれたような形状で構成されている。このため、テキスト275を横から見た際の表示276では、文字部分が雲のような吹き出しの中に含まれるように表現される。このため、観測者からは、文字が2次元のプレートとして表現され、吹き出しが3次元の雲やバルーンのような形状で表現されたように見える。 The text 275 is also configured in such a way that the characters appear to be contained within a speech bubble. For this reason, when the text 275 is viewed from the side, the display 276 shows the characters as if they were contained within a cloud-like speech bubble. For this reason, to the observer, the characters appear to be displayed as two-dimensional plates, and the speech bubble appears to be displayed in a three-dimensional cloud or balloon-like shape.
 なお、図16等で示した3次元表示は裏面表示も可能である。この場合、3次元表示における画像やテキストは、左右反転されて表示される。 The 3D display shown in Figure 16 etc. can also be displayed backwards. In this case, the images and text in the 3D display are displayed with the left and right reversed.
 このように、表示制御装置100は、会話におけるテキスト等についても多彩な表現を行うことができ、ユーザを視覚的に楽しませることができる。 In this way, the display control device 100 can express text in conversations in a variety of ways, providing visual entertainment to the user.
(1-4-4.アバター同士の近接効果)
 続いて、第4のバリエーションについて説明する。第4のバリエーションでは、表示制御装置100が、アバター同士が近接した場合に交流を促すような表現を行う例を示す。
(1-4-4. Proximity effect between avatars)
Next, a fourth variation will be described. In the fourth variation, an example will be shown in which the display control device 100 performs an expression that encourages interaction when avatars come close to each other.
 例えば、第4のバリエーションにおいて、判定部132は、第1のアバターが第2のアバターにより認識される範囲内に含まれるかを判定してもよい。そして、表示制御部133は、第1のアバターが第2のアバターにより認識される範囲内に含まれると判定された場合、第1のアバターが第2のアバターに近接するよう、第1のアバターの移動経路を調整してもよい。 For example, in a fourth variation, the determination unit 132 may determine whether the first avatar is included within a range recognized by the second avatar. Then, when it is determined that the first avatar is included within a range recognized by the second avatar, the display control unit 133 may adjust the movement path of the first avatar so that the first avatar approaches the second avatar.
 また、判定部132は、仮想空間において、第1のアバターが第2のアバターに近接したか否かを判定してもよい。表示制御部133は、第1のアバターが第2のアバターに近接したと判定された場合に、第1のアバターおよび第2のアバターの少なくとも1つに、近接に対応する特定の動作を反映する。 The determination unit 132 may also determine whether or not the first avatar has approached the second avatar in the virtual space. When it is determined that the first avatar has approached the second avatar, the display control unit 133 reflects a specific action corresponding to the approach in at least one of the first avatar and the second avatar.
 すなわち、第4のバリエーションにおいて、表示制御装置100は、アバター同士が互いに存在を認識したり、互いに近接してコミュニケーションが図れたりするよう、特殊な表示や作用を働かせることで、ユーザ同士の交流を促す。 In other words, in the fourth variation, the display control device 100 encourages interaction between users by using special displays and actions so that avatars can recognize each other's presence and communicate with each other in close proximity.
 第4のバリエーションについて、図17以下を用いて具体的に説明する。図17は、実施形態に係る表示制御処理の第4のバリエーションを説明するための図(1)である。 The fourth variation will be described in detail using FIG. 17 and subsequent figures. FIG. 17 is a diagram (1) for explaining the fourth variation of the display control process according to the embodiment.
 図17に示す第1図では、アバター10A、アバター10Bおよびアバター10Cが会話している場所に、アバター10Dが近付いてくる様子を示している。このとき、アバター10Aは、初期設定された視点認識によって、もしくは、近接が判定される範囲にアバター10Dが入ったことを認識する。 The first diagram in FIG. 17 shows avatar 10D approaching a location where avatar 10A, avatar 10B, and avatar 10C are having a conversation. At this time, avatar 10A recognizes that avatar 10D has entered a range where proximity is determined, either through initially set viewpoint recognition or by other means.
 アバター10Aは、図17に示す第2図のように、アバター10Dの近接を認識すると、特定の行動を発火させる。例えば、アバター10Aは、手を上げて挨拶を行うジェスチャーを実行する。このとき、表示制御装置100は、アバター10Aがアバター10Dへの会話に参加するよう促すよう、友好的なメッセージ等を自動的に表示してもよい。 When avatar 10A recognizes the proximity of avatar 10D as shown in FIG. 17 (FIG. 2), it initiates a specific action. For example, avatar 10A performs a gesture of greeting by raising its hand. At this time, display control device 100 may automatically display a friendly message or the like to encourage avatar 10A to join the conversation with avatar 10D.
 そして、アバター10Dがその挨拶に応答し、アバター10A等の会話に加わる意図を示すと、図17に示す第3図のように、アバター10Dが会話に参加する。 When avatar 10D responds to the greeting and indicates its intention to join the conversation with avatar 10A, etc., avatar 10D joins the conversation, as shown in FIG. 3 in FIG. 17.
 表示制御装置100は、これらの動作において、アバター10Aとアバター10Dとが視線を合わせるような補正や、アバター10Aとアバター10Dとに挨拶のジェスチャーを行わせるなどする。これにより、表示制御装置100は、アバター同士に積極的に交流をもたせることができる。 In these operations, the display control device 100 performs corrections so that the gaze of the avatar 10A and the avatar 10D meets, and causes the avatar 10A and the avatar 10D to perform a greeting gesture. In this way, the display control device 100 can allow the avatars to actively interact with each other.
 このような自動的な会話参加について、表示制御装置100は、いくつかの条件に基づいて制御処理を行う。例えば、表示制御装置100は、1人または複数のアバターが既に会話(チャット)状態に入っていることを判定する。また、表示制御装置100は、その会話が行われているエリアに、他のアバターが近接したことを判定する。これらの条件が満たされると、表示制御装置100は、アバター同士の視点が合いやすいように表示を補正したり、ジェスチャーを実行させたりする。 For such automatic conversation participation, the display control device 100 performs control processing based on several conditions. For example, the display control device 100 determines that one or more avatars are already in a conversation (chat) state. The display control device 100 also determines that another avatar has approached the area where the conversation is taking place. When these conditions are met, the display control device 100 corrects the display so that the avatars' viewpoints are more easily aligned, or causes them to perform gestures.
 この点について、図18を用いて説明する。図18は、実施形態に係る表示制御処理の第4のバリエーションを説明するための図(2)である。 This point will be explained using FIG. 18. FIG. 18 is a diagram (2) for explaining a fourth variation of the display control process according to the embodiment.
 図18に示すように、表示制御装置100は、アバター10Aの近接エリア280と、アバター10Dとの近接エリア281を認識する。また、表示制御装置100は、アバター10Aとアバター10Dとの距離282を認識する。そして、表示制御装置100は、距離282が所定距離以下になると、例えば、アバター10Aとアバター10Dとが身体を向きあうような補正を行うことで、互いの視線が合うように補正する。 As shown in FIG. 18, the display control device 100 recognizes a proximity area 280 of avatar 10A and a proximity area 281 with avatar 10D. The display control device 100 also recognizes a distance 282 between avatar 10A and avatar 10D. When the distance 282 becomes equal to or smaller than a predetermined distance, the display control device 100 corrects the positions of avatar 10A and avatar 10D so that their lines of sight meet, for example by making a correction so that their bodies face each other.
 その後、表示制御装置100は、アバター10Aの近接エリア280と、アバター10Dとの近接エリア281とが接触したことを認識する。この場合、表示制御装置100は、アバター10Aとアバター10Dとが互いに挨拶するジェスチャーを発火させる。 Then, the display control device 100 recognizes that the proximity area 280 of the avatar 10A and the proximity area 281 of the avatar 10D have come into contact. In this case, the display control device 100 causes the avatar 10A and the avatar 10D to make gestures to greet each other.
 さらに、表示制御装置100は、アバター10Aとアバター10Dとが互いの近接エリア内に入った場合、二人の間にボイスチャット等の会話を開始させる。このとき、表示制御装置100は、アバター10Aとアバター10Dとの会話が行われる会話エリア283を新たに認識する。 Furthermore, when avatar 10A and avatar 10D enter within each other's proximity area, display control device 100 starts a conversation such as a voice chat between the two. At this time, display control device 100 newly recognizes conversation area 283 in which the conversation between avatar 10A and avatar 10D takes place.
 なお、表示制御装置100は、視線を合わせた際や、挨拶のジェスチャーを行った際に、互いのユーザが会話を行うことを拒否する意思を示した場合、互いのアバターを近接させないよう補正してもよい。 In addition, when users make eye contact or make a greeting gesture and indicate their intention to refuse to have a conversation, the display control device 100 may correct the positions of their avatars so that they are not close to each other.
 また、表示制御装置100は、アバター同士が物理的に近接した際には、ボイスによるチャットを優先するような制御をおこなってもよい。これにより、表示制御装置100は、新たに参加したアバターにとって、既に会話に参加している不特定多数のアバターに対して呼びかけやすいような環境を構築することができる。なお、表示制御装置100は、ユーザがボイス入力をミュートにしているような場合、テキストによるチャットを優先するようにしてもよい。 The display control device 100 may also perform control to prioritize voice chat when avatars are physically close to each other. This allows the display control device 100 to create an environment in which a newly joining avatar can easily call out to the unspecified number of avatars already participating in the conversation. The display control device 100 may also prioritize text chat when the user has muted voice input.
 また、表示制御装置100は、アバター同士を近接させて交流を促す制御について、会話を促すエリアと、会話禁止エリアを予め設定しておいてもよい。この点について、図19を用いて説明する。図19は、実施形態に係る表示制御処理の第4のバリエーションを説明するための図(3)である。 Furthermore, the display control device 100 may pre-set areas where conversation is encouraged and areas where conversation is prohibited for control that encourages interaction by bringing avatars close to each other. This point will be explained using FIG. 19. FIG. 19 is a diagram (3) for explaining a fourth variation of the display control process according to the embodiment.
 図19には、仮想空間において何らかのイベントが実施される状況を示す。例えば、イベントは、ステージ290に特定のアバターが登場するものとする。この場合、表示制御装置100は、ステージ290に至るまでの通路やステージ290から離れた位置にアバターが滞留しないよう、通路等に会話禁止エリア292や会話禁止エリア293を設定する。一方、表示制御装置100は、ステージ290の周囲や、モニュメント291の周囲での会話を積極的に促すため、ステージ290の周囲等に会話エリア294を設定する。 FIG. 19 shows a situation in which some kind of event is taking place in a virtual space. For example, the event involves a specific avatar appearing on stage 290. In this case, the display control device 100 sets conversation-prohibited areas 292 and conversation-prohibited areas 293 in the passageways leading to stage 290 or in locations away from stage 290, so that the avatar does not remain in the passageways leading to stage 290 or in locations away from stage 290. On the other hand, the display control device 100 sets conversation areas 294 around stage 290, etc., to actively encourage conversation around stage 290 and around monument 291.
 この場合、表示制御装置100は、会話禁止エリア292や会話禁止エリア293ではアバター同士が会話を開始するための補正を行わない。一方、表示制御装置100は、会話エリア294においては、アバター同士が会話を開始するための補正を行う。これにより、表示制御装置100は、仮想空間において会話が推奨されるエリアにおけるアバター同士の会話を積極的に誘導することができる。 In this case, the display control device 100 does not make any corrections to allow avatars to start a conversation in the conversation prohibited area 292 or the conversation prohibited area 293. On the other hand, the display control device 100 makes corrections to allow avatars to start a conversation in the conversation area 294. This allows the display control device 100 to actively guide conversations between avatars in areas in the virtual space where conversation is encouraged.
 なお、表示制御装置100は、アバター10Dが会話に参加しやすくするよう、アバター10Dの移動経路を補正してもよい。例えば、表示制御装置100は、アバター10Dが歩行しようとする経路が、アバター10Aの近接エリアや、アバター10A等が会話をしているエリアに自動的に近付くよう、歩行経路や歩行速度を補正してもよい。また、表示制御装置100は、歩行中、アバター10Dがアバター10Aと視線を合わせやすくするよう、身体の向きを補正してもよい。 The display control device 100 may correct the movement path of the avatar 10D so that the avatar 10D can easily participate in the conversation. For example, the display control device 100 may correct the walking path or walking speed so that the path along which the avatar 10D is walking automatically approaches an area close to the avatar 10A or an area where the avatars 10A and the like are having a conversation. The display control device 100 may also correct the body orientation so that the avatar 10D can easily make eye contact with the avatar 10A while walking.
(1-4-5.アバターの表示や交流に関する表現)
 続いて、第5のバリエーションについて説明する。第5のバリエーションでは、表示制御装置100が、アバター10の容姿を変更したり、アバター同士で容姿を変更するためのアイテムを交換する際のUIを表示したりする例を示す。
(1-4-5. Expressions related to avatar display and interaction)
Next, a fifth variation will be described. In the fifth variation, an example is shown in which the display control device 100 changes the appearance of the avatar 10 and displays a UI for exchanging items for changing the appearance between avatars.
 例えば、第5のバリエーションでは、取得部131は、アバター10の外観を変更する旨の要求をユーザから取得する。そして、表示制御部133は、要求に基づく変更をアバター10に反映して、演出効果とともに仮想空間においてアバター10の外観を変更する。 For example, in the fifth variation, the acquisition unit 131 acquires a request from the user to change the appearance of the avatar 10. The display control unit 133 then reflects the change based on the request in the avatar 10, changing the appearance of the avatar 10 in the virtual space along with a dramatic effect.
 また、取得部131は、第1のアバターに紐付けられている複数の物体のいずれかを第2のアバターと交換する要求をユーザから取得してもよい。この場合、表示制御部133は、要求が取得されると、第1のアバターに紐付けられている複数の物体を第1のアバターの近傍に一覧表示する。 The acquisition unit 131 may also acquire from the user a request to exchange any one of a plurality of objects linked to the first avatar with a second avatar. In this case, when the request is acquired, the display control unit 133 displays a list of the plurality of objects linked to the first avatar near the first avatar.
 また、取得部131は、一覧表示された複数の物体のうち、第2のアバターと交換するいずれかの物体を選択する指示をユーザから取得してもよい。この場合、表示制御部133は、指示が取得されると、いずれかの物体を一覧表示から浮遊させるよう表示させるとともに、ユーザから受け付ける操作に従い、仮想空間において当該浮遊させた物体を第2のアバターに重なるよう移動させる。 The acquisition unit 131 may also acquire an instruction from the user to select one of the objects displayed in a list to be exchanged for the second avatar. In this case, when the display control unit 133 acquires the instruction, it displays one of the objects so that it floats from the list display, and moves the floating object in the virtual space so that it overlaps with the second avatar in accordance with the operation received from the user.
 このように、表示制御装置100は、第5のバリエーションにおいて、アバター10の外観を変更したり、アバター10同士のアイテム交換等を積極的に促したりするような表示制御を行う。これにより、表示制御装置100は、ユーザを視覚的に楽しませたり、ユーザ同士の交流を積極的に図ったりすることができる。 In this way, in the fifth variation, the display control device 100 performs display control such as changing the appearance of the avatar 10 and actively encouraging the exchange of items between avatars 10. This allows the display control device 100 to provide visual entertainment for the user and actively encourage interaction between users.
 第5のバリエーションについて、図20以下を用いて具体的に説明する。図20は、実施形態に係る表示制御処理の第5のバリエーションを説明するための図(1)である。 The fifth variation will be described in detail using FIG. 20 and subsequent figures. FIG. 20 is a diagram (1) for explaining the fifth variation of the display control process according to the embodiment.
 図20の第1図では、ユーザがアバター10Aの容姿を変更させるため、変更先の候補300をUI表示させている例を示す。この例では、ユーザは、候補301を選択したものとする。 In the first diagram of FIG. 20, an example is shown in which a UI displays candidate options 300 for changing the appearance of avatar 10A. In this example, it is assumed that the user has selected candidate 301.
 図20の第2図では、表示制御装置100が、アバター10Aの容姿を変更させている様子を示す。表示制御装置100は、アバター10Aの容姿を変更させる際に、アバター10Aが煙に覆われるような演出効果302を表示してもよい。 The second diagram in FIG. 20 shows the display control device 100 changing the appearance of the avatar 10A. When changing the appearance of the avatar 10A, the display control device 100 may display a dramatic effect 302 in which the avatar 10A appears to be covered in smoke.
 図20の第3図では、表示制御装置100が、アバター10Aの容姿を変更した様子を示す。表示制御装置100は、変更後のアバター10Aの容姿を仮想空間上に表示する。このように、ユーザは、アバター10Aの容姿を任意に変更することができる。 FIG. 3 in FIG. 20 shows the state in which the display control device 100 has changed the appearance of the avatar 10A. The display control device 100 displays the changed appearance of the avatar 10A in the virtual space. In this way, the user can arbitrarily change the appearance of the avatar 10A.
 仮想空間において、ユーザは、アバター10の容姿を変更するためのアイテムを他のユーザと交換することができる。この点について、図21を用いて説明する。図21は、実施形態に係る表示制御処理の第5のバリエーションを説明するための図(2)である。 In the virtual space, a user can exchange items with other users to change the appearance of the avatar 10. This will be explained using FIG. 21. FIG. 21 is a diagram (2) for explaining a fifth variation of the display control process according to the embodiment.
 図21には、アバター10Aと観測者とがアイテム交換を行う際のUI表示を示す。アバター10Aは、交換候補アイテム310を観測者に提示する。また、観測者も、交換候補アイテム311を提示する。そして、観測者が交換ボタン312を選択すると、互いのアイテムが交換される。なお、仮想空間では、各アイテムがどのユーザ(アバター)に所有されているかを示す所有権の概念がある場合がある。表示制御装置100は、アイテムの受け渡しとともに、かかる所有権についても交換がなされるよう、内部処理を行ってもよい。 FIG. 21 shows the UI display when avatar 10A and an observer exchange items. Avatar 10A presents exchange candidate items 310 to the observer. The observer also presents exchange candidate items 311. When the observer selects exchange button 312, the items are exchanged. Note that in virtual space, there may be a concept of ownership that indicates which user (avatar) owns each item. The display control device 100 may perform internal processing so that such ownership is also exchanged when the items are handed over.
 このようなアバター同士のアイテム交換は、さらに異なるUIで表示されてもよい。この点について、図22を用いて説明する。図22は、実施形態に係る表示制御処理の第5のバリエーションを説明するための図(3)である。 Such item exchanges between avatars may be displayed in a different UI. This will be explained with reference to FIG. 22. FIG. 22 is a diagram (3) for explaining a fifth variation of the display control process according to the embodiment.
 図22には、アバター10Aと、アバター10Bとがアイテム320を交換しようとしている様子を示す。この場合、アバター10Aに対応するユーザは、画面表示において、マウス等のポインティングデバイスを用いて、アイテム320を選択する。 FIG. 22 shows a state in which avatar 10A and avatar 10B are about to exchange item 320. In this case, the user corresponding to avatar 10A selects item 320 on the screen display using a pointing device such as a mouse.
 すると、表示制御装置100は、アイテム320周囲の表示を拡大して表示する。拡大表示において、ユーザは、実空間で腕をうごかすなどして、アイテム320をアバター10Bに近づけることができる。この場合、表示制御装置100は、ユーザの腕に対応したアバター10Aのボーン表示321を表示し、アバター10Aの腕付近に表示されたアイテム320をアバター10B側に移動させる。 Then, the display control device 100 enlarges the display around the item 320. In the enlarged display, the user can move the item 320 closer to the avatar 10B by, for example, moving their arm in real space. In this case, the display control device 100 displays a bone display 321 of the avatar 10A that corresponds to the user's arm, and moves the item 320 displayed near the arm of the avatar 10A toward the avatar 10B.
 アイテム320がアバター10Bのボーン表示322に対して所定距離より近づけられると、表示制御装置100は、アイテム320の譲渡が行われたと判定する。そして、表示制御装置100は、アイテム320をボーン表示322の近傍に表示するとともに、アイテム320の所有権をアバター10Bに移行する。 When the item 320 is brought closer than a predetermined distance to the bone display 322 of the avatar 10B, the display control device 100 determines that the item 320 has been transferred. The display control device 100 then displays the item 320 near the bone display 322 and transfers ownership of the item 320 to the avatar 10B.
 このように、表示制御装置100は、アイテム交換を示すUIを表示するのではなく、現実にアバター同士がアイテムを受け渡しているかのような表現で、アイテム交換を成立させることができる。 In this way, the display control device 100 can complete the item exchange in a way that makes it appear as if avatars are actually handing over items to each other, rather than displaying a UI that indicates the item exchange.
 また、上述のように、アバター10の容姿を変更する目的等にアイテムが用いられる場合、表示制御装置100は、変更する箇所に対応したUI表示を行ってもよい。この点について、図23を用いて説明する。図23は、実施形態に係る表示制御処理の第5のバリエーションを説明するための図(4)である。 Furthermore, as described above, when an item is used for the purpose of changing the appearance of the avatar 10, the display control device 100 may display a UI corresponding to the part to be changed. This point will be explained using FIG. 23. FIG. 23 is a diagram (4) for explaining a fifth variation of the display control process according to the embodiment.
 図23に示すように、アバター10Aには、容姿を変更する箇所として、頭部330、顔部332、身体部334、足部338が設定されている。この場合、ユーザは、各箇所において変更可能なアイテムを一覧表示させることができる。例えば、表示制御装置100は、頭部330に対応するアイテム候補331を表示する。同様に、表示制御装置100は、顔部332に対応するアイテム候補333、身体部334に対応するアイテム候補335、足部338に対応するアイテム候補337を表示する。また、表示制御装置100は、アバター10A全体の容姿を変更するためのアイテム候補339を表示することもできる。 As shown in FIG. 23, the avatar 10A has a head 330, a face 332, a body 334, and feet 338 set as parts of the appearance that can be changed. In this case, the user can display a list of items that can be changed in each part. For example, the display control device 100 displays an item candidate 331 that corresponds to the head 330. Similarly, the display control device 100 displays an item candidate 333 that corresponds to the face 332, an item candidate 335 that corresponds to the body 334, and an item candidate 337 that corresponds to the feet 338. The display control device 100 can also display an item candidate 339 for changing the overall appearance of the avatar 10A.
 すなわち、表示制御装置100は、アバター10Aの容姿変更が要求された箇所ごとに、対象アイテム候補表示などをポップアップさせて表示する。そして、表示制御装置100は、ユーザの操作に従い、アバター10Aの容姿を変更させるためのアイテムを選択する。 In other words, the display control device 100 displays a pop-up display of target item candidates for each part of the avatar 10A where a change in appearance is requested. Then, the display control device 100 selects an item for changing the appearance of the avatar 10A in accordance with the user's operation.
 なお、表示制御装置100は、アイテムのUI表示について、例えば、選択された1つのアイテムの下の階層に存在するアイテムを重畳して表示してもよい。例えば、表示制御装置100は、アバター10Aの顔部332を変更するためのアイテムとして眼鏡が選択された場合、眼鏡カテゴリに含まれる複数のアイテム候補をさらに表示してもよい。この場合、ユーザは、例えば一筆書きの要領で、アイテム候補333の1つのアイテムを選択したあと、その下の階層のアイテムを選択することができる。 Note that, when displaying the UI of an item, the display control device 100 may, for example, superimpose items present in a lower hierarchical level on a selected item. For example, when glasses are selected as an item for changing the face portion 332 of the avatar 10A, the display control device 100 may further display multiple candidate items included in the glasses category. In this case, the user can select one item from the candidate items 333, for example, in a single stroke, and then select an item in the lower hierarchical level.
 また、表示制御装置100は、アイテムの選択に際して、ユーザが用いるデバイスの種類に応じて、適当なショートカットを設定してもよい。例えば、表示制御装置100は、アイテム選択において、マウス等のポインティングデバイスやユーザの身振り等による入力とともに、キーボード等のキー操作による入力を受け付けてもよい。 The display control device 100 may also set an appropriate shortcut depending on the type of device used by the user when selecting an item. For example, when selecting an item, the display control device 100 may accept input by key operation on a keyboard or the like, in addition to input by a pointing device such as a mouse or by the user's gestures.
 また、表示制御装置100は、UI表示を利用したアイテムの受け渡し制御を行ってもよい。この点について、図24を用いて説明する。図24は、実施形態に係る表示制御処理の第5のバリエーションを説明するための図(5)である。 The display control device 100 may also control the delivery of items using a UI display. This will be described with reference to FIG. 24. FIG. 24 is a diagram (5) for explaining a fifth variation of the display control process according to the embodiment.
 図24に示す例では、アバター10Aが有するアイテム候補340のうち、ユーザがアイテム341を選択した様子を示す。例えば、ユーザは、画面に対して指やポインティングデバイスを長押しすることで、アイテム341を選択する。この場合、表示制御装置100は、アイテム341を、アイテム候補340から浮いたような表示(フローティング表示)に移行する。 The example shown in FIG. 24 shows a state in which the user selects item 341 from among item candidates 340 held by avatar 10A. For example, the user selects item 341 by pressing and holding a finger or a pointing device on the screen. In this case, the display control device 100 transitions item 341 to a display in which it appears to be floating above item candidates 340 (floating display).
 ユーザは、フローティング表示されたアイテム341をアバター10Bにドラッグする。かかる操作により、表示制御装置100は、アイテム341がアバター10Bに受け渡されたと判定する。このように、表示制御装置100は、様々なUIや画面表示を用いて、ユーザ同士のアイテム交換等の処理を行うことができる。なお、実施形態では、仮想空間でアバター同士が交換するものをアイテムと表記したが、アイテムがどのような態様をとるかは、仮想空間を提供するサービス側によって任意に設定可能である。例えば、アイテムは、NFT(Non-Fungible Token)等のデータであってもよい。 The user drags the floating item 341 to the avatar 10B. With this operation, the display control device 100 determines that the item 341 has been handed over to the avatar 10B. In this way, the display control device 100 can perform processes such as item exchange between users using various UIs and screen displays. Note that in the embodiment, the things exchanged between avatars in the virtual space are described as items, but the form that the items take can be arbitrarily set by the service side providing the virtual space. For example, the items may be data such as NFT (Non-Fungible Token).
(1-5.変形例)
 上記実施形態に係る処理は、様々な変形を伴ってもよい。例えば、図3における各々の装置は、表示制御システム1における機能を概念的に示すものであり、実施形態によって様々な態様をとりうる。
(1-5. Modified Examples)
The processing according to the above embodiment may involve various modifications. For example, each device in Fig. 3 conceptually shows a function in the display control system 1, and may take various forms depending on the embodiment.
 例えば、上記実施形態では、表示制御装置100が分析サーバ50と協働して処理を行う例を示した。しかし、表示制御装置100は、分析サーバ50が実行する処理を自装置で実行してもよい。 For example, in the above embodiment, an example was shown in which the display control device 100 performs processing in cooperation with the analysis server 50. However, the display control device 100 may also perform the processing performed by the analysis server 50 on its own device.
 また、表示制御装置100は、複数の装置で構成されてもよい。例えば、表示制御装置100は、ユーザの挙動を画像として取得し、取得した画像に関する処理を行う第1の情報処理装置と、ユーザの音声やテキスト等の入力に関する処理を行う第2の情報処理装置とに分けられてもよい。この場合、第1の情報処理装置は、ユーザの挙動を画像として取得し、画像に基づいて姿勢推定を行い、手上げ判定、頭向き判定、ハンドサイン判定等を行う。また、第2の情報処理装置は、音声に基づいて笑い声判定を行ったり、テキストに基づいて感情分析等を行ったりしてもよい。 The display control device 100 may also be composed of multiple devices. For example, the display control device 100 may be divided into a first information processing device that acquires the user's behavior as an image and processes the acquired image, and a second information processing device that processes input of the user's voice, text, etc. In this case, the first information processing device acquires the user's behavior as an image, performs posture estimation based on the image, and performs hand-raising determination, head direction determination, hand sign determination, etc. The second information processing device may also perform laughter determination based on voice and emotion analysis based on text.
 また、表示制御装置100は、ディスプレイ部と情報処理部とが別々に構成される装置であってもよい。この場合、表示制御装置100の情報処理部は、サーバやPCなど任意の情報処理装置であってもよい。 The display control device 100 may also be a device in which the display unit and the information processing unit are configured separately. In this case, the information processing unit of the display control device 100 may be any information processing device such as a server or a PC.
(2.その他の実施形態)
 上述した各実施形態に係る処理は、上記各実施形態以外にも種々の異なる形態にて実施されてよい。
2. Other Embodiments
The processing according to each of the above-described embodiments may be implemented in various different forms other than the above-described embodiments.
 また、上記各実施形態において説明した各処理のうち、自動的に行われるものとして説明した処理の全部または一部を手動的に行うこともでき、あるいは、手動的に行われるものとして説明した処理の全部または一部を公知の方法で自動的に行うこともできる。この他、上記文書中や図面中で示した処理手順、具体的名称、各種のデータやパラメータを含む情報については、特記する場合を除いて任意に変更することができる。例えば、各図に示した各種情報は、図示した情報に限られない。 Furthermore, among the processes described in each of the above embodiments, all or part of the processes described as being performed automatically can be performed manually, or all or part of the processes described as being performed manually can be performed automatically using known methods. In addition, the information including the processing procedures, specific names, various data and parameters shown in the above documents and drawings can be changed as desired unless otherwise specified. For example, the various information shown in each drawing is not limited to the information shown in the drawings.
 また、図示した各装置の各構成要素は機能概念的なものであり、必ずしも物理的に図示の如く構成されていることを要しない。すなわち、各装置の分散・統合の具体的形態は図示のものに限られず、その全部または一部を、各種の負荷や使用状況などに応じて、任意の単位で機能的または物理的に分散・統合して構成することができる。 Furthermore, each component of each device shown in the figure is a functional concept, and does not necessarily have to be physically configured as shown in the figure. In other words, the specific form of distribution and integration of each device is not limited to that shown in the figure, and all or part of them can be functionally or physically distributed and integrated in any unit depending on various loads, usage conditions, etc.
 また、上述してきた各実施形態および変形例は、処理内容を矛盾させない範囲で適宜組み合わせることが可能である。 Furthermore, the above-mentioned embodiments and variations can be combined as appropriate to the extent that they do not cause any contradictions in the processing content.
 また、本明細書に記載された効果はあくまで例示であって限定されるものでは無く、他の効果があってもよい。 Furthermore, the effects described in this specification are merely examples and are not limiting, and other effects may also be present.
(3.本開示に係る表示制御装置の効果)
 上述のように、本開示に係る表示制御装置(実施形態では表示制御装置100)は、取得部(実施形態では取得部131)と、判定部(実施形態では判定部132)と、表示制御部(実施形態では表示制御部133)とを備える。取得部は、実空間におけるユーザの挙動を入力情報として取得する。判定部は、予め登録された特定の動作に対応する入力が入力情報に含まれるか否かを判定する。表示制御部は、入力情報に対応するユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する。
(3. Effects of the display control device according to the present disclosure)
As described above, the display control device according to the present disclosure (the display control device 100 in the embodiment) includes an acquisition unit (the acquisition unit 131 in the embodiment), a determination unit (the determination unit 132 in the embodiment), and a display control unit (the display control unit 133 in the embodiment). The acquisition unit acquires a user's behavior in the real space as input information. The determination unit determines whether or not the input information includes an input corresponding to a specific action registered in advance. The display control unit reflects the user's behavior corresponding to the input information in an avatar in a virtual space, and when it is determined that the input information includes an input corresponding to a specific action, the display control unit further reflects the specific action in the avatar in the virtual space and displays the avatar in the virtual space.
 このように、本開示に係る表示制御装置は、ユーザの自然な動きに伴う無意識的動作と、ジェスチャーを発動させるための意識的動作とを組み合わせて、アバターを表示する。これにより、表示制御装置は、ユーザの動きをそのまま反映させるだけでなく、漫画的表現等も伴った、多様な表現で仮想空間のアバターを表示することができる。 In this way, the display control device according to the present disclosure displays an avatar by combining unconscious actions that accompany the user's natural movements with conscious actions to activate gestures. This allows the display control device to display an avatar in a virtual space not only in a way that directly reflects the user's movements, but also in a variety of expressions, including cartoon-like expressions, etc.
 また、取得部は、ユーザから特定の動作をアバターに反映するか否かの指定を受け付ける。判定部は、ユーザの指定がある場合に、特定の動作に対応する入力が入力情報に含まれるか否かを判定する。 The acquisition unit also receives a designation from the user as to whether or not a specific action should be reflected in the avatar. If there is a designation from the user, the determination unit determines whether or not the input information includes an input corresponding to the specific action.
 また、取得部は、ユーザの挙動が撮像される際の撮像装置からの位置を取得する。判定部は、撮像装置からの位置が所定距離より離れている場合には、特定の動作に対応する入力が入力情報に含まれるか否かを判定せず、当該撮像装置からの位置が所定距離以内である場合に、当該特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する。 The acquisition unit also acquires the position from the imaging device when the user's behavior is captured. If the position from the imaging device is farther than a predetermined distance, the determination unit does not determine whether the input information includes an input corresponding to a specific action, and if the position from the imaging device is within the predetermined distance, the determination unit determines whether the input information includes an input corresponding to the specific action.
 このように、本開示に係る表示制御装置は、意識的動作を行うことをユーザから指定されたり、特定のエリアでユーザのジェスチャーが行われたりした際に、意識的動作をアバターに反映する。これにより、表示制御装置は、ユーザの意図に沿ったアバター表現を行うことができる。 In this way, the display control device according to the present disclosure reflects the conscious action in the avatar when the user specifies that a conscious action should be performed or when the user makes a gesture in a specific area. This allows the display control device to display the avatar in accordance with the user's intention.
 また、取得部は、複数のユーザの挙動を入力情報として取得する。判定部は、複数のユーザの挙動に基づいて、仮想空間における当該複数のユーザ同士の交流の盛り上がり度を判定する。表示制御部は、複数のユーザ同士の近傍に、交流の盛り上がり度に応じた演出効果を表示する。 The acquisition unit also acquires the behavior of the multiple users as input information. The determination unit determines the level of excitement of the interaction between the multiple users in the virtual space based on the behavior of the multiple users. The display control unit displays a presentation effect in the vicinity of the multiple users according to the level of excitement of the interaction.
 また、判定部は、交流に参加するユーザの人数、ユーザ同士の会話の頻度、複数のユーザ同士の音声の重なり具合に基づいて、交流の盛り上がり度を判定する。 The determination unit also determines the level of excitement in the exchange based on the number of users participating in the exchange, the frequency of conversation between users, and the degree of overlap in the voices of multiple users.
 また、表示制御部は、交流の盛り上がり度、アバターの身体の向き、複数のユーザの会話の内容に基づいて、演出効果の表現もしくは発生頻度を決定する。 The display control unit also determines the expression or frequency of occurrence of the dramatic effect based on the level of excitement in the interaction, the orientation of the avatar's body, and the content of the conversation between multiple users.
 このように、本開示に係る表示制御装置は、アバター同士の会話の盛り上がりを視覚的に表示することで、会話に参加していないユーザに対しても、アバター同士が活発に交流している様子を効果的に見せることができる。 In this way, the display control device according to the present disclosure can visually display the excitement of a conversation between avatars, effectively showing even users who are not participating in the conversation how the avatars are actively interacting with each other.
 また、判定部は、特定の感情に紐付いた特定の動作に対応する入力が、入力情報に含まれるか否かを判定する。表示制御部は、入力情報に特定の感情に紐付いた特定の動作が含まれると判定された場合には、当該特定の感情に紐付いた特定の動作をアバターに反映して、仮想空間に当該アバターを表示する。 The determination unit also determines whether or not the input information includes an input corresponding to a specific action linked to a specific emotion. If it is determined that the input information includes a specific action linked to a specific emotion, the display control unit reflects the specific action linked to the specific emotion in the avatar and displays the avatar in the virtual space.
 また、判定部は、予めユーザから登録された声紋情報であって、特定の感情に紐付く声紋情報と、入力情報とを比較することで、特定の感情に紐付いた特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する。 The determination unit also compares the voiceprint information registered in advance by the user, which is associated with a specific emotion, with the input information to determine whether the input information includes an input corresponding to a specific action associated with the specific emotion.
 また、表示制御部は、特定の感情に紐付いた特定の動作を反映させたアバターを表示するとともに、当該特定の感情に紐付いた演出効果を当該アバターの近傍に表示する。 The display control unit also displays an avatar that reflects a specific action linked to a specific emotion, and displays a dramatic effect linked to the specific emotion near the avatar.
 このように、本開示に係る表示制御装置は、笑い声など特定の感情に結びついたジェスチャーをアバターに反映することができる。これにより、表示制御装置は、現実空間のように多彩な感情表現を仮想空間においても表現することができる。 In this way, the display control device according to the present disclosure can reflect gestures associated with specific emotions, such as laughter, in an avatar. This allows the display control device to express a wide variety of emotions in virtual space, just as it does in real space.
 また、表示制御部は、入力情報に基づいてアバターが発する言葉を表示するとともに、当該言葉の履歴が当該アバターを起点として上方向に遷移するよう表示する。 The display control unit also displays the words spoken by the avatar based on the input information, and displays the history of those words in an upward transition starting from the avatar.
 また、表示制御部は、入力情報に含まれる意味内容に基づいて、アバターが発する言葉を、当該意味内容を絵で表したアイコンで表示する。 The display control unit also displays the words uttered by the avatar as icons that pictorially represent the meaning contained in the input information.
 また、判定部は、言葉の表示を観測する仮想空間の位置と、当該言葉を発するアバターの位置との距離を判定する。表示制御部は、判定された距離に基づいて、言葉の履歴もしくはアイコンを、当該言葉およびアイコンの意味内容を含まない演出効果として表示する。 The determination unit also determines the distance between the position in the virtual space where the display of the words is observed and the position of the avatar uttering the words. Based on the determined distance, the display control unit displays the word history or an icon as a dramatic effect that does not include the meaning of the words and icons.
 また、判定部は、言葉の表示を観測する仮想空間の位置と、当該言葉を発するアバターの位置との位置関係を判定する。表示制御部は、判定された位置関係に基づいて、仮想空間で観測される言葉の表示の角度および3次元表示を決定する。 The determination unit also determines the positional relationship between the position in the virtual space where the display of the words is observed and the position of the avatar uttering the words. The display control unit determines the angle and three-dimensional display of the display of the words observed in the virtual space based on the determined positional relationship.
 このように、本開示に係る表示制御装置は、会話の履歴を様々な態様で表現することで、画面表示を煩雑にせずに、かつ、どこで活発なコミュニケーションが行われているかをユーザが一目で把握することができる表示を行うことができる。 In this way, the display control device according to the present disclosure can present the conversation history in various ways, without cluttering the screen display, and allows the user to see at a glance where active communication is taking place.
 また、判定部は、仮想空間において、第1のアバターが第2のアバターにより認識される範囲内に含まれるかを判定する。表示制御部は、第1のアバターが第2のアバターにより認識される範囲内に含まれると判定された場合、当該第1のアバターが当該第2のアバターに近接するよう、当該第1のアバターの移動経路を調整する。 The determination unit also determines whether the first avatar is included within a range recognized by the second avatar in the virtual space. If it is determined that the first avatar is included within a range recognized by the second avatar, the display control unit adjusts the movement path of the first avatar so that the first avatar approaches the second avatar.
 また、判定部は、仮想空間において、第1のアバターが第2のアバターに近接したか否かを判定する。表示制御部は、第1のアバターが第2のアバターに近接したと判定された場合に、当該第1のアバターおよび当該第2のアバターの少なくとも1つに近接に対応する特定の動作を反映して、当該第1のアバターおよび当該第2のアバターを表示する。 The determination unit also determines whether or not the first avatar has approached the second avatar in the virtual space. When it is determined that the first avatar has approached the second avatar, the display control unit displays the first avatar and the second avatar while reflecting a specific action corresponding to the approach of at least one of the first avatar and the second avatar.
 このように、本開示に係る表示制御装置は、アバター同士の挙動に補正を加えることで、アバター同士が積極的に交流するよう促すことができる。これにより、表示制御装置は、仮想空間においてユーザ間が会話する機会を増やすことができ、仮想空間上の交流を活性化することができる。 In this way, the display control device according to the present disclosure can correct the behavior of avatars, thereby encouraging the avatars to actively interact with each other. This allows the display control device to increase opportunities for conversations between users in the virtual space, and stimulate interaction in the virtual space.
 また、取得部は、アバターの外観を変更する旨の要求をユーザから取得する。表示制御部は、要求に基づく変更をアバターに反映して、演出効果とともに仮想空間において当該アバターの外観を変更する。 The acquisition unit also acquires a request from the user to change the appearance of the avatar. The display control unit reflects the changes based on the request in the avatar, and changes the appearance of the avatar in the virtual space along with the dramatic effects.
 また、取得部は、第1のアバターに紐付けられている複数の物体のいずれかを第2のアバターと交換する要求をユーザから取得する。表示制御部は、要求が取得されると、第1のアバターに紐付けられている複数の物体を当該第1のアバターの近傍に一覧表示する。 The acquisition unit also acquires a request from the user to exchange any one of the multiple objects linked to the first avatar for the second avatar. When the request is acquired, the display control unit displays a list of the multiple objects linked to the first avatar near the first avatar.
 また、取得部は、一覧表示された複数の物体のうち、第2のアバターと交換するいずれかの物体を選択する指示をユーザから取得する。表示制御部は、指示が取得されると、当該いずれかの物体を一覧表示から浮遊させるよう表示させるとともに、ユーザから受け付ける操作に従い、仮想空間において当該浮遊させた物体を第2のアバターに重なるよう移動させる。 The acquisition unit also acquires an instruction from the user to select one of the objects displayed in the list to be exchanged for the second avatar. When the display control unit acquires the instruction, it displays one of the objects in question so that it floats from the list display, and moves the floating object in the virtual space so that it overlaps with the second avatar in accordance with the operation received from the user.
 このように、本開示に係る表示制御装置は、操作性に優れたUIを提供することで、ユーザ同士のアイテム交換等の交流において、ユーザのストレスを低減させる。これにより、表示制御装置は、ユーザ同士の積極的な交流を支援することができる。 In this way, the display control device according to the present disclosure provides a UI with excellent operability, thereby reducing stress for users when interacting with each other, such as exchanging items. This allows the display control device to support active interaction between users.
(4.ハードウェア構成)
 上述してきた各実施形態に係る表示制御装置100等の情報機器は、例えば図25に示すような構成のコンピュータ1000によって実現される。以下、表示制御装置100を例に挙げて説明する。図25は、表示制御装置100の機能を実現するコンピュータ1000の一例を示すハードウェア構成図である。コンピュータ1000は、CPU1100、RAM1200、ROM(Read Only Memory)1300、HDD(Hard Disk Drive)1400、通信インターフェイス1500、および入出力インターフェイス1600を有する。コンピュータ1000の各部は、バス1050によって接続される。
(4. Hardware Configuration)
Information devices such as the display control device 100 according to each embodiment described above are realized by a computer 1000 having a configuration as shown in Fig. 25, for example. The display control device 100 will be described below as an example. Fig. 25 is a hardware configuration diagram showing an example of a computer 1000 that realizes the functions of the display control device 100. The computer 1000 has a CPU 1100, a RAM 1200, a ROM (Read Only Memory) 1300, a HDD (Hard Disk Drive) 1400, a communication interface 1500, and an input/output interface 1600. Each unit of the computer 1000 is connected by a bus 1050.
 CPU1100は、ROM1300またはHDD1400に格納されたプログラムに基づいて動作し、各部の制御を行う。例えば、CPU1100は、ROM1300またはHDD1400に格納されたプログラムをRAM1200に展開し、各種プログラムに対応した処理を実行する。 The CPU 1100 operates based on the programs stored in the ROM 1300 or the HDD 1400 and controls each component. For example, the CPU 1100 loads the programs stored in the ROM 1300 or the HDD 1400 into the RAM 1200 and executes processes corresponding to the various programs.
 ROM1300は、コンピュータ1000の起動時にCPU1100によって実行されるBIOS(Basic Input Output System)等のブートプログラムや、コンピュータ1000のハードウェアに依存するプログラム等を格納する。 The ROM 1300 stores boot programs such as the Basic Input Output System (BIOS) that is executed by the CPU 1100 when the computer 1000 starts up, as well as programs that depend on the hardware of the computer 1000.
 HDD1400は、CPU1100によって実行されるプログラム、および、かかるプログラムによって使用されるデータ等を非一時的に記録する、コンピュータが読み取り可能な記録媒体である。具体的には、HDD1400は、プログラムデータ1450の一例である、本開示に係る表示制御プログラムを記録する記録媒体である。 HDD 1400 is a computer-readable recording medium that non-temporarily records programs executed by CPU 1100 and data used by such programs. Specifically, HDD 1400 is a recording medium that records a display control program related to the present disclosure, which is an example of program data 1450.
 通信インターフェイス1500は、コンピュータ1000が外部ネットワーク1550(例えばインターネット)と接続するためのインターフェイスである。例えば、CPU1100は、通信インターフェイス1500を介して、他の機器からデータを受信したり、CPU1100が生成したデータを他の機器へ送信したりする。 The communication interface 1500 is an interface for connecting the computer 1000 to an external network 1550 (e.g., the Internet). For example, the CPU 1100 receives data from other devices and transmits data generated by the CPU 1100 to other devices via the communication interface 1500.
 入出力インターフェイス1600は、入出力デバイス1650とコンピュータ1000とを接続するためのインターフェイスである。例えば、CPU1100は、入出力インターフェイス1600を介して、キーボードやマウス等の入力デバイスからデータを受信する。また、CPU1100は、入出力インターフェイス1600を介して、ディスプレイやエッジーやプリンタ等の出力デバイスにデータを送信する。また、入出力インターフェイス1600は、所定の記録媒体(メディア)に記録されたプログラム等を読み取るメディアインターフェイスとして機能してもよい。メディアとは、例えばDVD(Digital Versatile Disc)、PD(Phase change rewritable Disk)等の光学記録媒体、MO(Magneto-Optical disk)等の光磁気記録媒体、テープ媒体、磁気記録媒体、または半導体メモリ等である。 The input/output interface 1600 is an interface for connecting the input/output device 1650 and the computer 1000. For example, the CPU 1100 receives data from an input device such as a keyboard or a mouse via the input/output interface 1600. The CPU 1100 also transmits data to an output device such as a display, edger, or printer via the input/output interface 1600. The input/output interface 1600 may also function as a media interface that reads programs and the like recorded on a specific recording medium. Examples of media include optical recording media such as DVDs (Digital Versatile Discs) and PDs (Phase change rewritable Disks), magneto-optical recording media such as MOs (Magneto-Optical Disks), tape media, magnetic recording media, and semiconductor memories.
 例えば、コンピュータ1000が実施形態に係る表示制御装置100として機能する場合、コンピュータ1000のCPU1100は、RAM1200上にロードされた表示制御プログラムを実行することにより、制御部130等の機能を実現する。また、HDD1400には、本開示に係る表示制御プログラムや、記憶部120内のデータが格納される。なお、CPU1100は、プログラムデータ1450をHDD1400から読み取って実行するが、他の例として、外部ネットワーク1550を介して、他の装置からこれらのプログラムを取得してもよい。 For example, when the computer 1000 functions as the display control device 100 according to the embodiment, the CPU 1100 of the computer 1000 executes a display control program loaded onto the RAM 1200 to realize the functions of the control unit 130, etc. Also, the display control program according to the present disclosure and data in the storage unit 120 are stored in the HDD 1400. The CPU 1100 reads and executes the program data 1450 from the HDD 1400, but as another example, the CPU 1100 may obtain these programs from other devices via the external network 1550.
 なお、本技術は以下のような構成も取ることができる。
(1)
 実空間におけるユーザの挙動を入力情報として取得する取得部と、
 予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する判定部と、
 前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する表示制御部と、
 を備える表示制御装置。
(2)
 前記取得部は、
 前記ユーザから前記特定の動作を前記アバターに反映するか否かの指定を受け付け、
 前記判定部は、
 前記ユーザの指定がある場合に、前記特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する、
 前記(1)に記載の表示制御装置。
(3)
 前記取得部は、
 前記ユーザの挙動が撮像される際の撮像装置からの位置を取得し、
 前記判定部は、
 前記撮像装置からの位置が所定距離より離れている場合には、前記特定の動作に対応する入力が前記入力情報に含まれるか否かを判定せず、当該撮像装置からの位置が所定距離以内である場合に、当該特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する、
 前記(1)または(2)に記載の表示制御装置。
(4)
 前記取得部は、
 複数のユーザの挙動を入力情報として取得し、
 前記判定部は、
 前記複数のユーザの挙動に基づいて、前記仮想空間における当該複数のユーザ同士の交流の盛り上がり度を判定し、
 前記表示制御部は、
 前記複数のユーザ同士の近傍に、前記交流の盛り上がり度に応じた演出効果を表示する、
 前記(1)~(3)のいずれか一つに記載の表示制御装置。
(5)
 前記判定部は、
 前記交流に参加するユーザの人数、ユーザ同士の会話の頻度、複数のユーザ同士の音声の重なり具合に基づいて、前記交流の盛り上がり度を判定する、
 前記(4)に記載の表示制御装置。
(6)
 前記表示制御部は、
 前記交流の盛り上がり度、前記アバターの身体の向き、前記複数のユーザの会話の内容に基づいて、前記演出効果の表現もしくは発生頻度を決定する、
 前記(5)に記載の表示制御装置。
(7)
 前記判定部は、
 特定の感情に紐付いた前記特定の動作に対応する入力が、前記入力情報に含まれるか否かを判定し、
 前記表示制御部は、
 前記入力情報に前記特定の感情に紐付いた特定の動作が含まれると判定された場合には、当該特定の感情に紐付いた特定の動作を前記アバターに反映して、前記仮想空間に当該アバターを表示する、
 前記(1)~(6)のいずれか一つに記載の表示制御装置。
(8)
 前記判定部は、
 予め前記ユーザから登録された声紋情報であって、前記特定の感情に紐付く声紋情報と、前記入力情報とを比較することで、前記特定の感情に紐付いた特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する、
 前記(7)に記載の表示制御装置。
(9)
 前記表示制御部は、
 前記特定の感情に紐付いた特定の動作を反映させた前記アバターを表示するとともに、当該特定の感情に紐付いた演出効果を当該アバターの近傍に表示する、
 前記(7)または(8)に記載の表示制御装置。
(10)
 前記表示制御部は、
 前記入力情報に基づいて前記アバターが発する言葉を表示するとともに、当該言葉の履歴が当該アバターを起点として上方向に遷移するよう表示する、
 前記(1)~(9)のいずれか一つに記載の表示制御装置。
(11)
 前記表示制御部は、
 前記入力情報に含まれる意味内容に基づいて、前記アバターが発する言葉を、当該意味内容を絵で表したアイコンで表示する、
 前記(10)に記載の表示制御装置。
(12)
 前記判定部は、
 前記言葉の表示を観測する仮想空間の位置と、当該言葉を発するアバターの位置との距離を判定し、
 前記表示制御部は、
 前記判定された距離に基づいて、前記言葉の履歴もしくはアイコンを、当該言葉およびアイコンの意味内容を含まない演出効果として表示する、
 前記(11)に記載の表示制御装置。
(13)
 前記判定部は、
 前記言葉の表示を観測する仮想空間の位置と、当該言葉を発するアバターの位置との位置関係を判定し、
 前記表示制御部は、
 前記判定された位置関係に基づいて、仮想空間で観測される前記言葉の表示の角度および3次元表示を決定する、
 前記(11)または(12)に記載の表示制御装置。
(14)
 前記判定部は、
 前記仮想空間において、第1のアバターが第2のアバターにより認識される範囲内に含まれるかを判定し、
 前記表示制御部は、
 前記第1のアバターが前記第2のアバターにより認識される範囲内に含まれると判定された場合、当該第1のアバターが当該第2のアバターに近接するよう、当該第1のアバターの移動経路を調整する、
 前記(1)~(13)のいずれか一つに記載の表示制御装置。
(15)
 前記判定部は、
 前記仮想空間において、前記第1のアバターが前記第2のアバターに近接したか否かを判定し、
 前記表示制御部は、
 前記第1のアバターが前記第2のアバターに近接したと判定された場合に、当該第1のアバターおよび当該第2のアバターの少なくとも1つに前記近接に対応する特定の動作を反映して、当該第1のアバターおよび当該第2のアバターを表示する、
 前記(14)に記載の表示制御装置。
(16)
 前記取得部は、
 前記アバターの外観を変更する旨の要求を前記ユーザから取得し、
 前記表示制御部は、
 前記要求に基づく変更を前記アバターに反映して、演出効果とともに前記仮想空間において当該アバターの外観を変更する、
 前記(1)~(15)のいずれか一つに記載の表示制御装置。
(17)
 前記取得部は、
 第1のアバターに紐付けられている複数の物体のいずれかを第2のアバターと交換する要求を前記ユーザから取得し、
 前記表示制御部は、
 前記要求が取得されると、前記第1のアバターに紐付けられている複数の物体を当該第1のアバターの近傍に一覧表示する、
 前記(1)~(16)のいずれか一つに記載の表示制御装置。
(18)
 前記取得部は、
 前記一覧表示された複数の物体のうち、前記第2のアバターと交換するいずれかの物体を選択する指示を前記ユーザから取得し、
 前記表示制御部は、
 前記指示が取得されると、当該いずれかの物体を一覧表示から浮遊させるよう表示させるとともに、前記ユーザから受け付ける操作に従い、仮想空間において当該浮遊させた物体を前記第2のアバターに重なるよう移動させる、
 前記(17)に記載の表示制御装置。
(19)
 コンピュータが、
 実空間におけるユーザの挙動を入力情報として取得し、
 予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定し、
 前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する、
 ことを含む表示制御方法。
(20)
 コンピュータを、
 実空間におけるユーザの挙動を入力情報として取得する取得部と、
 予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する判定部と、
 前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する表示制御部と、
 を備える表示制御装置として機能させるための表示制御プログラム。
The present technology can also be configured as follows.
(1)
an acquisition unit that acquires user behavior in a real space as input information;
a determination unit that determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance;
a display control unit that reflects a behavior of the user corresponding to the input information in an avatar in a virtual space, and, when it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space;
A display control device comprising:
(2)
The acquisition unit is
accepting a designation from the user as to whether or not the specific action is to be reflected in the avatar;
The determination unit is
determining whether or not an input corresponding to the specific action is included in the input information when the specific action is designated by the user;
The display control device according to (1).
(3)
The acquisition unit is
Acquire a position from an imaging device when the behavior of the user is imaged;
The determination unit is
When the position from the imaging device is farther than a predetermined distance, it is not determined whether or not the input corresponding to the specific action is included in the input information, and when the position from the imaging device is within a predetermined distance, it is determined whether or not the input corresponding to the specific action is included in the input information.
The display control device according to (1) or (2).
(4)
The acquisition unit is
Obtaining the behavior of multiple users as input information,
The determination unit is
determining a degree of excitement of an interaction between the plurality of users in the virtual space based on the behavior of the plurality of users;
The display control unit is
displaying a performance effect in the vicinity of the plurality of users according to the level of excitement of the exchange;
The display control device according to any one of (1) to (3).
(5)
The determination unit is
determining a level of excitement in the exchange based on the number of users participating in the exchange, the frequency of conversations between the users, and the degree of overlap of voices between the multiple users;
The display control device according to (4).
(6)
The display control unit is
determining an expression or occurrence frequency of the performance effect based on the level of excitement of the exchange, the body orientation of the avatar, and the content of the conversation between the plurality of users;
The display control device according to (5) above.
(7)
The determination unit is
determining whether an input corresponding to the specific action associated with a specific emotion is included in the input information;
The display control unit is
when it is determined that the input information includes a specific action associated with the specific emotion, reflecting the specific action associated with the specific emotion in the avatar and displaying the avatar in the virtual space;
A display control device according to any one of (1) to (6).
(8)
The determination unit is
comparing the input information with voiceprint information registered in advance by the user, the voiceprint information being associated with the specific emotion, to determine whether or not the input information includes an input corresponding to a specific action associated with the specific emotion;
The display control device according to (7) above.
(9)
The display control unit is
displaying the avatar reflecting a specific action associated with the specific emotion, and displaying a performance effect associated with the specific emotion in the vicinity of the avatar;
The display control device according to (7) or (8).
(10)
The display control unit is
displaying words uttered by the avatar based on the input information, and displaying a history of the words in an upward transition manner starting from the avatar;
A display control device according to any one of (1) to (9).
(11)
The display control unit is
displaying the words uttered by the avatar as icons pictorially representing the meanings contained in the input information;
The display control device according to (10).
(12)
The determination unit is
determining a distance between a position in a virtual space where the display of the words is observed and a position of an avatar uttering the words;
The display control unit is
displaying the history of the words or the icons as a dramatic effect not including the meaning of the words and the icons based on the determined distance;
The display control device according to (11).
(13)
The determination unit is
determining a positional relationship between a position in a virtual space where the display of the words is observed and a position of an avatar uttering the words;
The display control unit is
determining an angle and a three-dimensional representation of the representation of the words observed in a virtual space based on the determined positional relationship;
The display control device according to (11) or (12).
(14)
The determination unit is
determining whether a first avatar is included within a range recognized by a second avatar in the virtual space;
The display control unit is
When it is determined that the first avatar is included in a range recognized by the second avatar, a movement path of the first avatar is adjusted so that the first avatar approaches the second avatar;
The display control device according to any one of (1) to (13).
(15)
The determination unit is
determining whether the first avatar has come close to the second avatar in the virtual space;
The display control unit is
when it is determined that the first avatar is in proximity to the second avatar, displaying the first avatar and the second avatar by reflecting a specific action corresponding to the proximity in at least one of the first avatar and the second avatar;
The display control device according to (14).
(16)
The acquisition unit is
receiving a request from the user to change an appearance of the avatar;
The display control unit is
reflecting the change based on the request in the avatar, thereby changing the appearance of the avatar in the virtual space together with a dramatic effect;
The display control device according to any one of (1) to (15).
(17)
The acquisition unit is
receiving from the user a request to exchange any one of a plurality of objects associated with a first avatar with a second avatar;
The display control unit is
When the request is received, a list of a plurality of objects associated with the first avatar is displayed in the vicinity of the first avatar.
The display control device according to any one of (1) to (16).
(18)
The acquisition unit is
obtaining, from the user, an instruction to select one of the plurality of objects displayed in the list to be exchanged for the second avatar;
The display control unit is
When the instruction is acquired, the display device displays any one of the objects in a floating manner from the list display, and moves the floating object in a virtual space so as to overlap the second avatar in accordance with an operation received from the user.
The display control device according to (17).
(19)
The computer
The user's behavior in the real world is acquired as input information,
determining whether the input information includes an input corresponding to a specific action registered in advance;
reflecting the behavior of the user corresponding to the input information in an avatar in a virtual space, and when it is determined that the input information includes an input corresponding to the specific action, further reflecting the specific action in an avatar in the virtual space and displaying the avatar in the virtual space;
A display control method comprising:
(20)
Computer,
an acquisition unit that acquires user behavior in a real space as input information;
a determination unit that determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance;
a display control unit that reflects a behavior of the user corresponding to the input information in an avatar in a virtual space, and, when it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space;
A display control program for causing the display control device to function as a display control device having the display control program.
 10  アバター
 50  分析サーバ
 100 表示制御装置
 110 通信部
 120 記憶部
 130 制御部
 131 取得部
 132 判定部
 133 表示制御部
 140 センサ部
 141 カメラ
 142 マイクロホン
 150 表示部
REFERENCE SIGNS LIST 10 Avatar 50 Analysis server 100 Display control device 110 Communication unit 120 Storage unit 130 Control unit 131 Acquisition unit 132 Determination unit 133 Display control unit 140 Sensor unit 141 Camera 142 Microphone 150 Display unit

Claims (20)

  1.  実空間におけるユーザの挙動を入力情報として取得する取得部と、
     予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する判定部と、
     前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する表示制御部と、
     を備える表示制御装置。
    an acquisition unit that acquires user behavior in a real space as input information;
    a determination unit that determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance;
    a display control unit that reflects a behavior of the user corresponding to the input information in an avatar in a virtual space, and, when it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space;
    A display control device comprising:
  2.  前記取得部は、
     前記ユーザから前記特定の動作を前記アバターに反映するか否かの指定を受け付け、
     前記判定部は、
     前記ユーザの指定がある場合に、前記特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する、
     請求項1に記載の表示制御装置。
    The acquisition unit is
    accepting a designation from the user as to whether or not the specific action is to be reflected in the avatar;
    The determination unit is
    determining whether or not an input corresponding to the specific action is included in the input information when the specific action is designated by the user;
    The display control device according to claim 1 .
  3.  前記取得部は、
     前記ユーザの挙動が撮像される際の撮像装置からの位置を取得し、
     前記判定部は、
     前記撮像装置からの位置が所定距離より離れている場合には、前記特定の動作に対応する入力が前記入力情報に含まれるか否かを判定せず、当該撮像装置からの位置が所定距離以内である場合に、当該特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する、
     請求項1に記載の表示制御装置。
    The acquisition unit is
    Acquire a position from an imaging device when the behavior of the user is imaged;
    The determination unit is
    When the position from the imaging device is farther than a predetermined distance, it is not determined whether or not the input corresponding to the specific action is included in the input information, and when the position from the imaging device is within a predetermined distance, it is determined whether or not the input corresponding to the specific action is included in the input information.
    The display control device according to claim 1 .
  4.  前記取得部は、
     複数のユーザの挙動を入力情報として取得し、
     前記判定部は、
     前記複数のユーザの挙動に基づいて、前記仮想空間における当該複数のユーザ同士の交流の盛り上がり度を判定し、
     前記表示制御部は、
     前記複数のユーザ同士の近傍に、前記交流の盛り上がり度に応じた演出効果を表示する、
     請求項1に記載の表示制御装置。
    The acquisition unit is
    Obtaining the behavior of multiple users as input information,
    The determination unit is
    determining a level of excitement of an interaction between the plurality of users in the virtual space based on the behavior of the plurality of users;
    The display control unit is
    displaying a performance effect in the vicinity of the plurality of users according to the level of excitement of the exchange;
    The display control device according to claim 1 .
  5.  前記判定部は、
     前記交流に参加するユーザの人数、ユーザ同士の会話の頻度、複数のユーザ同士の音声の重なり具合に基づいて、前記交流の盛り上がり度を判定する、
     請求項4に記載の表示制御装置。
    The determination unit is
    determining a level of excitement in the exchange based on the number of users participating in the exchange, the frequency of conversation between the users, and the degree of overlap of voices between the multiple users;
    The display control device according to claim 4.
  6.  前記表示制御部は、
     前記交流の盛り上がり度、前記アバターの身体の向き、前記複数のユーザの会話の内容に基づいて、前記演出効果の表現もしくは発生頻度を決定する、
     請求項5に記載の表示制御装置。
    The display control unit is
    determining an expression or occurrence frequency of the performance effect based on the level of excitement of the exchange, the body orientation of the avatar, and the content of the conversation between the plurality of users;
    The display control device according to claim 5 .
  7.  前記判定部は、
     特定の感情に紐付いた前記特定の動作に対応する入力が、前記入力情報に含まれるか否かを判定し、
     前記表示制御部は、
     前記入力情報に前記特定の感情に紐付いた特定の動作が含まれると判定された場合には、当該特定の感情に紐付いた特定の動作を前記アバターに反映して、前記仮想空間に当該アバターを表示する、
     請求項1に記載の表示制御装置。
    The determination unit is
    determining whether an input corresponding to the specific action associated with a specific emotion is included in the input information;
    The display control unit is
    when it is determined that the input information includes a specific action associated with the specific emotion, the specific action associated with the specific emotion is reflected in the avatar, and the avatar is displayed in the virtual space.
    The display control device according to claim 1 .
  8.  前記判定部は、
     予め前記ユーザから登録された声紋情報であって、前記特定の感情に紐付く声紋情報と、前記入力情報とを比較することで、前記特定の感情に紐付いた特定の動作に対応する入力が当該入力情報に含まれるか否かを判定する、
     請求項7に記載の表示制御装置。
    The determination unit is
    comparing the input information with voiceprint information registered in advance by the user, the voiceprint information being associated with the specific emotion, to determine whether or not the input information includes an input corresponding to a specific action associated with the specific emotion;
    The display control device according to claim 7.
  9.  前記表示制御部は、
     前記特定の感情に紐付いた特定の動作を反映させた前記アバターを表示するとともに、当該特定の感情に紐付いた演出効果を当該アバターの近傍に表示する、
     請求項7に記載の表示制御装置。
    The display control unit is
    displaying the avatar reflecting a specific action associated with the specific emotion, and displaying a performance effect associated with the specific emotion in the vicinity of the avatar;
    The display control device according to claim 7.
  10.  前記表示制御部は、
     前記入力情報に基づいて前記アバターが発する言葉を表示するとともに、当該言葉の履歴が当該アバターを起点として上方向に遷移するよう表示する、
     請求項1に記載の表示制御装置。
    The display control unit is
    displaying words uttered by the avatar based on the input information, and displaying a history of the words in an upward transition manner starting from the avatar;
    The display control device according to claim 1 .
  11.  前記表示制御部は、
     前記入力情報に含まれる意味内容に基づいて、前記アバターが発する言葉を、当該意味内容を絵で表したアイコンで表示する、
     請求項10に記載の表示制御装置。
    The display control unit is
    displaying the words uttered by the avatar as an icon that pictorially represents the meaning based on the meaning included in the input information;
    The display control device according to claim 10.
  12.  前記判定部は、
     前記言葉の表示を観測する仮想空間の位置と、当該言葉を発するアバターの位置との距離を判定し、
     前記表示制御部は、
     前記判定された距離に基づいて、前記言葉の履歴もしくはアイコンを、当該言葉およびアイコンの意味内容を含まない演出効果として表示する、
     請求項11に記載の表示制御装置。
    The determination unit is
    determining a distance between a position in a virtual space where the display of the words is observed and a position of an avatar uttering the words;
    The display control unit is
    displaying the history of the words or the icons as a dramatic effect not including the meaning of the words and the icons based on the determined distance;
    The display control device according to claim 11.
  13.  前記判定部は、
     前記言葉の表示を観測する仮想空間の位置と、当該言葉を発するアバターの位置との位置関係を判定し、
     前記表示制御部は、
     前記判定された位置関係に基づいて、仮想空間で観測される前記言葉の表示の角度および3次元表示を決定する、
     請求項11に記載の表示制御装置。
    The determination unit is
    determining a positional relationship between a position in a virtual space where the display of the words is observed and a position of an avatar uttering the words;
    The display control unit is
    determining an angle and a three-dimensional representation of the representation of the words observed in a virtual space based on the determined positional relationship;
    The display control device according to claim 11.
  14.  前記判定部は、
     前記仮想空間において、第1のアバターが第2のアバターにより認識される範囲内に含まれるかを判定し、
     前記表示制御部は、
     前記第1のアバターが前記第2のアバターにより認識される範囲内に含まれると判定された場合、当該第1のアバターが当該第2のアバターに近接するよう、当該第1のアバターの移動経路を調整する、
     請求項1に記載の表示制御装置。
    The determination unit is
    determining whether a first avatar is within a range recognized by a second avatar in the virtual space;
    The display control unit is
    When it is determined that the first avatar is included in a range recognized by the second avatar, a movement path of the first avatar is adjusted so that the first avatar approaches the second avatar;
    The display control device according to claim 1 .
  15.  前記判定部は、
     前記仮想空間において、前記第1のアバターが前記第2のアバターに近接したか否かを判定し、
     前記表示制御部は、
     前記第1のアバターが前記第2のアバターに近接したと判定された場合に、当該第1のアバターおよび当該第2のアバターの少なくとも1つに前記近接に対応する特定の動作を反映して、当該第1のアバターおよび当該第2のアバターを表示する、
     請求項14に記載の表示制御装置。
    The determination unit is
    determining whether the first avatar has come close to the second avatar in the virtual space;
    The display control unit is
    when it is determined that the first avatar is in proximity to the second avatar, displaying the first avatar and the second avatar by reflecting a specific action corresponding to the proximity in at least one of the first avatar and the second avatar;
    The display control device according to claim 14.
  16.  前記取得部は、
     前記アバターの外観を変更する旨の要求を前記ユーザから取得し、
     前記表示制御部は、
     前記要求に基づく変更を前記アバターに反映して、演出効果とともに前記仮想空間において当該アバターの外観を変更する、
     請求項1に記載の表示制御装置。
    The acquisition unit is
    receiving a request from the user to change an appearance of the avatar;
    The display control unit is
    reflecting the change based on the request in the avatar, thereby changing the appearance of the avatar in the virtual space together with a dramatic effect;
    The display control device according to claim 1 .
  17.  前記取得部は、
     第1のアバターに紐付けられている複数の物体のいずれかを第2のアバターと交換する要求を前記ユーザから取得し、
     前記表示制御部は、
     前記要求が取得されると、前記第1のアバターに紐付けられている複数の物体を当該第1のアバターの近傍に一覧表示する、
     請求項1に記載の表示制御装置。
    The acquisition unit is
    receiving a request from the user to exchange any one of a plurality of objects associated with a first avatar with a second avatar;
    The display control unit is
    When the request is received, a list of a plurality of objects associated with the first avatar is displayed in the vicinity of the first avatar.
    The display control device according to claim 1 .
  18.  前記取得部は、
     前記一覧表示された複数の物体のうち、前記第2のアバターと交換するいずれかの物体を選択する指示を前記ユーザから取得し、
     前記表示制御部は、
     前記指示が取得されると、当該いずれかの物体を一覧表示から浮遊させるよう表示させるとともに、前記ユーザから受け付ける操作に従い、仮想空間において当該浮遊させた物体を前記第2のアバターに重なるよう移動させる、
     請求項17に記載の表示制御装置。
    The acquisition unit is
    obtaining, from the user, an instruction to select one of the plurality of objects displayed in the list to be exchanged for the second avatar;
    The display control unit is
    When the instruction is acquired, the display device displays any one of the objects in a floating manner from the list display, and moves the floating object in a virtual space so as to overlap the second avatar in accordance with an operation received from the user.
    The display control device according to claim 17.
  19.  コンピュータが、
     実空間におけるユーザの挙動を入力情報として取得し、
     予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定し、
     前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する、
     ことを含む表示制御方法。
    The computer
    The user's behavior in the real world is acquired as input information,
    determining whether the input information includes an input corresponding to a specific action registered in advance;
    reflecting the behavior of the user corresponding to the input information in an avatar in a virtual space, and when it is determined that the input information includes an input corresponding to the specific action, further reflecting the specific action in an avatar in the virtual space and displaying the avatar in the virtual space;
    A display control method comprising:
  20.  コンピュータを、
     実空間におけるユーザの挙動を入力情報として取得する取得部と、
     予め登録された特定の動作に対応する入力が前記入力情報に含まれるか否かを判定する判定部と、
     前記入力情報に対応する前記ユーザの挙動を仮想空間のアバターに反映するとともに、当該入力情報に前記特定の動作に対応する入力が含まれると判定された場合には、さらに当該特定の動作を仮想空間のアバターに反映して、当該仮想空間に当該アバターを表示する表示制御部と、
     を備える表示制御装置として機能させるための表示制御プログラム。
    Computer,
    an acquisition unit that acquires user behavior in a real space as input information;
    a determination unit that determines whether or not the input information includes an input corresponding to a specific action that has been registered in advance;
    a display control unit that reflects a behavior of the user corresponding to the input information in an avatar in a virtual space, and, when it is determined that the input information includes an input corresponding to the specific action, further reflects the specific action in an avatar in the virtual space and displays the avatar in the virtual space;
    A display control program for causing the display control device to function as a display control device having the display control program.
PCT/JP2023/034925 2022-10-13 2023-09-26 Display control device, display control method, and display control program WO2024080135A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-165071 2022-10-13
JP2022165071 2022-10-13

Publications (1)

Publication Number Publication Date
WO2024080135A1 true WO2024080135A1 (en) 2024-04-18

Family

ID=90669086

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/034925 WO2024080135A1 (en) 2022-10-13 2023-09-26 Display control device, display control method, and display control program

Country Status (1)

Country Link
WO (1) WO2024080135A1 (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016071558A (en) * 2014-09-29 2016-05-09 シャープ株式会社 Display control device, control method, control program, and recording medium
WO2018020766A1 (en) * 2016-07-28 2018-02-01 ソニー株式会社 Information processing device, information processing method, and program
JP2018092523A (en) * 2016-12-07 2018-06-14 株式会社コロプラ Method for communicating via virtual space, program for causing computer to execute method, and information processing apparatus for executing program
JP2019106220A (en) * 2019-03-27 2019-06-27 株式会社コロプラ Program executed by computer to provide virtual space via head mount device, method, and information processing device
JP2020162882A (en) * 2019-03-29 2020-10-08 株式会社バンダイナムコエンターテインメント Server system and play data community system
JP2021077255A (en) * 2019-11-13 2021-05-20 富士フイルム株式会社 Image processing device, image processing method, and image processing system
WO2022209450A1 (en) * 2021-03-30 2022-10-06 ソニーグループ株式会社 Information processing device, information processing system, and information processing method

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016071558A (en) * 2014-09-29 2016-05-09 シャープ株式会社 Display control device, control method, control program, and recording medium
WO2018020766A1 (en) * 2016-07-28 2018-02-01 ソニー株式会社 Information processing device, information processing method, and program
JP2018092523A (en) * 2016-12-07 2018-06-14 株式会社コロプラ Method for communicating via virtual space, program for causing computer to execute method, and information processing apparatus for executing program
JP2019106220A (en) * 2019-03-27 2019-06-27 株式会社コロプラ Program executed by computer to provide virtual space via head mount device, method, and information processing device
JP2020162882A (en) * 2019-03-29 2020-10-08 株式会社バンダイナムコエンターテインメント Server system and play data community system
JP2021077255A (en) * 2019-11-13 2021-05-20 富士フイルム株式会社 Image processing device, image processing method, and image processing system
WO2022209450A1 (en) * 2021-03-30 2022-10-06 ソニーグループ株式会社 Information processing device, information processing system, and information processing method

Similar Documents

Publication Publication Date Title
US11398067B2 (en) Virtual reality presentation of body postures of avatars
US10489960B2 (en) Virtual reality presentation of eye movement and eye contact
US11532112B2 (en) Emoji recording and sending
JP4395687B2 (en) Information processing device
US20190004639A1 (en) Providing living avatars within virtual meetings
Varona et al. Hands-free vision-based interface for computer accessibility
US20220165013A1 (en) Artificial Reality Communications
KR20200132995A (en) Creative camera
JP7183450B2 (en) Avatar integration with multiple applications
US20140068526A1 (en) Method and apparatus for user interaction
JP2014149856A (en) Enhanced camera-based input
CN109086860B (en) Interaction method and system based on virtual human
WO2022079933A1 (en) Communication supporting program, communication supporting method, communication supporting system, terminal device, and nonverbal expression program
JP7278307B2 (en) Computer program, server device, terminal device and display method
US11824821B2 (en) Modifying an avatar to reflect a user's expression in a messaging platform
KR20180132364A (en) Method and device for videotelephony based on character
KR20220123576A (en) Integrated input/output (i/o) for a three-dimensional (3d) environment
WO2024080135A1 (en) Display control device, display control method, and display control program
WO2023079847A1 (en) Information processing device, information processing method, and storage medium
JP7505666B1 (en) COMMUNICATION SUPPORT PROGRAM, COMMUNICATION SUPPORT METHOD, AND COMMUNICATION SUPPORT SYSTEM
US20240112389A1 (en) Intentional virtual user expressiveness
US20240104870A1 (en) AR Interactions and Experiences
US20230410441A1 (en) Generating user interfaces displaying augmented reality graphics
US20240078732A1 (en) Avatar facial expressions based on semantical context
KR20230082374A (en) An electronic apparatus for adaptive generation of an avatar and a method therefore

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23877141

Country of ref document: EP

Kind code of ref document: A1