WO2018216327A1 - Information processing device, information processing method, and program - Google Patents

Information processing device, information processing method, and program Download PDF

Info

Publication number
WO2018216327A1
WO2018216327A1 PCT/JP2018/010433 JP2018010433W WO2018216327A1 WO 2018216327 A1 WO2018216327 A1 WO 2018216327A1 JP 2018010433 W JP2018010433 W JP 2018010433W WO 2018216327 A1 WO2018216327 A1 WO 2018216327A1
Authority
WO
WIPO (PCT)
Prior art keywords
control unit
information processing
display
remote user
display control
Prior art date
Application number
PCT/JP2018/010433
Other languages
French (fr)
Japanese (ja)
Inventor
高橋 慧
石川 毅
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Publication of WO2018216327A1 publication Critical patent/WO2018216327A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working

Definitions

  • the present disclosure relates to an information processing apparatus, an information processing method, and a program.
  • Communication tools for remote communication are used in various fields such as business, education, and entertainment.
  • a remote user a user existing in a remote place
  • a local user existing in a local environment thereby realizing natural communication closer to reality.
  • Patent Literature 1 describes a technique for presenting an interaction between a virtual object and a real object.
  • the position of the camera that captures the local user is different from the position of the display device that displays the remote user. For this reason, there are cases where the eyes of each other may be felt as if they are not in line with each other, even though they are looking at each other. Even if the AR technology described above is applied to a communication tool and a remote user is presented as a virtual object in real space, depending on the position of the presented remote user, it may be felt that the lines of sight of each other do not match. is there. As a result, smooth communication may be hindered.
  • the present disclosure proposes an information processing apparatus, an information processing method, and a program capable of realizing smoother communication in remote communication.
  • an object indicating a remote user is located at a position in the real space according to a position of the imaging device based on relative position information indicating a relative positional relationship with the imaging device existing in the real space.
  • An information processing apparatus including a display control unit that controls display so as to be visually recognized by a user is provided.
  • An information processing method includes a processor controlling the display so that a local user can view the image.
  • a remote user is placed at a position in the real space according to the position of the imaging device based on relative position information indicating a relative positional relationship with the imaging device existing in the real space.
  • a program is provided for realizing a function of controlling display so that an object indicating is visually recognized by a local user.
  • FIG. 12 is a schematic diagram illustrating a local user's field of view V10 when the display control unit 135 according to Modification 1 displays an avatar in the vicinity of the imaging device 30.
  • FIG. 10 is a block diagram illustrating a configuration example of a communication system 1-2 according to Modification 2. It is explanatory drawing which shows the example of the animation which the display control part 135 which concerns on the modification displays.
  • FIG. 10 is a block diagram illustrating a configuration example of a communication system 1-3 according to Modification 3. It is explanatory drawing which shows the example of the animation which the display control part 135 which concerns on the same modification displays the avatar of a some remote user. It is explanatory drawing which shows the hardware structural example.
  • a plurality of constituent elements having substantially the same functional configuration may be distinguished by adding different alphabets after the same reference numeral.
  • it is not necessary to particularly distinguish each of a plurality of constituent elements having substantially the same functional configuration only the same reference numerals are given.
  • FIGS. 1 and 2 are explanatory diagrams for describing an overview of a communication system according to an embodiment of the present disclosure.
  • the communication system according to the present embodiment is an information processing system that realizes remote communication between a local user LU1 existing in the local environment 1000 shown in FIG. 1 and a remote user RU1 existing in the remote environment 2000.
  • the local environment 1000 and the remote environment 2000 may be any environment in a real space where a real object may exist. Further, the local environment 1000 and the remote environment 2000 may be far away, and for example, it may be a situation where direct face-to-face communication is difficult.
  • the local terminal 10 shown in FIG. 1 is an information processing apparatus that exists in the local environment 1000 and is used by the local user LU1.
  • the local terminal 10 has a transmissive (optical see-through) display unit arranged in front of one or both eyes of the local user LU1, and is mounted on the head of the local user LU1. Glasses-type device.
  • the remote terminal 20 shown in FIG. 1 is an information processing apparatus that exists in the remote environment 2000 and is used by the remote user RU1.
  • the remote terminal 20 has a display unit arranged in front of one or both of the eyes of the remote user RU1, and is an immersive HMD (Head Mounted Display) attached to the head of the remote user RU1. ).
  • HMD Head Mounted Display
  • the imaging device 30 shown in FIG. 1 is a so-called omnidirectional camera that can acquire 360-degree omnidirectional images in all directions in the vertical and horizontal directions by imaging.
  • an image is not limited to a still image, but is used as an expression including a moving image.
  • the local terminal 10, the remote terminal 20, and the imaging device 30 according to the present embodiment are connected to each other via a communication network (not shown).
  • the communication system according to the present embodiment for example, in addition to a message intercommunication function and a message output (display or sound output) function between the local terminal 10 and the remote terminal 20, can be remotely controlled by a display function described below. Realize communication.
  • the apparatus used for remote communication is not limited to the example shown in FIG. 1, and examples of other apparatuses will be described later.
  • the remote terminal 20 displays a display image G ⁇ b> 20 based on the imaging of the imaging device 30 existing in the local environment 1000.
  • the remote terminal 20 may generate a display image G20 by cutting out an area corresponding to the face direction of the remote user RU1 from the omnidirectional image acquired by imaging of the imaging device 30.
  • the remote user RU1 can observe the local environment 1000 in all directions from the viewpoint of the imaging device 30 by changing the face orientation.
  • the remote user RU1 can obtain a sense of facing the local user LU1 by turning his face so that the local user LU1 is included in the display image G20.
  • the local terminal 10 displays a virtual object (hereinafter referred to as an avatar) indicating the remote user RU1.
  • an avatar a virtual object indicating the remote user RU1.
  • the local terminal 10 has a transmissive display unit (not shown)
  • the local user LU1 wearing the local terminal 10 visually recognizes the avatar together with the real space through the display unit of the local terminal 10. It is possible. Therefore, the local terminal 10 can obtain a feeling as if the avatar of the remote user RU1 exists in the real space in front of you.
  • the field of view V10 viewed by the local user LU1 through the display unit of the local terminal 10 includes the avatar A1 of the remote user RU1 together with the real space.
  • the local terminal 10 according to the present embodiment has the avatar A1 at the position in the real space according to the position of the imaging device 30 based on the relative position information indicating the relative positional relationship with the imaging device 30 in the real space. Is displayed on the transmissive display unit. For example, in the field of view V10 shown in FIG. 1, the avatar A1 is superimposed on the imaging device 30 and is visually recognized by the local user LU1.
  • the local user LU1 turns his line of sight toward the imaging device 30 when he / she tries to see the avatar A1.
  • the remote user RU1 sees the local user LU1 in the display image G20 displayed on the remote terminal 20
  • the remote user RU1 feels that the line of sight is aligned with the local user LU1. can get.
  • the local user LU1 feels as if the remote user RU1 exists at the position of the imaging device 30, and the remote user RU1 observes the local environment 1000 from the position of the imaging device 30. As described above, since the remote user RU1 sees an image based on the imaging of the imaging device 30, the smooth communication can be realized by the local user LU1 feeling as described above.
  • the local terminal 10 may change the orientation (face orientation) of the face A11 of the avatar A1 to be displayed according to the face orientation of the remote user RU1.
  • Information regarding the face orientation of the remote user RU1 may be acquired by a sensor included in the remote terminal 20, for example, and provided to the local terminal 10.
  • FIG. 2 shows a display example of the local terminal 10 and the remote terminal 20 when the remote user RU1 faces the right side.
  • the region where the remote terminal 20 cuts out the display image from the omnidirectional image is also changed. Therefore, comparing FIG. 1 with FIG. Yes.
  • the orientation of the face A11 of the avatar A1 displayed by the local terminal 10 also changes according to the face orientation of the remote user RU1.
  • the local user LU1 confirms the angle of view in the imaging device 30 corresponding to the field of view of the remote user RU1, that is, what range the remote user RU1 is viewing in the local environment 1000. Can be grasped. With this configuration, the local user LU1 can perform communication while grasping the scenery that the remote user RU1 is viewing, and can perform communication more smoothly.
  • FIG. 3 is a block diagram illustrating a configuration example of the communication system according to the present embodiment.
  • the communication system 1 includes an information processing system including a local terminal 10, a remote terminal 20, an imaging device 30, a sensor device 40, a distribution server 50, a sensor device 60, and a communication network 70. It is.
  • the communication system 1 includes an information processing system including a local terminal 10, a remote terminal 20, an imaging device 30, a sensor device 40, a distribution server 50, a sensor device 60, and a communication network 70.
  • the local terminal 10 is a glasses-type device that is worn on the head of a local user existing in the local environment 1000 as described with reference to FIG.
  • the local terminal 10 has a transmissive display unit that displays a remote user's avatar at a position corresponding to the position of the imaging device 30 in real space. A more detailed configuration of the local terminal 10 will be described later with reference to FIG.
  • the remote terminal 20 is an immersive HMD attached to the head of a remote user existing in the remote environment 2000 as described with reference to FIG. Further, the remote terminal 20 displays an image based on the imaging of the imaging device 30. A more detailed configuration of the remote terminal 20 will be described later with reference to FIG.
  • the imaging device 30 is an omnidirectional camera that can acquire 360-degree omnidirectional images in all directions in the vertical and horizontal directions by imaging.
  • the imaging device 30 may have a plurality of imaging units, for example, and may acquire an omnidirectional image by performing image processing for combining images obtained by the plurality of imaging units.
  • the imaging device 30 transmits the omnidirectional image to the distribution server 50 via the communication network 70.
  • the imaging apparatus 30 may not have an image processing function. In such a case, for example, another information processing apparatus connected to the imaging apparatus 30 or a distribution server 50 described later is associated with the imaging apparatus 30 instead. It may have an image processing function.
  • the sensor device 40 acquires sensing data related to the local environment 1000 by sensing.
  • the sensor device 40 may include a plurality of sensors, for example, various sensors such as a camera, an infrared camera, a microphone, a depth sensor, an illuminance sensor, and a human sensor.
  • the sensor device 40 provides (transmits) the acquired sensing data to the local terminal 10.
  • the distribution server 50 is an information processing apparatus that distributes (transmits) the omnidirectional image received from the imaging apparatus 30 via the communication network 70 to another apparatus (for example, the remote terminal 20).
  • the distribution server 50 may perform streaming distribution while caching the omnidirectional image received from the imaging device 30.
  • the sensor device 60 acquires sensing data related to the remote environment 2000 by sensing.
  • the sensor device 60 may include a plurality of sensors, for example, various sensors such as a camera, an infrared camera, a microphone, a depth sensor, an illuminance sensor, and a human sensor.
  • the sensor device 60 provides (transmits) the acquired sensing data to the remote terminal 20.
  • the communication network 70 is a wired or wireless transmission path for information transmitted from a device or system connected to the communication network 70.
  • the communication network 70 may include a public line network such as the Internet, a telephone line network, a satellite communication network, various LANs (Local Area Network) including Ethernet (registered trademark), WAN (Wide Area Network), and the like.
  • the communication network 70 may include a dedicated line network such as IP-VPN (Internet Protocol-Virtual Private Network).
  • FIG. 4 is a block diagram illustrating a configuration example of the local terminal 10 according to the present embodiment.
  • the local terminal 10 according to the present embodiment is an information processing apparatus including a sensor unit 11, a communication unit 12, a control unit 13, a display unit 14, a sound output unit 15, and a storage unit 16.
  • the sensor unit 11 acquires sensing data regarding the local user wearing the local terminal 10 and the surrounding environment (local environment 1000) by sensing.
  • the sensor unit 11 may include, for example, an acceleration sensor, a gyro sensor, a camera, a microphone, a geomagnetic sensor, a force sensor, and the like.
  • the sensor data acquired by the sensor unit 11 may include information on the position and orientation of the local terminal 10.
  • the sensor unit 11 provides the acquired sensing data to the control unit 13.
  • the communication unit 12 is a communication interface that mediates communication between the local terminal 10 and other devices.
  • the communication unit 12 supports an arbitrary wireless communication protocol or wired communication protocol, and establishes a communication connection with another device, for example, via the communication network 70 described with reference to FIG. 3 or directly.
  • the control unit 13 controls the operation of each component of the local terminal 10.
  • the control unit 13 also functions as a communication control unit 131, a relative position acquisition unit 133, a display control unit 135, and a sound output control unit 137, as shown in FIG.
  • the communication control unit 131 shown in FIG. 4 controls communication by the communication unit 12, and acquires various types of information from other devices, or transmits them to other devices.
  • the communication control unit 131 causes the remote terminal 20 to transmit voice data acquired by a microphone included in the sensor unit 11 as a message.
  • the communication control unit 131 may cause text data input via an input device (not shown) to be transmitted to the remote terminal 20 as a message.
  • the communication control unit 131 may receive user information regarding the remote user wearing the remote terminal 20 from the remote terminal 20.
  • the user information may include, for example, remote user identification information, remote user status information, remote user attitude information, messages (text data, voice data, etc.) transmitted by the remote user, and the like. Good.
  • the communication control unit 131 may receive sensing data from the sensor device 40 described with reference to FIG.
  • the relative position acquisition unit 133 acquires relative position information indicating a relative positional relationship with the imaging device 30 in real space.
  • the relative position information may be information (coordinates) indicating the position of the imaging device 30 expressed in a coordinate system based on the current local terminal 10, for example.
  • the relative position acquisition unit 133 can acquire the relative position information by various methods. Several examples of the relative position information acquisition method will be described below.
  • the relative position acquisition unit 133 may acquire the relative position information by detecting the imaging device 30 from the image acquired by the camera included in the sensor unit 11.
  • the imaging device 30 may include a detection marker, or the imaging device 30 may include a light emitting unit, and the light emitting unit may emit light with a predetermined light emission pattern. Also good.
  • the relative position acquisition unit 133 may acquire the relative position information by detecting the local terminal 10 by image recognition from the omnidirectional image acquired by the imaging device 30.
  • the local terminal 10 may be provided with a detection marker, or the local terminal 10 may include a light emitting unit, and the light emitting unit may emit light with a predetermined light emission pattern. Also good.
  • the local terminal 10 may acquire an omnidirectional image directly from the imaging device 30 or via the communication network 70 or the distribution server 50.
  • the relative position acquisition unit 133 acquires relative position information based on the coordinates of the imaging device 30 in the absolute coordinate system (hereinafter referred to as absolute coordinates) and the position and orientation information of the local terminal 10. It is also possible to do.
  • the position and orientation information of the local terminal 10 may be included in the sensing data acquired by the sensor unit 11, or based on the sensing data, using self-position estimation technology such as SLAM (Simultaneous Localization and Mapping). It may be specified.
  • the absolute coordinates of the imaging device 30 may be stored, for example, in the storage unit 16 described later, or may be acquired from another device (for example, the imaging device 30) via the communication network 70.
  • the relative position acquisition unit 133 may detect the imaging device 30 from the image acquired by the camera included in the sensor device 40 and specify the absolute coordinates of the imaging device 30.
  • a detection marker may be provided in the imaging device 30, or the imaging device 30 may include a light emitting unit to facilitate detection by a light emission pattern.
  • the relative position acquisition unit 133 may specify the absolute coordinates of the imaging device 30 by matching the omnidirectional image acquired by the imaging device 30 with an image obtained by capturing the local environment 1000 in advance. Good. By dynamically specifying the absolute coordinates of the imaging device 30 in this way, even when the imaging device 30 moves, it is possible to specify the absolute coordinates of the imaging device 30 and acquire the relative position information. It is.
  • the relative position information may be acquired by a method other than the above.
  • the relative position information may be acquired by detecting the imaging device 30 from sensing data other than the image acquired by the sensor unit 11.
  • the relative position information may be acquired by the communication unit 12 receiving the relative position information specified by another device.
  • the display control unit 135 is a transmissive display unit so that the local user can visually recognize the object indicating the remote user at a position in the real space corresponding to the position of the imaging device 30 in the real space. 14 controls display.
  • the object indicating the remote user may be, for example, a virtual avatar, or may be an image of the remote user captured by the camera included in the sensor device 60 described with reference to FIG.
  • the display control unit 135 displays a remote user's avatar as an object indicating the remote user.
  • the display control unit 135 may display the avatar of the remote user so that the imaging device 30 and the avatar in the real space are superimposed and viewed by the local user.
  • the imaging device 30 and the avatar are superimposed and visually recognized means that at least a part of the imaging device 30 overlaps at least a part of the avatar and is visually recognized by the local user.
  • the local user turns his / her line of sight toward the imaging device 30 when he / she tries to see the avatar of the remote user. Then, when the remote user sees the local user in the image displayed on the remote terminal 20, an effect is obtained in which the remote user feels that his / her line of sight matches the local user.
  • the local user feels as if the remote user is present at the position of the imaging device 30 and the remote user observes the local environment 1000 with the position of the imaging device 30 as a viewpoint.
  • the remote user since the remote user sees an image based on the imaging of the imaging device 30 (an image displayed on the remote terminal 20), the local user can feel more as described above. Smooth communication can be realized.
  • the display control unit 135 is visually recognized by superimposing an imaging unit (not shown) of the imaging device 30 and the eyes included in the avatar. As shown, an avatar may be displayed. With such a configuration, the local user can more easily imagine the viewpoint of the remote user, and smoother communication can be realized.
  • the display control unit 135 may control the attitude of the avatar based on the attitude information indicating the attitude of the remote user received from the remote terminal 20 via the communication unit 12. For example, when the posture information includes information related to the face orientation of the remote user and the avatar includes a face (a part that appears to be a face), the display control unit 135 includes the information related to the face orientation. The direction of the face to be displayed may be controlled.
  • the local user confirms the range of the local environment 1000 by the remote user by confirming the face direction of the avatar. It becomes possible. Then, the local user can communicate while grasping the scenery viewed by the remote user, and can communicate more smoothly.
  • posture control by the display control unit 135 is not limited to face orientation control.
  • the information included in the posture information and the posture of the part corresponding to the human body included in the avatar can be controlled.
  • the display control unit 135 may control the avatar's hand, arm, and body posture, respectively. With such a configuration, the local user can feel stronger as if the remote user exists in the local environment 1000.
  • the display control unit 135 may display an avatar corresponding to the remote user.
  • a plurality of avatars are associated with remote user identification information and stored in the storage unit 16, and the display control unit 135 responds to the remote user identification information received from the remote terminal 20 via the communication unit 12. May be selected to display the selected avatar.
  • the local user can identify the remote user via the avatar.
  • the display control unit 135 may display a message transmitted by the remote user.
  • the message transmitted by the remote user is received from the remote terminal 20 via the communication unit 12, for example.
  • the message displayed by the display control unit 135 is not limited to text, and may include electronic data such as image data and document data.
  • the display control unit 135 may display an icon indicating electronic data.
  • the icon displayed by the display control unit 135 may be an image.
  • FIG. 5 is a schematic diagram showing the field of view V10 of the local user when the display control unit 135 displays a message.
  • the display control unit 135 displays the message M10 in the vicinity of the avatar A1.
  • the display control unit 135 displays the message M10 as a balloon from the position of the avatar A1.
  • the display control unit 135 may display an icon M11 indicating electronic data as shown in FIG.
  • the display control unit 135 may control the display according to the state of the remote user.
  • the display control unit 135 may display an indicator (for example, an icon or text) indicating the state of the remote user in the vicinity of the remote user's avatar.
  • the display control unit 135 may control parameters related to the display of the avatar according to the state of the remote user. Parameters relating to avatar display may include, for example, luminance, color, saturation, transparency, posture, and the like.
  • information indicating the status of the remote user may be received from the remote terminal 20 via the communication unit 12.
  • the information indicating the status of the remote user includes whether the remote user is online, whether the remote user is wearing the remote terminal 20, whether the remote user has permitted display of the avatar, and the like. Information may be included.
  • the local user can grasp the status of the remote user.
  • the display control unit 135 receives information on whether or not the image of the omnidirectional camera is normally transmitted to the remote terminal 20 from the distribution server 50 or the imaging device 30, and controls display based on the information. May be. In such a case, the display control unit 135 may display an icon or text indicating whether or not the image of the omnidirectional camera is normally transmitted to the remote terminal 20, or may control a parameter related to the display of the avatar. Good.
  • the display control unit 135 may control the display according to whether or not the voice communication between the local terminal 10 and the remote terminal 20 is normally performed. In such a case, the display control unit 135 may display an icon or text indicating whether or not the voice communication between the local terminal 10 and the remote terminal 20 is normally performed, and controls parameters related to the display of the avatar. May be.
  • the sound output control unit 137 may control the sound output unit 15 to sound output a message transmitted by the remote user.
  • the sound output control unit 137 may control the sound output so that the message can be heard from the position in the real space where the avatar is visually recognized by the local user. With this configuration, the local user can obtain a stronger feeling as if the remote user exists in the local environment 1000.
  • the display unit 14 is a display that is controlled by the display control unit 135 to display various information including objects such as avatars.
  • the display unit 14 may be a transmissive (optical see-through) display. With such a configuration, the local user wearing the local terminal 10 can view the real space and the information displayed on the display unit 14 at the same time.
  • the display unit 14 may be a non-transparent display unit, and such a case will be described later as a modified example.
  • the display unit 14 may be able to present different images to both eyes of the local user wearing the local terminal 10, for example, and the local user can display the shape of the displayed object, It is possible to recognize the position three-dimensionally.
  • the sound output unit 15 performs sound output under the control of the sound output control unit 137.
  • the sound output unit 15 may include, for example, a plurality of speakers, and the sound output unit 15 may be capable of outputting sound in a three-dimensional manner (stereo sound).
  • the storage unit 16 stores a program for the control unit 13 to execute the above processes and various data.
  • the storage unit 16 may store the avatar information described above and a history of information received from the remote terminal 20 via the communication unit 12.
  • the function of the storage unit 16 may exist in an external device, and the local terminal 10 may receive information stored in a storage unit included in the external device via the communication unit 12, for example.
  • the configuration example of the local terminal 10 has been described above with reference to FIG. 4, but the example illustrated in FIG. 4 is an example, and the present technology is not limited to the example.
  • some functions shown in FIG. 4 may be provided in another information processing apparatus connected to the local terminal 10.
  • some or all of the functions of the control unit 13 are provided in another information processing apparatus connected to the local terminal 10, and the local terminal 10 is controlled by the other information processing apparatus to display or output sound. You may go.
  • the relative position information may be information indicating a relative position between the local terminal 10 that performs display and the imaging device 30.
  • FIG. 6 is a block diagram illustrating a configuration example of the remote terminal 20 according to the present embodiment.
  • the remote terminal 20 according to the present embodiment is an information processing apparatus including a sensor unit 21, a communication unit 22, a control unit 23, a display unit 24, a sound output unit 25, and a storage unit 26.
  • the sensor unit 21 acquires sensing data related to the remote user wearing the remote terminal 20 and the surrounding environment (remote environment 2000) by sensing.
  • the sensor unit 21 may include, for example, an acceleration sensor, a gyro sensor, a camera, a microphone, a geomagnetic sensor, a force sensor, and the like.
  • the sensor data acquired by the sensor unit 21 may include posture information related to the posture (for example, the face orientation) of the remote user.
  • the sensor unit 21 provides the acquired sensing data to the control unit 23.
  • the communication unit 22 is a communication interface that mediates communication between the remote terminal 20 and other devices.
  • the communication unit 22 supports an arbitrary wireless communication protocol or wired communication protocol, and establishes a communication connection with another device, for example, via the communication network 70 described with reference to FIG.
  • the control unit 23 controls the operation of each component of the remote terminal 20. Further, as shown in FIG. 6, the control unit 23 also functions as a communication control unit 231, a display control unit 235, and a sound output control unit 237.
  • the communication control unit 231 illustrated in FIG. 6 controls communication by the communication unit 22 and acquires various types of information from other devices, or transmits them to other devices.
  • the communication control unit 231 transmits user information including information indicating the status of the remote user, attitude information of the remote user, messages (text data, voice data, and the like) transmitted by the remote user to the local terminal 10.
  • the communication control unit 231 may cause the voice data acquired by the microphone included in the sensor unit 21 to be transmitted as a message to the local terminal 10, or text data input via an input device (not shown) as a message. May be transmitted to the local terminal 10.
  • the communication control unit 231 may receive a message (text data, voice data, etc.) transmitted from the local terminal 10 by the local user.
  • the communication control unit 231 may receive the omnidirectional image from the distribution server 50 described with reference to FIG.
  • the communication control unit 231 may receive sensing data from the sensor device 60 described with reference to FIG.
  • the display control unit 235 controls display by the display unit 24.
  • the display control unit 235 may generate a display image by cutting out a region corresponding to the face orientation of the remote user acquired by the sensor unit 21 from the omnidirectional image, and display the display image on the display unit 24. .
  • a remote user wearing the remote terminal 20 can observe the local environment 1000 in all directions from the viewpoint of the imaging device 30 by changing the face orientation. Further, as shown in FIG. 1, the remote user can obtain a sense of facing the local user by facing the face in a direction that is included in the display image.
  • the display control unit 235 may further perform image processing on the area cut out from the omnidirectional image to generate a display image.
  • the image processing performed by the display control unit 235 includes a processing process that detects a local user from an area cut out from the omnidirectional image and processes the local user area, or converts the local user area into another object (for example, A replacement process may be included.
  • the processing performed by the display control unit 235 may include, for example, processing for removing the local terminal 10 worn by the local user and restoring the facial expression of the local user.
  • the image processing performed by the display control unit 235 may include processing for synthesizing a message transmitted by the local user.
  • the sound output control unit 237 illustrated in FIG. the sound output control unit 237 may control the sound output unit 25 to sound output a message transmitted by the local user.
  • the sound output control unit 237 may control the sound output so that a message can be heard from the position of the local user displayed on the display unit 24. With this configuration, the remote user can obtain a more immersive feeling.
  • the display unit 24 is a display that displays a display image under the control of the display control unit 235.
  • the display unit 24 may be able to present different images to both eyes of a remote user wearing the remote terminal 20, for example. With such a configuration, the remote user can observe the local environment 1000 in a three-dimensional manner, and can obtain a more immersive feeling.
  • the sound output unit 25 performs sound output under the control of the sound output control unit 237.
  • the sound output unit 25 may include, for example, a plurality of speakers, and the sound output unit 25 may be capable of outputting sound in a three-dimensional manner (stereo sound).
  • the storage unit 26 stores a program for the control unit 23 to execute the above processes and various data.
  • the configuration example of the remote terminal 20 has been described above with reference to FIG. 6, but the example illustrated in FIG. 6 is an example, and the present technology is not limited to the example.
  • some functions shown in FIG. 6 may be provided in another information processing apparatus connected to the remote terminal 20.
  • a part or all of the functions of the control unit 23 shown in FIG. 6 are provided in another information processing apparatus connected to the remote terminal 20, and the remote terminal 20 displays and is controlled by the other information processing apparatus. Or sound output may be performed.
  • FIG. 7 is a flowchart showing an operation example of the communication system 1 according to the present embodiment.
  • imaging is performed by the imaging device 30 (S102), and then image processing such as synthesis processing based on the captured image is performed, and an omnidirectional image is acquired (S104). Subsequently, the omnidirectional image is transmitted from the imaging device 30 to the distribution server 50 (S108). Further, the distribution server 50 distributes (transmits) the omnidirectional image received from the imaging device 30 to the remote terminal 20 (S108).
  • the sensor unit 21 of the remote terminal 20 acquires sensing data including, for example, information on the face direction of the remote user by sensing (S110). Further, the display control unit of the remote terminal 20 cuts out an area corresponding to the remote user's face direction from the omnidirectional image, generates a display image, and causes the display unit 24 to display the display image (S112).
  • the remote terminal 20 locally stores user information including remote user identification information, information indicating the status of the remote user, attitude information of the remote user, messages (text data, voice data, etc.) transmitted by the remote user, and the like. It transmits to the terminal 10 (S114).
  • the relative position acquisition unit 133 of the local terminal 10 acquires relative position information indicating the relative positional relationship between the local terminal 10 and the imaging device existing in the real space (S116). Further, the display control unit 135 of the local terminal 10 recognizes the remote user's avatar by the local user at a position in the real space according to the position of the imaging device 30 in the real space based on the user information and the relative position information. As shown, the avatar is displayed (S118).
  • FIG. 7 shows an example, and the present technology is not limited to the example.
  • the processes in steps S102 to S118 shown in FIG. 7 may be repeated.
  • the processing order of step S108 and step S110 may be reversed, and the processing order of step S114 and step S116 may be reversed.
  • FIG. 8 is a schematic diagram showing the field of view V10 of the local user when the display control unit 135 displays an avatar in the vicinity of the imaging device 30.
  • the remote user's avatar A ⁇ b> 1 is displayed beside the imaging device 30.
  • the position where the avatar is visually recognized is not limited to the side of the imaging device 30 and may be above, below, inside, or in front of the imaging device 30.
  • FIG. 9 is a block diagram showing a configuration example of the communication system 1-2 according to this modification.
  • the configuration of the communication system 1-2 illustrated in FIG. 9 is the same as the configuration of the communication system 1 described with reference to FIG. 3 except that the two imaging devices 30A and 30B exist in the local environment 1000. Therefore, it will be described while omitting as appropriate.
  • the imaging device 30A and the imaging device 30B are omnidirectional cameras that can acquire 360-degree omnidirectional images in all directions in the vertical and horizontal directions by imaging, as with the imaging device 30 described above.
  • the imaging device 30 ⁇ / b> A and the imaging device 30 ⁇ / b> B transmit an omnidirectional image to the distribution server 50 via the communication network 70.
  • the remote terminal 20 may display an image based on the imaging of one imaging device selected by the remote user from the imaging device 30A or the imaging device 30B.
  • the remote user may perform an input operation related to the selection via, for example, the sensor device 60, the sensor unit 21 of the remote terminal 20, or an input device (not shown).
  • the distribution server 50 may transmit both the omnidirectional image captured by the imaging device 30A and the omnidirectional image captured by the imaging device 30B to the remote terminal 20, or the imaging selected by the remote user. Only the omnidirectional image captured by the apparatus may be transmitted to the remote terminal 20.
  • the local terminal 10 may have the following functions in addition to the functions described with reference to FIG.
  • the display control unit 135 of the local terminal 10 has the remote user's avatar at a position in the real space corresponding to the position of one imaging device selected by the remote user among the plurality of imaging devices.
  • the display is controlled so as to be visually recognized by the local user.
  • the display control unit 135 when the position where the avatar is visually recognized is switched according to the selection of the remote user, the display control unit 135 according to this modification displays the avatar so that the avatar fades out and then fades in to a new position. You may let them.
  • the display control unit 135 according to the present modification is currently selected from the position corresponding to the imaging device selected immediately before when the position where the avatar is visually recognized is switched according to the selection of the remote user. You may display the animation which an avatar moves to the position according to an imaging device. In this animation, a video effect in which the avatar moves three-dimensionally may be added.
  • FIG. 10 is an explanatory diagram showing an example of an animation displayed by the display control unit 135 according to this modification.
  • FIG. 10 shows a state in which the imaging device 30B is selected by the remote user RU1 after the imaging device 30A has been selected.
  • the remote terminal 20 displays a display image G20 generated by cutting out an area corresponding to the face direction of the remote user RU1 from the omnidirectional image acquired by imaging of the imaging device 30B. Further, in the field of view V10 of the local user LU1, the avatar A1 that has been visually recognized while being superimposed with the imaging device 30A has moved to a position that is visually recognized with being superimposed with the imaging device 30B.
  • the remote user can observe the local environment 1000 from various viewpoints, and the local user can easily grasp the imaging device corresponding to the current viewpoint of the remote user. .
  • FIG. 11 is a block diagram showing a configuration example of the communication system 1-3 according to this modification.
  • the configuration of the communication system 1-3 illustrated in FIG. 11 is partially the same as the configuration of the communication system 1 described with reference to FIG.
  • the communication system 1-3 includes two remote terminals 20A and 20B, and is shown in FIG. 3 in that it includes two sensor devices 60A and 60B. Different from communication system 1. Further, the remote terminal 20A and the sensor device 60A exist in the remote environment 2000A, and the remote terminal 20B and the sensor device 60B exist in the remote environment 2000B. The remote environment 2000A and the remote environment 2000B may be the same (single) environment. In such a case, the sensor device 60A and the sensor device 60B may be the same (single) device.
  • the remote terminal 20A may be an information processing device used by a first remote user
  • the remote terminal 20B may be an information processing device used by a second remote user different from the first remote user. Note that the configurations of the remote terminal 20A and the remote terminal 20B according to this modification are substantially the same as the configuration of the remote terminal 20 described with reference to FIG.
  • the distribution server 50 distributes (transmits) the omnidirectional image obtained by the imaging of the imaging device 30 to both the remote terminal 20A and the remote terminal 20B.
  • the local terminal 10 according to this modification may have the following functions in addition to the functions described with reference to FIG.
  • the display control unit 135 of the local terminal 10 according to this modification may display an object (for example, an avatar) indicating a plurality of remote users.
  • FIG. 12 is an explanatory diagram showing an example of an animation in which the display control unit 135 according to this modification displays a plurality of remote user avatars.
  • the display control unit 135 displays both the avatar A1 of the first remote user RU1 wearing the remote terminal 20A and the avatar A2 of the second remote user RU2 wearing the remote terminal 20B. ing.
  • the display control unit 135 overlaps with other avatars.
  • An avatar may be displayed so as to be visually recognized in the vicinity of the imaging device 30 that should not be.
  • the display control unit 135 may control the display so that a plurality of avatars do not overlap with each other and can be visually recognized while being superimposed on the imaging device 30 by displaying the avatar in a reduced size.
  • the display control unit 135 of the local terminal 10 has a function of displaying an avatar corresponding to the remote user as described above, and in the example illustrated in FIG. 12, the avatar A1 and the avatar A2 are different avatars.
  • the display control unit 135 may control the avatar posture of each remote user in accordance with the posture information of each remote user.
  • the face of the avatar A1's face A11 is displayed in a direction corresponding to the face direction of the remote user RU1
  • the face of the avatar A2's face A21 is displayed in a direction corresponding to the face direction of the remote user RU2. Yes.
  • the remote user RU1 existing in the remote environment 2000A is wearing the remote terminal 20A. Further, on the remote terminal 20A, a display image G21 generated by cutting out an area corresponding to the face orientation of the remote user RU1 from the omnidirectional image obtained by imaging by the imaging device 30 is displayed.
  • the remote user RU2 existing in the remote environment 2000B is wearing the remote terminal 20B.
  • the remote terminal 20B displays a display image G22 generated by cutting out an area corresponding to the face direction of the remote user RU2 from the omnidirectional image obtained by imaging by the imaging device 30.
  • each remote terminal 20 may display the avatar of another remote user. For example, when the remote user RU1 turns to the left, the remote terminal 20A displays, as a display image, an image obtained by combining the avatar of the remote user RU2 with an image obtained by cutting out an area corresponding to the face direction from the omnidirectional image. May be.
  • the remote terminal 20A displays a message in the vicinity of the avatar of the remote user RU2, or outputs a sound so that a message transmitted by the remote user RU2 can be heard from the position of the avatar of the remote user RU2. Also good.
  • Modification 2 may be combined with Modification 2 described above.
  • a plurality of remote users select one imaging device from a plurality of imaging devices, and the display control unit 135 of the local terminal 10 displays the remote user's avatar at a position corresponding to the position of the imaging device selected by each remote user.
  • the display may be controlled so as to be visually recognized.
  • each remote terminal 20 may display the avatar of the other remote user at a position corresponding to the position of the imaging device selected by the remote user other than the remote user wearing the remote terminal 20.
  • the remote terminal 20 displays a message in the vicinity of the avatar of the other remote user, or outputs a sound so that a message transmitted by the remote user can be heard from the position of the avatar of the remote user. Also good.
  • the local terminal 10 is a glasses-type device having a transmissive display unit
  • the remote terminal 20 is an immersive HMD
  • the imaging device 30 is an omnidirectional camera.
  • the embodiment according to the invention is not limited to the example, and can be realized by various apparatus configurations.
  • the display unit 14 included in the local terminal 10 may not be a transmissive type.
  • an avatar an example of an object indicating a remote user
  • a real space image that is included in the sensor unit 11 included in the local terminal 10 and acquired by a camera that captures the field of view (real space) of the local user. May be displayed.
  • the display control unit 135 of the local terminal 10 allows the local user to visually recognize the avatar at a position in the real space according to the position of the imaging device 30. It is possible to control the display.
  • the local terminal 10 may be a smartphone or a tablet terminal. Even in such a case, a real space image and an avatar acquired by a camera included in the sensor unit 11 may be combined and displayed.
  • the display unit 14 of the local terminal 10 may be a projector.
  • the local terminal 10 may project an avatar image at a position in the real space according to the position of the imaging device 30.
  • the display control unit 135 of the local terminal 10 controls the projector so that the avatar is positioned at a position in the real space according to the position of the imaging device 30.
  • the display can be controlled to be viewed by a local user.
  • the display unit 24 of the remote terminal 20 may be a flat display.
  • the area corresponding to the field of view of the remote user may be specified according to an operation via an input device such as a remote controller or a gesture operation acquired by the sensor unit 21.
  • the remote terminal 20 may be a smartphone or a tablet terminal. Even in such a case, the region corresponding to the visual field of the remote user may be specified according to the touch operation or the posture of the remote terminal 20 acquired by the sensor unit 21.
  • the imaging device 30 may not be a spherical camera, and may be a camera having an imaging angle of view of 180 degrees, for example. Then, the display control unit 135 of the local terminal 10 may limit the face direction of the avatar to be displayed according to the imaging angle of view of the imaging device 30. With such a configuration, it is possible to suppress inconsistency between the avatar's face direction and the field of view actually seen by the remote user.
  • FIG. 13 is a block diagram illustrating an example of a hardware configuration of the information processing apparatus according to the present embodiment.
  • the information processing apparatus 900 illustrated in FIG. 13 can implement the local terminal 10, the remote terminal 20, and the distribution server 50, for example.
  • Information processing by the local terminal 10, the remote terminal 20, and the distribution server 50 according to the present embodiment is realized by cooperation of software and hardware described below.
  • the information processing apparatus 900 includes a CPU (Central Processing Unit) 901, a ROM (Read Only Memory) 902, a RAM (Random Access Memory) 903, and a host bus 904a.
  • the information processing apparatus 900 includes a bridge 904, an external bus 904b, an interface 905, an input device 906, an output device 907, a storage device 908, a drive 909, a connection port 911, a communication device 913, and a sensor 915.
  • the information processing apparatus 900 may include a processing circuit such as a DSP or an ASIC in place of or in addition to the CPU 901.
  • the CPU 901 functions as an arithmetic processing unit and a control unit, and controls the overall operation in the information processing apparatus 900 according to various programs. Further, the CPU 901 may be a microprocessor.
  • the ROM 902 stores programs used by the CPU 901, calculation parameters, and the like.
  • the RAM 903 temporarily stores programs used in the execution of the CPU 901, parameters that change as appropriate during the execution, and the like. For example, the CPU 901 can form the control unit 13 and the control unit 23.
  • the CPU 901, ROM 902, and RAM 903 are connected to each other by a host bus 904a including a CPU bus.
  • the host bus 904 a is connected to an external bus 904 b such as a PCI (Peripheral Component Interconnect / Interface) bus via a bridge 904.
  • an external bus 904 b such as a PCI (Peripheral Component Interconnect / Interface) bus
  • PCI Peripheral Component Interconnect / Interface
  • the host bus 904a, the bridge 904, and the external bus 904b do not necessarily have to be configured separately, and these functions may be mounted on one bus.
  • the input device 906 is realized by a device in which information is input by the user, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, and a lever.
  • the input device 906 may be, for example, a remote control device using infrared rays or other radio waves, or may be an external connection device such as a mobile phone or a PDA that supports the operation of the information processing device 900.
  • the input device 906 may include, for example, an input control circuit that generates an input signal based on information input by the user using the above-described input means and outputs the input signal to the CPU 901.
  • a user of the information processing apparatus 900 can input various data and instruct a processing operation to the information processing apparatus 900 by operating the input device 906.
  • the output device 907 is formed of a device that can notify the user of the acquired information visually or audibly. Examples of such devices include CRT display devices, liquid crystal display devices, plasma display devices, EL display devices, display devices such as lamps, audio output devices such as speakers and headphones, printer devices, and the like.
  • the output device 907 outputs results obtained by various processes performed by the information processing device 900. Specifically, the display device visually displays results obtained by various processes performed by the information processing device 900 in various formats such as text, images, tables, and graphs.
  • the audio output device converts an audio signal composed of reproduced audio data, acoustic data, and the like into an analog signal and outputs it aurally.
  • the output device 907 can form, for example, the display unit 14, the sound output unit 15, the display unit 24, and the sound output unit 25.
  • the storage device 908 is a data storage device formed as an example of a storage unit of the information processing device 900.
  • the storage apparatus 908 is realized by, for example, a magnetic storage device such as an HDD, a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like.
  • the storage device 908 may include a storage medium, a recording device that records data on the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded on the storage medium, and the like.
  • the storage device 908 stores programs executed by the CPU 901, various data, various data acquired from the outside, and the like.
  • the storage device 908 can form the storage unit 16 and the storage unit 26, for example.
  • the drive 909 is a storage medium reader / writer, and is built in or externally attached to the information processing apparatus 900.
  • the drive 909 reads information recorded on a removable storage medium such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and outputs the information to the RAM 903.
  • the drive 909 can also write information to a removable storage medium.
  • connection port 911 is an interface connected to an external device, and is a connection port with an external device capable of transmitting data by USB (Universal Serial Bus), for example.
  • USB Universal Serial Bus
  • the communication device 913 is a communication interface formed by a communication device or the like for connecting to the network 920, for example.
  • the communication device 913 is, for example, a communication card for wired or wireless LAN (Local Area Network), LTE (Long Term Evolution), Bluetooth (registered trademark), or WUSB (Wireless USB).
  • the communication device 913 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), a modem for various communication, or the like.
  • the communication device 913 can transmit and receive signals and the like according to a predetermined protocol such as TCP / IP, for example, with the Internet and other communication devices.
  • the communication device 913 can form the communication unit 12 and the communication unit 22, for example.
  • the sensor 915 is various sensors such as an acceleration sensor, a gyro sensor, a geomagnetic sensor, an optical sensor, a sound sensor, a distance measuring sensor, and a force sensor.
  • the sensor 915 acquires information on the state of the information processing apparatus 900 itself, such as the posture and movement speed of the information processing apparatus 900, and information on the surrounding environment of the information processing apparatus 900, such as brightness and noise around the information processing apparatus 900.
  • Sensor 915 may also include a GPS sensor that receives GPS signals and measures the latitude, longitude, and altitude of the device.
  • the sensor 915 can form the sensor part 11 and the sensor part 21, for example.
  • the network 920 is a wired or wireless transmission path for information transmitted from a device connected to the network 920.
  • the network 920 may include a public line network such as the Internet, a telephone line network, and a satellite communication network, various LANs including the Ethernet (registered trademark), a wide area network (WAN), and the like.
  • the network 920 may include a dedicated line network such as an IP-VPN (Internet Protocol-Virtual Private Network).
  • IP-VPN Internet Protocol-Virtual Private Network
  • each of the above components may be realized using a general-purpose member, or may be realized by hardware specialized for the function of each component. Therefore, it is possible to change the hardware configuration to be used as appropriate according to the technical level at the time of carrying out this embodiment.
  • a computer program for realizing each function of the information processing apparatus 900 according to the present embodiment as described above can be produced and mounted on a PC or the like.
  • a computer-readable recording medium storing such a computer program can be provided.
  • the recording medium is, for example, a magnetic disk, an optical disk, a magneto-optical disk, a flash memory, or the like.
  • the above computer program may be distributed via a network, for example, without using a recording medium.
  • the local terminal 10 displays the avatar that is a virtual object as the object indicating the remote user
  • the present technology is not limited to the example.
  • the local terminal 10 may display an image obtained by photographing the remote user with a camera included in the sensor device 60 as an object indicating the remote user.
  • an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device.
  • An information processing apparatus comprising a display control unit that controls display.
  • the display control unit displays the object so that the imaging apparatus and the object are superimposed and viewed.
  • the display control unit displays the object so that an imaging unit of the imaging apparatus and eyes included in the object are superimposed and viewed.
  • the information processing apparatus controls display according to a state of the remote user.
  • the display control unit displays an indicator indicating the state of the remote user in the vicinity of the object.
  • the display control unit controls a parameter related to display of the object according to a state of the remote user.
  • the display control unit controls display so that the object is visually recognized at a position in the real space according to the position of one imaging device selected from a plurality of imaging devices.
  • the display control unit changes from the position corresponding to the imaging device selected immediately before to the position corresponding to the currently selected imaging device.
  • the display control unit displays an object indicating a plurality of remote users.
  • An acoustic output control unit that acoustically outputs a message transmitted by the remote user, and the acoustic output control unit controls the acoustic output so that the message can be heard from a position in the real space where the object is visually recognized.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

[Problem] To provide an information processing device, an information processing method, and a program. [Solution] This information processing device is provided with a display control unit which, on the basis of relative positional information indicating a relative positional relationship with an imaging device present in real space, controls display such that an object indicating a remote user is visually recognized by a local user at a position in the real space that corresponds to the position of the imaging device.

Description

情報処理装置、情報処理方法、及びプログラムInformation processing apparatus, information processing method, and program
 本開示は、情報処理装置、情報処理方法、及びプログラムに関する。 The present disclosure relates to an information processing apparatus, an information processing method, and a program.
 ビジネス、教育、及び娯楽等の様々な分野において、リモートコミュニケーション用のコミュニケーションツールが活用されている。そうしたコミュニケーションツールでは、例えばリモートユーザ(遠隔地に存在するユーザ)を撮影して、ローカル環境に存在するローカルユーザに提示することで、より現実に近い自然なコミュニケーションを実現している。 ・ Communication tools for remote communication are used in various fields such as business, education, and entertainment. In such a communication tool, for example, a remote user (a user existing in a remote place) is photographed and presented to a local user existing in a local environment, thereby realizing natural communication closer to reality.
 一方、最近では拡張現実感(Augmented Reality:AR)技術のように、仮想物体を実空間に存在する実物体と同時にユーザに提示する技術が浸透しつつある。例えば、下記特許文献1には、仮想物体と実物体とのインタラクションを提示する技術が記載されている。 On the other hand, recently, a technique of presenting a virtual object to a user simultaneously with a real object existing in a real space, such as augmented reality (AR) technology, is spreading. For example, Patent Literature 1 below describes a technique for presenting an interaction between a virtual object and a real object.
国際公開第2014/171200号International Publication No. 2014/171200
 しかし、上記のようなコミュニケーションツールでは、多くの場合、ローカルユーザを撮影するカメラの位置と、リモートユーザを表示する表示装置の位置が異なる。そのため、お互いに相手が撮影された画像を見ているにも関わらず、お互いの視線が合っていないように感じてしまう場合がある。仮に上述したAR技術をコミュニケーションツールに適用し、リモートユーザを仮想物体として実空間に提示したとしても、提示されたリモートユーザの位置によっては、お互いの視線が合っていないように感じてしまう場合がある。その結果、円滑なコミュニケーションが阻害される恐れがあった。 However, in the communication tools as described above, in many cases, the position of the camera that captures the local user is different from the position of the display device that displays the remote user. For this reason, there are cases where the eyes of each other may be felt as if they are not in line with each other, even though they are looking at each other. Even if the AR technology described above is applied to a communication tool and a remote user is presented as a virtual object in real space, depending on the position of the presented remote user, it may be felt that the lines of sight of each other do not match. is there. As a result, smooth communication may be hindered.
 そこで、本開示では、リモートコミュニケーションにおいて、より円滑なコミュニケーションを実現することが可能な、情報処理装置、情報処理方法、及びプログラムを提案する。 Therefore, the present disclosure proposes an information processing apparatus, an information processing method, and a program capable of realizing smoother communication in remote communication.
 本開示によれば、実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、表示を制御する表示制御部を備える、情報処理装置が提供される。 According to the present disclosure, an object indicating a remote user is located at a position in the real space according to a position of the imaging device based on relative position information indicating a relative positional relationship with the imaging device existing in the real space. An information processing apparatus including a display control unit that controls display so as to be visually recognized by a user is provided.
 また、本開示によれば、実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、プロセッサが表示を制御すること、を含む情報処理方法が提供される。 According to the present disclosure, an object indicating a remote user at a position in the real space according to the position of the imaging device based on relative position information indicating a relative positional relationship with the imaging device existing in the real space. An information processing method is provided that includes a processor controlling the display so that a local user can view the image.
 また、本開示によれば、コンピュータに、実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、表示を制御する機能を実現させるための、プログラムが提供される。 In addition, according to the present disclosure, a remote user is placed at a position in the real space according to the position of the imaging device based on relative position information indicating a relative positional relationship with the imaging device existing in the real space. A program is provided for realizing a function of controlling display so that an object indicating is visually recognized by a local user.
 以上説明したように本開示によれば、リモートコミュニケーションにおいて、より円滑なコミュニケーションを実現することが可能である。 As described above, according to the present disclosure, it is possible to realize smoother communication in remote communication.
 なお、上記の効果は必ずしも限定的なものではなく、上記の効果とともに、または上記の効果に代えて、本明細書に示されたいずれかの効果、または本明細書から把握され得る他の効果が奏されてもよい。 Note that the above effects are not necessarily limited, and any of the effects shown in the present specification, or other effects that can be grasped from the present specification, together with or in place of the above effects. May be played.
本開示の一実施形態に係るコミュニケーションシステムの概要を説明するための説明図である。It is explanatory drawing for demonstrating the outline | summary of the communication system which concerns on one Embodiment of this indication. 同実施形態に係るコミュニケーションシステムの概要を説明するための説明図である。It is explanatory drawing for demonstrating the outline | summary of the communication system which concerns on the embodiment. 同実施形態に係るコミュニケーションシステムの構成例を示すブロック図である。It is a block diagram which shows the structural example of the communication system which concerns on the same embodiment. 同実施形態に係るローカル端末10の構成例を示すブロック図である。It is a block diagram showing an example of composition of local terminal 10 concerning the embodiment. 同実施形態に係る表示制御部135がメッセージを表示させた場合のローカルユーザの視界V10を示す模式図である。It is a mimetic diagram showing field of view V10 of a local user when display control part 135 concerning the embodiment displays a message. 同実施形態に係るリモート端末20の構成例を示すブロック図である。It is a block diagram showing an example of composition of remote terminal 20 concerning the embodiment. 同実施形態に係るコミュニケーションシステム1の動作例を示すフローチャート図である。It is a flowchart figure which shows the operation example of the communication system 1 which concerns on the same embodiment. 変形例1に係る表示制御部135が撮像装置30の近傍にアバタを表示させた場合のローカルユーザの視界V10を示す模式図である。FIG. 12 is a schematic diagram illustrating a local user's field of view V10 when the display control unit 135 according to Modification 1 displays an avatar in the vicinity of the imaging device 30. 変形例2に係るコミュニケーションシステム1-2の構成例を示すブロック図である。FIG. 10 is a block diagram illustrating a configuration example of a communication system 1-2 according to Modification 2. 同変形例に係る表示制御部135が表示させるアニメーションの例を示す説明図である。It is explanatory drawing which shows the example of the animation which the display control part 135 which concerns on the modification displays. 変形例3に係るコミュニケーションシステム1-3の構成例を示すブロック図である。FIG. 10 is a block diagram illustrating a configuration example of a communication system 1-3 according to Modification 3. 同変形例に係る表示制御部135が複数のリモートユーザのアバタを表示させるアニメーションの例を示す説明図である。It is explanatory drawing which shows the example of the animation which the display control part 135 which concerns on the same modification displays the avatar of a some remote user. ハードウェア構成例を示す説明図である。It is explanatory drawing which shows the hardware structural example.
 以下に添付図面を参照しながら、本開示の好適な実施の形態について詳細に説明する。なお、本明細書及び図面において、実質的に同一の機能構成を有する構成要素については、同一の符号を付することにより重複説明を省略する。 Hereinafter, preferred embodiments of the present disclosure will be described in detail with reference to the accompanying drawings. In addition, in this specification and drawing, about the component which has the substantially same function structure, duplication description is abbreviate | omitted by attaching | subjecting the same code | symbol.
 また、本明細書及び図面において、実質的に同一の機能構成を有する複数の構成要素を、同一の符号の後に異なるアルファベットを付して区別する場合もある。ただし、実質的に同一の機能構成を有する複数の構成要素の各々を特に区別する必要がない場合、同一符号のみを付する。 In the present specification and drawings, a plurality of constituent elements having substantially the same functional configuration may be distinguished by adding different alphabets after the same reference numeral. However, when it is not necessary to particularly distinguish each of a plurality of constituent elements having substantially the same functional configuration, only the same reference numerals are given.
 なお、説明は以下の順序で行うものとする。
 <<1.概要>>
 <<2.構成>>
  <2-1.全体構成>
  <2-2.ローカル端末の構成>
  <2-3.リモート端末の構成>
 <<3.動作>>
 <<4.変形例>>
  <4-1.変形例1>
  <4-2.変形例2>
  <4-3.変形例3>
  <4-4.変形例4>
 <<5.ハードウェア構成例>>
 <<6.むすび>>
The description will be made in the following order.
<< 1. Overview >>
<< 2. Configuration >>
<2-1. Overall configuration>
<2-2. Configuration of local terminal>
<2-3. Configuration of remote terminal>
<< 3. Operation >>
<< 4. Modification >>
<4-1. Modification 1>
<4-2. Modification 2>
<4-3. Modification 3>
<4-4. Modification 4>
<< 5. Hardware configuration example >>
<< 6. Conclusion >>
 <<1.概要>>
 まず、図1、図2を参照ながら本開示の一実施形態の概要を説明する。図1、及び図2は、本開示の一実施形態に係るコミュニケーションシステムの概要を説明するための説明図である。本実施形態に係るコミュニケーションシステムは、図1に示すローカル環境1000に存在するローカルユーザLU1と、リモート環境2000に存在するリモートユーザRU1との間のリモートコミュニケーションを実現する情報処理システムである。
<< 1. Overview >>
First, an outline of an embodiment of the present disclosure will be described with reference to FIGS. 1 and 2. 1 and 2 are explanatory diagrams for describing an overview of a communication system according to an embodiment of the present disclosure. The communication system according to the present embodiment is an information processing system that realizes remote communication between a local user LU1 existing in the local environment 1000 shown in FIG. 1 and a remote user RU1 existing in the remote environment 2000.
 ローカル環境1000、及びリモート環境2000は実物体が存在し得る実空間内の任意の環境であってよい。また、ローカル環境1000とリモート環境2000とは、遠く離れていてもよく、例えば直接対面してのコミュニケーションが困難な状況であってもよい。 The local environment 1000 and the remote environment 2000 may be any environment in a real space where a real object may exist. Further, the local environment 1000 and the remote environment 2000 may be far away, and for example, it may be a situation where direct face-to-face communication is difficult.
 図1に示すローカル端末10は、ローカル環境1000に存在し、ローカルユーザLU1が用いる情報処理装置である。図1に示す例において、ローカル端末10はローカルユーザLU1の目の一方、または双方の前方に配置される透過型(光学シースルー型)の表示部を有し、ローカルユーザLU1の頭部に装着された眼鏡型デバイスである。 The local terminal 10 shown in FIG. 1 is an information processing apparatus that exists in the local environment 1000 and is used by the local user LU1. In the example shown in FIG. 1, the local terminal 10 has a transmissive (optical see-through) display unit arranged in front of one or both eyes of the local user LU1, and is mounted on the head of the local user LU1. Glasses-type device.
 図1に示すリモート端末20は、リモート環境2000に存在し、リモートユーザRU1が用いる情報処理装置である。図1に示す例において、リモート端末20はリモートユーザRU1の目の一方、または双方の前方に配置される表示部を有し、リモートユーザRU1の頭部に装着された没入型HMD(Head Mounted Display)である。 The remote terminal 20 shown in FIG. 1 is an information processing apparatus that exists in the remote environment 2000 and is used by the remote user RU1. In the example shown in FIG. 1, the remote terminal 20 has a display unit arranged in front of one or both of the eyes of the remote user RU1, and is an immersive HMD (Head Mounted Display) attached to the head of the remote user RU1. ).
 図1に示す撮像装置30は、撮像により上下左右全方位の360度の全天球画像を取得することが可能な所謂全天球カメラである。なお、本明細書において、画像とは静止画像に限定されず、動画像を含んだ表現として用いられる。 The imaging device 30 shown in FIG. 1 is a so-called omnidirectional camera that can acquire 360-degree omnidirectional images in all directions in the vertical and horizontal directions by imaging. In this specification, an image is not limited to a still image, but is used as an expression including a moving image.
 本実施形態に係るローカル端末10、リモート端末20、及び撮像装置30は、互いに不図示の通信網を介して接続される。また本実施形態に係るコミュニケーションシステムはローカル端末10とリモート端末20との間で、例えばメッセージの相互通信機能、及びメッセージの出力(表示または音響出力)機能に加え、以下に説明する表示機能によるリモートコミュニケーションを実現する。なお、リモートコミュニケーションのために用いられる装置は図1に示した例に限定されず、他の装置の例については後述する。 The local terminal 10, the remote terminal 20, and the imaging device 30 according to the present embodiment are connected to each other via a communication network (not shown). In addition, the communication system according to the present embodiment, for example, in addition to a message intercommunication function and a message output (display or sound output) function between the local terminal 10 and the remote terminal 20, can be remotely controlled by a display function described below. Realize communication. The apparatus used for remote communication is not limited to the example shown in FIG. 1, and examples of other apparatuses will be described later.
 図1に示すように、リモート端末20は、ローカル環境1000に存在する撮像装置30の撮像に基づく表示画像G20を表示する。例えば、リモート端末20は、撮像装置30の撮像により取得された全天球画像から、リモートユーザRU1の顔向きに応じた領域を切出して表示画像G20を生成してもよい。係る構成により、リモートユーザRU1は、顔向きを変えることで、撮像装置30を視点とし、上下左右全方位にローカル環境1000を観察することが可能である。また、図1に示すように、ローカルユーザLU1が表示画像G20に含まれるような向きに顔を向けることで、リモートユーザRU1はローカルユーザLU1と対面した感覚を得ることが可能となる。 As shown in FIG. 1, the remote terminal 20 displays a display image G <b> 20 based on the imaging of the imaging device 30 existing in the local environment 1000. For example, the remote terminal 20 may generate a display image G20 by cutting out an area corresponding to the face direction of the remote user RU1 from the omnidirectional image acquired by imaging of the imaging device 30. With such a configuration, the remote user RU1 can observe the local environment 1000 in all directions from the viewpoint of the imaging device 30 by changing the face orientation. Also, as shown in FIG. 1, the remote user RU1 can obtain a sense of facing the local user LU1 by turning his face so that the local user LU1 is included in the display image G20.
 一方、ローカル端末10は、リモートユーザRU1を示す仮想的なオブジェクト(以下、アバタと呼称する)を表示させる。上述したように、ローカル端末10は透過型の表示部(不図示)を有しているため、ローカル端末10を装着したローカルユーザLU1は、ローカル端末10の表示部を通して実空間と共にアバタを視認することが可能である。そのため、ローカル端末10は、あたかも目の前の実空間にリモートユーザRU1のアバタが存在するかのような感覚を得ることが可能となる。 On the other hand, the local terminal 10 displays a virtual object (hereinafter referred to as an avatar) indicating the remote user RU1. As described above, since the local terminal 10 has a transmissive display unit (not shown), the local user LU1 wearing the local terminal 10 visually recognizes the avatar together with the real space through the display unit of the local terminal 10. It is possible. Therefore, the local terminal 10 can obtain a feeling as if the avatar of the remote user RU1 exists in the real space in front of you.
 図1に示すようにローカルユーザLU1がローカル端末10の表示部を通して見た視界V10には、実空間と共にリモートユーザRU1のアバタA1が含まれる。また、本実施形態に係るローカル端末10は、実空間における撮像装置30との相対的な位置関係を示す相対位置情報に基づいて、撮像装置30の位置に応じた実空間における位置に、アバタA1が視認されるようにアバタA1を透過型の表示部に表示する。例えば、図1に示す視界V10では、アバタA1が撮像装置30に重畳されてローカルユーザLU1に視認されている。 As shown in FIG. 1, the field of view V10 viewed by the local user LU1 through the display unit of the local terminal 10 includes the avatar A1 of the remote user RU1 together with the real space. Further, the local terminal 10 according to the present embodiment has the avatar A1 at the position in the real space according to the position of the imaging device 30 based on the relative position information indicating the relative positional relationship with the imaging device 30 in the real space. Is displayed on the transmissive display unit. For example, in the field of view V10 shown in FIG. 1, the avatar A1 is superimposed on the imaging device 30 and is visually recognized by the local user LU1.
 係る構成により、ローカルユーザLU1は、アバタA1を見ようとした場合に、撮像装置30へ視線を向けることになる。その結果、リモートユーザRU1が、リモート端末20に表示される表示画像G20中のローカルユーザLU1を見た場合に、リモートユーザRU1は、ローカルユーザLU1と視線が合っているように感じる、という効果が得られる。 With this configuration, the local user LU1 turns his line of sight toward the imaging device 30 when he / she tries to see the avatar A1. As a result, when the remote user RU1 sees the local user LU1 in the display image G20 displayed on the remote terminal 20, the remote user RU1 feels that the line of sight is aligned with the local user LU1. can get.
 また、上述したように、ローカルユーザLU1は、あたかもリモートユーザRU1が撮像装置30の位置に存在し、リモートユーザRU1が撮像装置30の位置からローカル環境1000を観察しているように感じる。上述したように、リモートユーザRU1は撮像装置30の撮像に基づく画像を見ているため、上記のようにローカルユーザLU1が感じることで、より円滑なコミュニケーションが実現され得る。 Further, as described above, the local user LU1 feels as if the remote user RU1 exists at the position of the imaging device 30, and the remote user RU1 observes the local environment 1000 from the position of the imaging device 30. As described above, since the remote user RU1 sees an image based on the imaging of the imaging device 30, the smooth communication can be realized by the local user LU1 feeling as described above.
 また、ローカル端末10は、リモートユーザRU1の顔向きに応じて、表示するアバタA1の顔A11の向き(顔向き)を変化させてもよい。リモートユーザRU1の顔向きに関する情報は、例えばリモート端末20が有するセンサにより取得され、ローカル端末10へ提供されてもよい。 Also, the local terminal 10 may change the orientation (face orientation) of the face A11 of the avatar A1 to be displayed according to the face orientation of the remote user RU1. Information regarding the face orientation of the remote user RU1 may be acquired by a sensor included in the remote terminal 20, for example, and provided to the local terminal 10.
 リモートユーザRU1が顔を右側に向けた場合のローカル端末10、及びリモート端末20の表示例を図2に示す。上述したようにリモートユーザRU1の顔向きに応じて、リモート端末20が全天球画像から表示画像を切り出す領域も変更されるため、図1と図2を比較すると、表示画像G20が変化している。また、図1と図2を比較すると、リモートユーザRU1の顔向きに応じて、ローカル端末10が表示するアバタA1の顔A11の向きも変化している。ローカルユーザLU1は、アバタA1の顔A11の向きを確認することで、リモートユーザRU1の視界に相当する撮像装置30における画角、つまり、リモートユーザRU1がローカル環境1000のどの範囲を見ているのかを把握することが可能となる。係る構成により、ローカルユーザLU1は、リモートユーザRU1が見ている景色を把握しながらコミュニケーションを行うことが可能となり、より円滑にコミュニケーションを行うことが可能となる。 FIG. 2 shows a display example of the local terminal 10 and the remote terminal 20 when the remote user RU1 faces the right side. As described above, according to the face orientation of the remote user RU1, the region where the remote terminal 20 cuts out the display image from the omnidirectional image is also changed. Therefore, comparing FIG. 1 with FIG. Yes. Further, comparing FIG. 1 with FIG. 2, the orientation of the face A11 of the avatar A1 displayed by the local terminal 10 also changes according to the face orientation of the remote user RU1. By confirming the orientation of the face A11 of the avatar A1, the local user LU1 confirms the angle of view in the imaging device 30 corresponding to the field of view of the remote user RU1, that is, what range the remote user RU1 is viewing in the local environment 1000. Can be grasped. With this configuration, the local user LU1 can perform communication while grasping the scenery that the remote user RU1 is viewing, and can perform communication more smoothly.
 以上、本開示の一実施形態に係るコミュニケーションシステムの概要について説明した。続いて、本開示の一実施形態に係るコミュニケーションシステムが上述した効果を実現するための構成例について説明する。 Heretofore, an overview of the communication system according to an embodiment of the present disclosure has been described. Next, a configuration example for realizing the above-described effect by the communication system according to an embodiment of the present disclosure will be described.
 <<2.構成>>
 図3は、本実施形態に係るコミュニケーションシステムの構成例を示すブロック図である。図3に示すように、本実施形態に係るコミュニケーションシステム1は、ローカル端末10、リモート端末20、撮像装置30、センサ装置40、配信サーバ50、センサ装置60、及び通信網70を備える情報処理システムである。以下では、まずコミュニケーションシステム1の全体的な構成を説明した後に、ローカル端末10、及びリモート端末20のより詳細な構成について説明する。
<< 2. Configuration >>
FIG. 3 is a block diagram illustrating a configuration example of the communication system according to the present embodiment. As illustrated in FIG. 3, the communication system 1 according to the present embodiment includes an information processing system including a local terminal 10, a remote terminal 20, an imaging device 30, a sensor device 40, a distribution server 50, a sensor device 60, and a communication network 70. It is. In the following, after describing the overall configuration of the communication system 1, more detailed configurations of the local terminal 10 and the remote terminal 20 will be described.
  <2-1.全体構成>
 図3に示すように、ローカル端末10、撮像装置30、センサ装置40は、図1を参照して説明したローカル環境1000に存在し、リモート端末20、及びセンサ装置60は図1を参照して説明したリモート環境2000に存在する。
<2-1. Overall configuration>
As shown in FIG. 3, the local terminal 10, the imaging device 30, and the sensor device 40 exist in the local environment 1000 described with reference to FIG. 1, and the remote terminal 20 and the sensor device 60 refer to FIG. Exists in the remote environment 2000 described.
 本実施形態に係るローカル端末10は、図1を参照して説明したように、ローカル環境1000に存在するローカルユーザの頭部に装着される眼鏡型のデバイスである。また、ローカル端末10は実空間における撮像装置30の位置に応じた位置に、リモートユーザのアバタを表示する透過型の表示部を有する。なお、ローカル端末10のより詳細な構成については図4を参照して後述する。 The local terminal 10 according to the present embodiment is a glasses-type device that is worn on the head of a local user existing in the local environment 1000 as described with reference to FIG. In addition, the local terminal 10 has a transmissive display unit that displays a remote user's avatar at a position corresponding to the position of the imaging device 30 in real space. A more detailed configuration of the local terminal 10 will be described later with reference to FIG.
 本実施形態に係るリモート端末20は、図1を参照して説明したようにリモート環境2000に存在するリモートユーザの頭部に装着される没入型HMDである。また、リモート端末20は撮像装置30の撮像に基づく画像を表示する。なお、リモート端末20のより詳細な構成については図6を参照して後述する。 The remote terminal 20 according to the present embodiment is an immersive HMD attached to the head of a remote user existing in the remote environment 2000 as described with reference to FIG. Further, the remote terminal 20 displays an image based on the imaging of the imaging device 30. A more detailed configuration of the remote terminal 20 will be described later with reference to FIG.
 本実施形態に係る撮像装置30は、図1を参照して説明したように、撮像により上下左右全方位の360度の全天球画像を取得することが可能な全天球カメラである。撮像装置30は、例えば複数の撮像部を有し、複数の撮像部により得られた画像を合成する画像処理を行うことで、全天球画像を取得してもよい。撮像装置30は、通信網70を介して配信サーバ50へ全天球画像を送信する。なお、撮像装置30は画像処理機能を有していなくてもよく、係る場合には例えば撮像装置30に接続された他の情報処理装置や、後述する配信サーバ50が撮像装置30の代わりに係る画像処理機能を有していてもよい。 As described with reference to FIG. 1, the imaging device 30 according to the present embodiment is an omnidirectional camera that can acquire 360-degree omnidirectional images in all directions in the vertical and horizontal directions by imaging. The imaging device 30 may have a plurality of imaging units, for example, and may acquire an omnidirectional image by performing image processing for combining images obtained by the plurality of imaging units. The imaging device 30 transmits the omnidirectional image to the distribution server 50 via the communication network 70. Note that the imaging apparatus 30 may not have an image processing function. In such a case, for example, another information processing apparatus connected to the imaging apparatus 30 or a distribution server 50 described later is associated with the imaging apparatus 30 instead. It may have an image processing function.
 センサ装置40は、センシングによりローカル環境1000に関するセンシングデータを取得する。センサ装置40は、複数のセンサを含んでもよく、例えば、カメラ、赤外線カメラ、マイクロフォン、深度センサ、照度センサ、人感センサ等の各種センサを含んでもよい。センサ装置40は、取得したセンシングデータをローカル端末10に提供(送信)する。 The sensor device 40 acquires sensing data related to the local environment 1000 by sensing. The sensor device 40 may include a plurality of sensors, for example, various sensors such as a camera, an infrared camera, a microphone, a depth sensor, an illuminance sensor, and a human sensor. The sensor device 40 provides (transmits) the acquired sensing data to the local terminal 10.
 配信サーバ50は、通信網70を介して撮像装置30から受信した全天球画像を他の装置(例えばリモート端末20)へ配信(送信)する情報処理装置である。例えば、配信サーバ50は、撮像装置30から受信した全天球画像をキャッシュしながらストリーミング配信を行ってもよい。 The distribution server 50 is an information processing apparatus that distributes (transmits) the omnidirectional image received from the imaging apparatus 30 via the communication network 70 to another apparatus (for example, the remote terminal 20). For example, the distribution server 50 may perform streaming distribution while caching the omnidirectional image received from the imaging device 30.
 センサ装置60は、センシングによりリモート環境2000に関するセンシングデータを取得する。センサ装置60は、複数のセンサを含んでもよく、例えば、カメラ、赤外線カメラ、マイクロフォン、深度センサ、照度センサ、人感センサ等の各種センサを含んでもよい。センサ装置60は、取得したセンシングデータをリモート端末20に提供(送信)する。 The sensor device 60 acquires sensing data related to the remote environment 2000 by sensing. The sensor device 60 may include a plurality of sensors, for example, various sensors such as a camera, an infrared camera, a microphone, a depth sensor, an illuminance sensor, and a human sensor. The sensor device 60 provides (transmits) the acquired sensing data to the remote terminal 20.
 通信網70は、通信網70に接続されている装置、またはシステムから送信される情報の有線、または無線の伝送路である。例えば、通信網70は、インターネット、電話回線網、衛星通信網等の公衆回線網や、Ethernet(登録商標)を含む各種のLAN(Local Area Network)、WAN(Wide Area Network)等を含んでもよい。また、通信網70は、IP-VPN(Internet Protocol-Virtual Private Network)等の専用回線網を含んでもよい。 The communication network 70 is a wired or wireless transmission path for information transmitted from a device or system connected to the communication network 70. For example, the communication network 70 may include a public line network such as the Internet, a telephone line network, a satellite communication network, various LANs (Local Area Network) including Ethernet (registered trademark), WAN (Wide Area Network), and the like. . Further, the communication network 70 may include a dedicated line network such as IP-VPN (Internet Protocol-Virtual Private Network).
 以上、本実施形態に係るコミュニケーションシステム1の全体構成を説明した。続いて、コミュニケーションシステム1が有するローカル端末10、及びリモート端末20の構成例について順次説明する。 The overall configuration of the communication system 1 according to the present embodiment has been described above. Next, configuration examples of the local terminal 10 and the remote terminal 20 included in the communication system 1 will be sequentially described.
  <2-2.ローカル端末の構成>
 図4は、本実施形態に係るローカル端末10の構成例を示すブロック図である。図4に示すように、本実施形態に係るローカル端末10は、センサ部11、通信部12、制御部13、表示部14、音響出力部15、及び記憶部16を備える情報処理装置である。
<2-2. Configuration of local terminal>
FIG. 4 is a block diagram illustrating a configuration example of the local terminal 10 according to the present embodiment. As illustrated in FIG. 4, the local terminal 10 according to the present embodiment is an information processing apparatus including a sensor unit 11, a communication unit 12, a control unit 13, a display unit 14, a sound output unit 15, and a storage unit 16.
 センサ部11は、センシングによりローカル端末10を装着したローカルユーザ、及び周辺環境(ローカル環境1000)に関するセンシングデータを取得する。センサ部11は、例えば加速度センサ、ジャイロセンサ、カメラ、マイクロフォン、地磁気センサ、力センサ等を含んでもよい。また、センサ部11により取得されるセンサデータは、ローカル端末10の位置や姿勢に関する情報を含んでいてもよい。センサ部11は、取得したセンシングデータを制御部13へ提供する。 The sensor unit 11 acquires sensing data regarding the local user wearing the local terminal 10 and the surrounding environment (local environment 1000) by sensing. The sensor unit 11 may include, for example, an acceleration sensor, a gyro sensor, a camera, a microphone, a geomagnetic sensor, a force sensor, and the like. The sensor data acquired by the sensor unit 11 may include information on the position and orientation of the local terminal 10. The sensor unit 11 provides the acquired sensing data to the control unit 13.
 通信部12は、ローカル端末10による他の装置との間の通信を仲介する通信インタフェースである。通信部12は、任意の無線通信プロトコルまたは有線通信プロトコルをサポートし、例えば図3を参照して説明した通信網70を介して、または直接に他の装置との間の通信接続を確立する。 The communication unit 12 is a communication interface that mediates communication between the local terminal 10 and other devices. The communication unit 12 supports an arbitrary wireless communication protocol or wired communication protocol, and establishes a communication connection with another device, for example, via the communication network 70 described with reference to FIG. 3 or directly.
 制御部13は、ローカル端末10の各構成の動作を制御する。また、制御部13は、図4に示すように、通信制御部131、相対位置取得部133、表示制御部135、及び音響出力制御部137としても機能する。 The control unit 13 controls the operation of each component of the local terminal 10. The control unit 13 also functions as a communication control unit 131, a relative position acquisition unit 133, a display control unit 135, and a sound output control unit 137, as shown in FIG.
 図4に示す通信制御部131は、通信部12による通信を制御し、様々な情報を他の装置から受信させることで取得し、または他の装置へ送信させる。例えば、通信制御部131は、リモート端末20へ、センサ部11に含まれるマイクロフォンにより取得された音声データをメッセージとして送信させる。また、通信制御部131は、不図示の入力装置を介して入力されたテキストデータをメッセージとしてリモート端末20へ送信させてもよい。また、通信制御部131は、リモート端末20からリモート端末20を装着したリモートユーザに関するユーザ情報を受信させてもよい。なお、ユーザ情報には、例えば、リモートユーザの識別情報、リモートユーザの状態を示す情報、リモートユーザの姿勢情報、リモートユーザにより発信されたメッセージ(テキストデータ、音声データ等)等が含まれてもよい。また、通信制御部131は、図3を参照して説明したセンサ装置40からセンシングデータを受信させてもよい。 The communication control unit 131 shown in FIG. 4 controls communication by the communication unit 12, and acquires various types of information from other devices, or transmits them to other devices. For example, the communication control unit 131 causes the remote terminal 20 to transmit voice data acquired by a microphone included in the sensor unit 11 as a message. Further, the communication control unit 131 may cause text data input via an input device (not shown) to be transmitted to the remote terminal 20 as a message. Further, the communication control unit 131 may receive user information regarding the remote user wearing the remote terminal 20 from the remote terminal 20. The user information may include, for example, remote user identification information, remote user status information, remote user attitude information, messages (text data, voice data, etc.) transmitted by the remote user, and the like. Good. Further, the communication control unit 131 may receive sensing data from the sensor device 40 described with reference to FIG.
 相対位置取得部133は、実空間における撮像装置30との相対的な位置関係を示す相対位置情報を取得する。本実施形態において、相対位置情報は、例えば現在のローカル端末10を基準とした座標系において表現された撮像装置30の位置を示す情報(座標)であってもよい。相対位置取得部133は、多様な方法で相対位置情報を取得し得る。以下に相対位置情報の取得方法について、いくつかの例を説明する。 The relative position acquisition unit 133 acquires relative position information indicating a relative positional relationship with the imaging device 30 in real space. In the present embodiment, the relative position information may be information (coordinates) indicating the position of the imaging device 30 expressed in a coordinate system based on the current local terminal 10, for example. The relative position acquisition unit 133 can acquire the relative position information by various methods. Several examples of the relative position information acquisition method will be described below.
 相対位置取得部133は、センサ部11に含まれるカメラにより取得された画像から、撮像装置30を検出することで、相対位置情報を取得してもよい。係る場合、撮像装置30の検出を容易にするため、撮像装置30が検出用のマーカを備えてもよいし、撮像装置30が発光部を備え、当該発光部が所定の発光パターンで発光してもよい。 The relative position acquisition unit 133 may acquire the relative position information by detecting the imaging device 30 from the image acquired by the camera included in the sensor unit 11. In this case, in order to facilitate detection of the imaging device 30, the imaging device 30 may include a detection marker, or the imaging device 30 may include a light emitting unit, and the light emitting unit may emit light with a predetermined light emission pattern. Also good.
 または、相対位置取得部133は、撮像装置30により取得された全天球画像から、ローカル端末10を画像認識により検出することで、相対位置情報を取得してもよい。係る場合、ローカル端末10の検出を容易にするため、ローカル端末10が検出用のマーカを備えてもよいし、ローカル端末10が発光部を備え、当該発光部が所定の発光パターンで発光してもよい。なお、ローカル端末10は、撮像装置30から直接的に、または通信網70や配信サーバ50を介して全天球画像を取得してもよい。 Alternatively, the relative position acquisition unit 133 may acquire the relative position information by detecting the local terminal 10 by image recognition from the omnidirectional image acquired by the imaging device 30. In this case, in order to facilitate detection of the local terminal 10, the local terminal 10 may be provided with a detection marker, or the local terminal 10 may include a light emitting unit, and the light emitting unit may emit light with a predetermined light emission pattern. Also good. Note that the local terminal 10 may acquire an omnidirectional image directly from the imaging device 30 or via the communication network 70 or the distribution server 50.
 または、相対位置取得部133は、絶対的な座標系における撮像装置30の座標(以下、絶対座標と呼ぶ)と、ローカル端末10の位置、及び姿勢の情報とに基づいて、相対位置情報を取得することも可能である。ローカル端末10の位置、及び姿勢の情報は、センサ部11が取得したセンシングデータに含まれてもよいし、センシングデータに基づいて、SLAM(Simultaneous Localization and Mapping)等の自己位置推定技術を用いて特定されてもよい。 Alternatively, the relative position acquisition unit 133 acquires relative position information based on the coordinates of the imaging device 30 in the absolute coordinate system (hereinafter referred to as absolute coordinates) and the position and orientation information of the local terminal 10. It is also possible to do. The position and orientation information of the local terminal 10 may be included in the sensing data acquired by the sensor unit 11, or based on the sensing data, using self-position estimation technology such as SLAM (Simultaneous Localization and Mapping). It may be specified.
 なお、撮像装置30の絶対座標は、例えば後述する記憶部16に記憶されていてもよいし、通信網70を介して他の装置(例えば撮像装置30)から取得されてもよい。 The absolute coordinates of the imaging device 30 may be stored, for example, in the storage unit 16 described later, or may be acquired from another device (for example, the imaging device 30) via the communication network 70.
 もしくは、相対位置取得部133は、センサ装置40に含まれるカメラにより取得された画像から、撮像装置30を検出して、撮像装置30の絶対座標を特定してもよい。係る場合、撮像装置30には検出用のマーカが設置されてもよいし、撮像装置30が発光部を備えて発光パターンにより検出し易くしてもよい。あるいは、相対位置取得部133は、撮像装置30により取得された全天球画像と、予めローカル環境1000を撮影して得られた画像とのマッチングにより、撮像装置30の絶対座標を特定してもよい。このように動的に撮像装置30の絶対座標を特定することで、撮像装置30が移動した場合であっても、撮像装置30の絶対座標を特定して、相対位置情報を取得することが可能である。 Alternatively, the relative position acquisition unit 133 may detect the imaging device 30 from the image acquired by the camera included in the sensor device 40 and specify the absolute coordinates of the imaging device 30. In such a case, a detection marker may be provided in the imaging device 30, or the imaging device 30 may include a light emitting unit to facilitate detection by a light emission pattern. Alternatively, the relative position acquisition unit 133 may specify the absolute coordinates of the imaging device 30 by matching the omnidirectional image acquired by the imaging device 30 with an image obtained by capturing the local environment 1000 in advance. Good. By dynamically specifying the absolute coordinates of the imaging device 30 in this way, even when the imaging device 30 moves, it is possible to specify the absolute coordinates of the imaging device 30 and acquire the relative position information. It is.
 以上、相対位置取得部133による相対位置情報の取得方法についていくつか説明したが、上記以外の方法により相対位置情報が取得されてもよい。例えば、センサ部11で取得される画像以外のセンシングデータから撮像装置30を検出することで相対位置情報が取得されてもよい。または、他の装置により特定された相対位置情報を通信部12で受信することにより、相対位置情報が取得されてもよい。 As mentioned above, although several methods for acquiring relative position information by the relative position acquisition unit 133 have been described, the relative position information may be acquired by a method other than the above. For example, the relative position information may be acquired by detecting the imaging device 30 from sensing data other than the image acquired by the sensor unit 11. Alternatively, the relative position information may be acquired by the communication unit 12 receiving the relative position information specified by another device.
 表示制御部135は、相対位置情報に基づいて、実空間における撮像装置30の位置に応じた実空間における位置に、リモートユーザを示すオブジェクトがローカルユーザにより視認されるように、透過型の表示部14による表示を制御する。リモートユーザを示すオブジェクトは、例えば仮想的なアバタであってもよいし、図3を参照して説明したセンサ装置60に含まれるカメラにより撮影されたリモートユーザの画像であってもよい。なお、以下では、表示制御部135がリモートユーザを示すオブジェクトとして、リモートユーザのアバタを表示する例について説明する。 Based on the relative position information, the display control unit 135 is a transmissive display unit so that the local user can visually recognize the object indicating the remote user at a position in the real space corresponding to the position of the imaging device 30 in the real space. 14 controls display. The object indicating the remote user may be, for example, a virtual avatar, or may be an image of the remote user captured by the camera included in the sensor device 60 described with reference to FIG. Hereinafter, an example in which the display control unit 135 displays a remote user's avatar as an object indicating the remote user will be described.
 表示制御部135は、例えば図1を参照して説明したように、実空間における撮像装置30とアバタとが重畳されてローカルユーザにより視認されるようにリモートユーザのアバタを表示させてもよい。なお、撮像装置30とアバタとが重畳されて視認されるとは、撮像装置30のうち少なくとも一部がアバタの少なくとも一部と重なってローカルユーザにより視認されることを意味する。 For example, as described with reference to FIG. 1, the display control unit 135 may display the avatar of the remote user so that the imaging device 30 and the avatar in the real space are superimposed and viewed by the local user. Note that “the imaging device 30 and the avatar are superimposed and visually recognized” means that at least a part of the imaging device 30 overlaps at least a part of the avatar and is visually recognized by the local user.
 係る構成により、ローカルユーザはリモートユーザのアバタを見ようとした場合に撮像装置30へ視線を向けることとなる。そして、リモートユーザがリモート端末20に表示される画像中のローカルユーザを見た場合に、リモートユーザはローカルユーザと視線が合っているように感じる、という効果が得られる。 With this configuration, the local user turns his / her line of sight toward the imaging device 30 when he / she tries to see the avatar of the remote user. Then, when the remote user sees the local user in the image displayed on the remote terminal 20, an effect is obtained in which the remote user feels that his / her line of sight matches the local user.
 また、ローカルユーザは、あたかもリモートユーザが撮像装置30の位置に存在し、リモートユーザが撮像装置30の位置を視点としてローカル環境1000を観察しているように感じる。図1を参照して説明したように、リモートユーザは撮像装置30の撮像に基づく画像(リモート端末20に表示される画像)を見ているため、上記のようにローカルユーザが感じることで、より円滑なコミュニケーションが実現され得る。 Further, the local user feels as if the remote user is present at the position of the imaging device 30 and the remote user observes the local environment 1000 with the position of the imaging device 30 as a viewpoint. As described with reference to FIG. 1, since the remote user sees an image based on the imaging of the imaging device 30 (an image displayed on the remote terminal 20), the local user can feel more as described above. Smooth communication can be realized.
 また、リモートユーザのアバタが目(目であるように見える部分)を含む場合、表示制御部135は、撮像装置30の撮像部(不図示)とアバタに含まれる目とが重畳されて視認されるように、アバタを表示させてもよい。係る構成により、ローカルユーザがリモートユーザの視点をより想像し易くなり、より円滑なコミュニケーションが実現され得る。 When the remote user's avatar includes eyes (portion that looks like eyes), the display control unit 135 is visually recognized by superimposing an imaging unit (not shown) of the imaging device 30 and the eyes included in the avatar. As shown, an avatar may be displayed. With such a configuration, the local user can more easily imagine the viewpoint of the remote user, and smoother communication can be realized.
 また、表示制御部135は、通信部12を介してリモート端末20から受信したリモートユーザの姿勢を示す姿勢情報に基づいて、アバタの姿勢を制御してもよい。例えば、姿勢情報にリモートユーザの顔向きに関する情報が含まれ、アバタが顔(顔であるように見える部分)を含む場合、表示制御部135は、当該顔向きに関する情報に応じて、アバタに含まれる顔の向きを制御してもよい。 Further, the display control unit 135 may control the attitude of the avatar based on the attitude information indicating the attitude of the remote user received from the remote terminal 20 via the communication unit 12. For example, when the posture information includes information related to the face orientation of the remote user and the avatar includes a face (a part that appears to be a face), the display control unit 135 includes the information related to the face orientation. The direction of the face to be displayed may be controlled.
 係る構成により、例えば図1、図2を参照して説明したように、ローカルユーザは、アバタの顔向きを確認することで、リモートユーザがローカル環境1000のどの範囲を見ているのかを把握することが可能となる。そして、ローカルユーザは、リモートユーザが見ている景色を把握しながらコミュニケーションを行うことが可能となり、より円滑にコミュニケーションを行うことが可能となる。 With this configuration, for example, as described with reference to FIGS. 1 and 2, the local user confirms the range of the local environment 1000 by the remote user by confirming the face direction of the avatar. It becomes possible. Then, the local user can communicate while grasping the scenery viewed by the remote user, and can communicate more smoothly.
 なお、表示制御部135による姿勢の制御は顔向きの制御に限定されない。例えば、姿勢情報に含まれる情報、及びアバタに含まれる人体に対応する部分の姿勢が制御され得る。例えば、姿勢情報にリモートユーザの手、腕、身体の姿勢が含まれる場合、表示制御部135は、アバタの手、腕、身体の姿勢をそれぞれ制御してもよい。係る構成により、ローカルユーザは、あたかもリモートユーザがローカル環境1000に存在するように、より強く感じることが可能となる。 Note that posture control by the display control unit 135 is not limited to face orientation control. For example, the information included in the posture information and the posture of the part corresponding to the human body included in the avatar can be controlled. For example, when the posture information includes a remote user's hand, arm, and body posture, the display control unit 135 may control the avatar's hand, arm, and body posture, respectively. With such a configuration, the local user can feel stronger as if the remote user exists in the local environment 1000.
 また、表示制御部135は、リモートユーザに応じたアバタを表示させてもよい。例えば、複数のアバタがリモートユーザの識別情報と対応付けられて記憶部16に記憶され、表示制御部135は、通信部12を介してリモート端末20から受信したリモートユーザの識別情報に応じてアバタを選択し、選択されたアバタを表示させてもよい。係る構成により、ローカルユーザは、アバタを介してリモートユーザを識別することが可能となる。 Further, the display control unit 135 may display an avatar corresponding to the remote user. For example, a plurality of avatars are associated with remote user identification information and stored in the storage unit 16, and the display control unit 135 responds to the remote user identification information received from the remote terminal 20 via the communication unit 12. May be selected to display the selected avatar. With this configuration, the local user can identify the remote user via the avatar.
 また、表示制御部135は、リモートユーザにより発信されるメッセージを表示させてもよい。なお、リモートユーザにより発信されるメッセージは、例えば通信部12を介してリモート端末20から受信される。また、表示制御部135が表示させるメッセージはテキストに限定されず、例えば画像データ、ドキュメントデータ等の電子データを含んでもよい。係る場合、表示制御部135は、電子データを示すアイコンを表示させてもよい。なお、電子データが画像データである場合、表示制御部135が表示させるアイコンは、画像であってもよい。 Further, the display control unit 135 may display a message transmitted by the remote user. Note that the message transmitted by the remote user is received from the remote terminal 20 via the communication unit 12, for example. The message displayed by the display control unit 135 is not limited to text, and may include electronic data such as image data and document data. In such a case, the display control unit 135 may display an icon indicating electronic data. When the electronic data is image data, the icon displayed by the display control unit 135 may be an image.
 図5は、表示制御部135がメッセージを表示させた場合のローカルユーザの視界V10を示す模式図である。図5に示すように、表示制御部135はメッセージM10をアバタA1の近傍に表示させている。また、図5に示す例では、表示制御部135はアバタA1の位置からの吹き出しとしてメッセージM10を表示させている。係る構成により、ローカルユーザは、リモートユーザが発したメッセージを把握しやすい。また、メッセージに電子データが含まれる場合、図5に示すように、表示制御部135は電子データを示すアイコンM11を表示させてもよい。 FIG. 5 is a schematic diagram showing the field of view V10 of the local user when the display control unit 135 displays a message. As shown in FIG. 5, the display control unit 135 displays the message M10 in the vicinity of the avatar A1. In the example shown in FIG. 5, the display control unit 135 displays the message M10 as a balloon from the position of the avatar A1. With this configuration, the local user can easily grasp the message issued by the remote user. When electronic data is included in the message, the display control unit 135 may display an icon M11 indicating electronic data as shown in FIG.
 また、表示制御部135は、リモートユーザの状態に応じて、表示を制御してもよい。例えば、表示制御部135は、リモートユーザの状態を示すインジケータ(例えばアイコンやテキスト)をリモートユーザのアバタの近傍に表示させてもよい。また、表示制御部135は、リモートユーザの状態に応じてアバタの表示に関するパラメータを制御してもよい。アバタの表示に関するパラメータは、例えば輝度、色、彩度、透過度、姿勢等を含んでもよい。 Further, the display control unit 135 may control the display according to the state of the remote user. For example, the display control unit 135 may display an indicator (for example, an icon or text) indicating the state of the remote user in the vicinity of the remote user's avatar. Further, the display control unit 135 may control parameters related to the display of the avatar according to the state of the remote user. Parameters relating to avatar display may include, for example, luminance, color, saturation, transparency, posture, and the like.
 なお、リモートユーザの状態を示す情報は、通信部12を介してリモート端末20から受信されてよい。また、リモートユーザの状態を示す情報は、リモートユーザがオンライン状態であるか否か、リモートユーザがリモート端末20を装着しているか否か、リモートユーザがアバタの表示を許可したか否か等の情報を含んでもよい。 Note that information indicating the status of the remote user may be received from the remote terminal 20 via the communication unit 12. The information indicating the status of the remote user includes whether the remote user is online, whether the remote user is wearing the remote terminal 20, whether the remote user has permitted display of the avatar, and the like. Information may be included.
 係る構成により、ローカルユーザはリモートユーザの状態を把握することが可能となる。 With this configuration, the local user can grasp the status of the remote user.
 また、表示制御部135は、全天球カメラの映像が正常にリモート端末20へ送信されているか否かの情報を配信サーバ50または撮像装置30から受信し、当該情報に基づいて表示を制御してもよい。係る場合、表示制御部135は全天球カメラの映像が正常にリモート端末20へ送信されているか否かを示すアイコンやテキストを表示させてもよいし、アバタの表示に関するパラメータを制御してもよい。 In addition, the display control unit 135 receives information on whether or not the image of the omnidirectional camera is normally transmitted to the remote terminal 20 from the distribution server 50 or the imaging device 30, and controls display based on the information. May be. In such a case, the display control unit 135 may display an icon or text indicating whether or not the image of the omnidirectional camera is normally transmitted to the remote terminal 20, or may control a parameter related to the display of the avatar. Good.
 また、表示制御部135は、ローカル端末10とリモート端末20との間の音声通信が正常に行われているか否かに応じて、表示を制御してもよい。係る場合、表示制御部135ローカル端末10とリモート端末20との間の音声通信が正常に行われているか否かを示すアイコンやテキストを表示させてもよいし、アバタの表示に関するパラメータを制御してもよい。 Further, the display control unit 135 may control the display according to whether or not the voice communication between the local terminal 10 and the remote terminal 20 is normally performed. In such a case, the display control unit 135 may display an icon or text indicating whether or not the voice communication between the local terminal 10 and the remote terminal 20 is normally performed, and controls parameters related to the display of the avatar. May be.
 図4に示す音響出力制御部137は、音響出力部15による音響出力を制御する。例えば、音響出力制御部137は、音響出力部15を制御して、リモートユーザにより発信されるメッセージを音響出力させてもよい。また、音響出力制御部137は、アバタがローカルユーザによって視認される実空間における位置からメッセージが聞こえるように、音響出力を制御してもよい。係る構成により、ローカルユーザは、あたかもリモートユーザがローカル環境1000に存在するかのような感覚をより強く得ることが可能となる。 4 controls the sound output from the sound output unit 15. The sound output control unit 137 illustrated in FIG. For example, the sound output control unit 137 may control the sound output unit 15 to sound output a message transmitted by the remote user. In addition, the sound output control unit 137 may control the sound output so that the message can be heard from the position in the real space where the avatar is visually recognized by the local user. With this configuration, the local user can obtain a stronger feeling as if the remote user exists in the local environment 1000.
 表示部14は表示制御部135に制御されて、アバタ等のオブジェクトを含む種々の情報を表示するディスプレイである。表示部14は、図1に示したように、透過型(光学シースルー型)のディスプレイであってもよい。係る構成により、ローカル端末10を装着するローカルユーザは、実空間と表示部14に表示された情報とを同時に視認することが可能である。なお、表示部14は非透過の表示部であってもよく、係る場合については変形例として後述する。 The display unit 14 is a display that is controlled by the display control unit 135 to display various information including objects such as avatars. As shown in FIG. 1, the display unit 14 may be a transmissive (optical see-through) display. With such a configuration, the local user wearing the local terminal 10 can view the real space and the information displayed on the display unit 14 at the same time. The display unit 14 may be a non-transparent display unit, and such a case will be described later as a modified example.
 また、表示部14は、例えばローカル端末10を装着するローカルユーザの両目に対してそれぞれ別の画像を提示することが可能であってもよく、係る構成によりローカルユーザは表示されたオブジェクトの形状や位置を立体的に認識することが可能である。 Further, the display unit 14 may be able to present different images to both eyes of the local user wearing the local terminal 10, for example, and the local user can display the shape of the displayed object, It is possible to recognize the position three-dimensionally.
 音響出力部15は音響出力制御部137に制御されて音響出力を行う。音響出力部15は、例えば複数のスピーカを含んでよく、音響出力部15は立体的に音響(立体音響)を出力することが可能であってもよい。 The sound output unit 15 performs sound output under the control of the sound output control unit 137. The sound output unit 15 may include, for example, a plurality of speakers, and the sound output unit 15 may be capable of outputting sound in a three-dimensional manner (stereo sound).
 記憶部16は、制御部13が上記各処理を実行するためのプログラムや、様々なデータを記憶する。例えば、記憶部16は、上述したアバタの情報や、通信部12を介してリモート端末20から受信した情報の履歴を記憶してもよい。なお、記憶部16の機能は、外部の装置に存在してもよく、ローカル端末10は例えば通信部12を介して、外部の装置が有する記憶部に記憶された情報を受信してもよい。 The storage unit 16 stores a program for the control unit 13 to execute the above processes and various data. For example, the storage unit 16 may store the avatar information described above and a history of information received from the remote terminal 20 via the communication unit 12. The function of the storage unit 16 may exist in an external device, and the local terminal 10 may receive information stored in a storage unit included in the external device via the communication unit 12, for example.
 以上、図4を参照してローカル端末10の構成例について説明したが、図4に示したのは一例であって、本技術は係る例に限定されない。例えば、図4に示す一部の機能がローカル端末10に接続された他の情報処理装置に備えられてもよい。例えば、制御部13の一部またはすべての機能がローカル端末10に接続された他の情報処理装置に備えられ、ローカル端末10は当該他の情報処理装置により制御されて、表示、または音響出力を行ってもよい。なお、係る場合、相対位置情報は、表示を行うローカル端末10と撮像装置30との相対的な位置を示す情報であってよい。 The configuration example of the local terminal 10 has been described above with reference to FIG. 4, but the example illustrated in FIG. 4 is an example, and the present technology is not limited to the example. For example, some functions shown in FIG. 4 may be provided in another information processing apparatus connected to the local terminal 10. For example, some or all of the functions of the control unit 13 are provided in another information processing apparatus connected to the local terminal 10, and the local terminal 10 is controlled by the other information processing apparatus to display or output sound. You may go. In this case, the relative position information may be information indicating a relative position between the local terminal 10 that performs display and the imaging device 30.
  <2-3.リモート端末の構成>
 図6は、本実施形態に係るリモート端末20の構成例を示すブロック図である。図6に示すように、本実施形態に係るリモート端末20は、センサ部21、通信部22、制御部23、表示部24、音響出力部25、及び記憶部26を備える情報処理装置である。
<2-3. Configuration of remote terminal>
FIG. 6 is a block diagram illustrating a configuration example of the remote terminal 20 according to the present embodiment. As illustrated in FIG. 6, the remote terminal 20 according to the present embodiment is an information processing apparatus including a sensor unit 21, a communication unit 22, a control unit 23, a display unit 24, a sound output unit 25, and a storage unit 26.
 センサ部21は、センシングによりリモート端末20を装着したリモートユーザ、及び周辺環境(リモート環境2000)に関するセンシングデータを取得する。センサ部21は、例えば加速度センサ、ジャイロセンサ、カメラ、マイクロフォン、地磁気センサ、力センサ等を含んでもよい。また、センサ部21により取得されるセンサデータは、リモートユーザの姿勢(例えば顔向き等)に関する姿勢情報を含んでいてもよい。センサ部21は、取得したセンシングデータを制御部23へ提供する。 The sensor unit 21 acquires sensing data related to the remote user wearing the remote terminal 20 and the surrounding environment (remote environment 2000) by sensing. The sensor unit 21 may include, for example, an acceleration sensor, a gyro sensor, a camera, a microphone, a geomagnetic sensor, a force sensor, and the like. The sensor data acquired by the sensor unit 21 may include posture information related to the posture (for example, the face orientation) of the remote user. The sensor unit 21 provides the acquired sensing data to the control unit 23.
 通信部22は、リモート端末20による他の装置との間の通信を仲介する通信インタフェースである。通信部22は、任意の無線通信プロトコルまたは有線通信プロトコルをサポートし、例えば図3を参照して説明した通信網70を介して、または直接に他の装置との間の通信接続を確立する。 The communication unit 22 is a communication interface that mediates communication between the remote terminal 20 and other devices. The communication unit 22 supports an arbitrary wireless communication protocol or wired communication protocol, and establishes a communication connection with another device, for example, via the communication network 70 described with reference to FIG.
 制御部23は、リモート端末20の各構成の動作を制御する。また、制御部23は、図6に示すように、通信制御部231、表示制御部235、及び音響出力制御部237としても機能する。 The control unit 23 controls the operation of each component of the remote terminal 20. Further, as shown in FIG. 6, the control unit 23 also functions as a communication control unit 231, a display control unit 235, and a sound output control unit 237.
 図6に示す通信制御部231は、通信部22による通信を制御し、様々な情報を他の装置から受信させることで取得し、または他の装置へ送信させる。例えば、通信制御部231は、リモートユーザの状態を示す情報、リモートユーザの姿勢情報、リモートユーザにより発信されたメッセージ(テキストデータ、音声データ等)等を含むユーザ情報をローカル端末10へ送信させる。なお、通信制御部231は、センサ部21に含まれるマイクロフォンにより取得された音声データをメッセージとしてローカル端末10へ送信させてもよいし、不図示の入力装置を介して入力されたテキストデータをメッセージとしてローカル端末10へ送信させてもよい。また、通信制御部231は、ローカル端末10からローカルユーザにより発信されたメッセージ(テキストデータ、音声データ等)等を受信させてもよい。また、通信制御部231は、図3を参照して説明した配信サーバ50から全天球画像を受信させてもよい。また、通信制御部231は、図3を参照して説明したセンサ装置60からセンシングデータを受信させてもよい。 The communication control unit 231 illustrated in FIG. 6 controls communication by the communication unit 22 and acquires various types of information from other devices, or transmits them to other devices. For example, the communication control unit 231 transmits user information including information indicating the status of the remote user, attitude information of the remote user, messages (text data, voice data, and the like) transmitted by the remote user to the local terminal 10. Note that the communication control unit 231 may cause the voice data acquired by the microphone included in the sensor unit 21 to be transmitted as a message to the local terminal 10, or text data input via an input device (not shown) as a message. May be transmitted to the local terminal 10. Further, the communication control unit 231 may receive a message (text data, voice data, etc.) transmitted from the local terminal 10 by the local user. In addition, the communication control unit 231 may receive the omnidirectional image from the distribution server 50 described with reference to FIG. The communication control unit 231 may receive sensing data from the sensor device 60 described with reference to FIG.
 表示制御部235は、表示部24による表示を制御する。例えば、表示制御部235は、全天球画像からセンサ部21により取得されたリモートユーザの顔向きに応じた領域を切出して表示画像を生成し、表示画像を表示部24に表示させてもよい。 The display control unit 235 controls display by the display unit 24. For example, the display control unit 235 may generate a display image by cutting out a region corresponding to the face orientation of the remote user acquired by the sensor unit 21 from the omnidirectional image, and display the display image on the display unit 24. .
 係る構成により、リモート端末20を装着したリモートユーザは、顔向きを変えることで、撮像装置30を視点とし、上下左右全方位にローカル環境1000を観察することが可能である。また、図1に示すように、ローカルユーザが表示画像に含まれるような向きに顔を向けることで、リモートユーザはローカルユーザと対面した感覚を得ることが可能となる。 With this configuration, a remote user wearing the remote terminal 20 can observe the local environment 1000 in all directions from the viewpoint of the imaging device 30 by changing the face orientation. Further, as shown in FIG. 1, the remote user can obtain a sense of facing the local user by facing the face in a direction that is included in the display image.
 なお、表示制御部235は、全天球画像から切り出された領域に対して、さらに画像処理を施して表示画像を生成してもよい。例えば、表示制御部235が施す画像処理は、全天球画像から切り出された領域からローカルユーザを検出してローカルユーザの領域を加工する加工処理や、当該ローカルユーザの領域を別のオブジェクト(例えばアバタ)に置き換える置き換え処理を含んでもよい。表示制御部235が施す加工処理は、例えばローカルユーザが装着したローカル端末10を除去して、ローカルユーザの表情を復元する処理を含んでもよい。また、表示制御部235が施す画像処理は、ローカルユーザにより発信されるメッセージを合成する処理を含んでもよい。 Note that the display control unit 235 may further perform image processing on the area cut out from the omnidirectional image to generate a display image. For example, the image processing performed by the display control unit 235 includes a processing process that detects a local user from an area cut out from the omnidirectional image and processes the local user area, or converts the local user area into another object (for example, A replacement process may be included. The processing performed by the display control unit 235 may include, for example, processing for removing the local terminal 10 worn by the local user and restoring the facial expression of the local user. In addition, the image processing performed by the display control unit 235 may include processing for synthesizing a message transmitted by the local user.
 図6に示す音響出力制御部237は、音響出力部25による音響出力を制御する。例えば、音響出力制御部237は、音響出力部25を制御して、ローカルユーザにより発信されるメッセージを音響出力させてもよい。また、音響出力制御部237は、表示部24に表示されたローカルユーザの位置からメッセージが聞こえるように、音響出力を制御してもよい。係る構成により、リモートユーザはより没入感を得ることが可能となる。 6 controls the sound output by the sound output unit 25. The sound output control unit 237 illustrated in FIG. For example, the sound output control unit 237 may control the sound output unit 25 to sound output a message transmitted by the local user. In addition, the sound output control unit 237 may control the sound output so that a message can be heard from the position of the local user displayed on the display unit 24. With this configuration, the remote user can obtain a more immersive feeling.
 表示部24は表示制御部235に制御されて、表示画像を表示するディスプレイである。表示部24は、例えばリモート端末20を装着するリモートユーザの両目に対してそれぞれ別の画像を提示することが可能であってもよい。係る構成により、リモートユーザはローカル環境1000を立体的に観察可能であり、より没入感を得ることが可能となる。 The display unit 24 is a display that displays a display image under the control of the display control unit 235. The display unit 24 may be able to present different images to both eyes of a remote user wearing the remote terminal 20, for example. With such a configuration, the remote user can observe the local environment 1000 in a three-dimensional manner, and can obtain a more immersive feeling.
 音響出力部25は音響出力制御部237に制御されて音響出力を行う。音響出力部25は、例えば複数のスピーカを含んでよく、音響出力部25は立体的に音響(立体音響)を出力することが可能であってもよい。 The sound output unit 25 performs sound output under the control of the sound output control unit 237. The sound output unit 25 may include, for example, a plurality of speakers, and the sound output unit 25 may be capable of outputting sound in a three-dimensional manner (stereo sound).
 記憶部26は、制御部23が上記各処理を実行するためのプログラムや、様々なデータを記憶する。 The storage unit 26 stores a program for the control unit 23 to execute the above processes and various data.
 以上、図6を参照してリモート端末20の構成例について説明したが、図6に示したのは一例であって、本技術は係る例に限定されない。例えば、図6に示す一部の機能がリモート端末20に接続された他の情報処理装置に備えられてもよい。例えば、図6に示す制御部23の一部またはすべての機能がリモート端末20に接続された他の情報処理装置に備えられ、当該他の情報処理装置により制御されて、リモート端末20が表示、または音響出力を行ってもよい。 The configuration example of the remote terminal 20 has been described above with reference to FIG. 6, but the example illustrated in FIG. 6 is an example, and the present technology is not limited to the example. For example, some functions shown in FIG. 6 may be provided in another information processing apparatus connected to the remote terminal 20. For example, a part or all of the functions of the control unit 23 shown in FIG. 6 are provided in another information processing apparatus connected to the remote terminal 20, and the remote terminal 20 displays and is controlled by the other information processing apparatus. Or sound output may be performed.
 <<3.動作>>
 以上、本実施形態に係るコミュニケーションシステム1、ローカル端末10、及びリモート端末20の構成例について説明した。続いて、本実施形態に係るコミュニケーションシステム1の動作例について図7を参照して説明する。図7は、本実施形態に係るコミュニケーションシステム1の動作例を示すフローチャート図である。
<< 3. Operation >>
The configuration examples of the communication system 1, the local terminal 10, and the remote terminal 20 according to the present embodiment have been described above. Next, an operation example of the communication system 1 according to the present embodiment will be described with reference to FIG. FIG. 7 is a flowchart showing an operation example of the communication system 1 according to the present embodiment.
 図7に示すように、まず撮像装置30により撮像が行われた(S102)後、撮像された画像に基づく合成処理等の画像処理が行われて全天球画像が取得される(S104)。続いて、撮像装置30から配信サーバ50へ全天球画像が送信される(S108)。さらに、配信サーバ50は撮像装置30から受信した全天球画像をリモート端末20へ配信(送信)する(S108)。 As shown in FIG. 7, first, imaging is performed by the imaging device 30 (S102), and then image processing such as synthesis processing based on the captured image is performed, and an omnidirectional image is acquired (S104). Subsequently, the omnidirectional image is transmitted from the imaging device 30 to the distribution server 50 (S108). Further, the distribution server 50 distributes (transmits) the omnidirectional image received from the imaging device 30 to the remote terminal 20 (S108).
 続いて、リモート端末20のセンサ部21がセンシングにより、例えばリモートユーザの顔向きの情報等を含むセンシングデータを取得する(S110)。さらに、リモート端末20の表示制御部は、全天球画像からリモートユーザの顔向きに応じた領域を切出して表示画像を生成し、表示部24に表示させる(S112)、 Subsequently, the sensor unit 21 of the remote terminal 20 acquires sensing data including, for example, information on the face direction of the remote user by sensing (S110). Further, the display control unit of the remote terminal 20 cuts out an area corresponding to the remote user's face direction from the omnidirectional image, generates a display image, and causes the display unit 24 to display the display image (S112).
 続いて、リモート端末20は、リモートユーザの識別情報、リモートユーザの状態を示す情報、リモートユーザの姿勢情報、リモートユーザにより発信されたメッセージ(テキストデータ、音声データ等)等を含むユーザ情報をローカル端末10へ送信する(S114)。 Subsequently, the remote terminal 20 locally stores user information including remote user identification information, information indicating the status of the remote user, attitude information of the remote user, messages (text data, voice data, etc.) transmitted by the remote user, and the like. It transmits to the terminal 10 (S114).
 続いて、ローカル端末10の相対位置取得部133は、ローカル端末10と実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報を取得する(S116)。さらに、ローカル端末10の表示制御部135は、ユーザ情報、及び相対位置情報に基づいて、実空間における撮像装置30の位置に応じた実空間における位置に、リモートユーザのアバタがローカルユーザにより視認されるように、当該アバタを表示させる(S118)。 Subsequently, the relative position acquisition unit 133 of the local terminal 10 acquires relative position information indicating the relative positional relationship between the local terminal 10 and the imaging device existing in the real space (S116). Further, the display control unit 135 of the local terminal 10 recognizes the remote user's avatar by the local user at a position in the real space according to the position of the imaging device 30 in the real space based on the user information and the relative position information. As shown, the avatar is displayed (S118).
 以上、本実施形態に係るコミュニケーションシステム1の動作例について説明したが、図7に示したのは一例であって、本技術は係る例に限定されない。例えば、図7に示したステップS102~S118の処理は繰り返されてもよい。また、ステップS108とステップS110の処理順が逆であってもよいし、ステップS114とステップS116の処理順が逆であってもよい。 The operation example of the communication system 1 according to the present embodiment has been described above. However, FIG. 7 shows an example, and the present technology is not limited to the example. For example, the processes in steps S102 to S118 shown in FIG. 7 may be repeated. Moreover, the processing order of step S108 and step S110 may be reversed, and the processing order of step S114 and step S116 may be reversed.
 <<4.変形例>>
 以上、本開示の一実施形態を説明した。以下では、本実施形態の幾つかの変形例を説明する。なお、以下に説明する各変形例は、単独で本実施形態に適用されてもよいし、組み合わせで本実施形態に適用されてもよい。また、各変形例は、本実施形態で説明した構成に代えて適用されてもよいし、本実施形態で説明した構成に対して追加的に適用されてもよい。
<< 4. Modification >>
The embodiment of the present disclosure has been described above. Below, some modifications of this embodiment are explained. Each modified example described below may be applied to the present embodiment alone, or may be applied to the present embodiment in combination. Each modification may be applied instead of the configuration described in the present embodiment, or may be additionally applied to the configuration described in the present embodiment.
  <4-1.変形例1>
 上記実施形態では、表示制御部135がリモートユーザのアバタと撮像装置30とが重畳されて視認されるようにアバタを表示させる例を説明したが、本技術は係る例に限定されない。例えば、表示制御部135は、撮像装置30の近傍にアバタを表示させてもよい。係る例について変形例1として図8を参照して説明する。
<4-1. Modification 1>
In the above-described embodiment, the example in which the display control unit 135 displays the avatar so that the avatar of the remote user and the imaging device 30 are superimposed and viewed has been described, but the present technology is not limited to the example. For example, the display control unit 135 may display an avatar in the vicinity of the imaging device 30. Such an example will be described as a first modification with reference to FIG.
 図8は表示制御部135が撮像装置30の近傍にアバタを表示させた場合のローカルユーザの視界V10を示す模式図である。図8に示す例では、撮像装置30の横にリモートユーザのアバタA1が視認されるように表示されている。なお、アバタが視認される位置は撮像装置30の横に限定されず、撮像装置30の上、下、奥、または手前であってもよい。 FIG. 8 is a schematic diagram showing the field of view V10 of the local user when the display control unit 135 displays an avatar in the vicinity of the imaging device 30. In the example illustrated in FIG. 8, the remote user's avatar A <b> 1 is displayed beside the imaging device 30. Note that the position where the avatar is visually recognized is not limited to the side of the imaging device 30 and may be above, below, inside, or in front of the imaging device 30.
  <4-2.変形例2>
 上記実施形態では、コミュニケーションシステム1が1の撮像装置30を備える例を説明したが、本技術は係る例に限定されない。例えば、複数の撮像装置30が備えられてもよい。係る例について変形例2として図9、図10を参照して説明する。
<4-2. Modification 2>
In the above-described embodiment, an example in which the communication system 1 includes one imaging device 30 has been described. However, the present technology is not limited to such an example. For example, a plurality of imaging devices 30 may be provided. Such an example will be described as a second modification with reference to FIGS.
 図9は本変形例に係るコミュニケーションシステム1-2の構成例を示すブロック図である。図9に示すコミュニケーションシステム1-2の構成は、ローカル環境1000に2つの撮像装置30A、及び撮像装置30Bが存在することを除いて図3を参照して説明したコミュニケーションシステム1の構成と同様であるため、適宜省略ながら説明する。 FIG. 9 is a block diagram showing a configuration example of the communication system 1-2 according to this modification. The configuration of the communication system 1-2 illustrated in FIG. 9 is the same as the configuration of the communication system 1 described with reference to FIG. 3 except that the two imaging devices 30A and 30B exist in the local environment 1000. Therefore, it will be described while omitting as appropriate.
 撮像装置30A、及び撮像装置30Bは、上述した撮像装置30と同様に、撮像により上下左右全方位の360度の全天球画像を取得することが可能な全天球カメラである。また、撮像装置30A、及び撮像装置30Bは通信網70を介して配信サーバ50へ全天球画像を送信する。 The imaging device 30A and the imaging device 30B are omnidirectional cameras that can acquire 360-degree omnidirectional images in all directions in the vertical and horizontal directions by imaging, as with the imaging device 30 described above. In addition, the imaging device 30 </ b> A and the imaging device 30 </ b> B transmit an omnidirectional image to the distribution server 50 via the communication network 70.
 本変形例に係るリモート端末20は、撮像装置30A、または撮像装置30Bのうち、リモートユーザにより選択された1の撮像装置の撮像に基づいた画像を表示してもよい。リモートユーザは、例えばセンサ装置60、リモート端末20のセンサ部21、または不図示の入力装置を介して当該選択に係る入力操作を行ってもよい。なお、配信サーバ50は、撮像装置30Aの撮像による全天球画像、及び撮像装置30Bの撮像による全天球画像の両方をリモート端末20へ送信してもよいし、リモートユーザにより選択された撮像装置の撮像による全天球画像のみをリモート端末20へ送信してもよい。 The remote terminal 20 according to the present modification may display an image based on the imaging of one imaging device selected by the remote user from the imaging device 30A or the imaging device 30B. The remote user may perform an input operation related to the selection via, for example, the sensor device 60, the sensor unit 21 of the remote terminal 20, or an input device (not shown). Note that the distribution server 50 may transmit both the omnidirectional image captured by the imaging device 30A and the omnidirectional image captured by the imaging device 30B to the remote terminal 20, or the imaging selected by the remote user. Only the omnidirectional image captured by the apparatus may be transmitted to the remote terminal 20.
 また、本変形例に係るローカル端末10は、図4を参照して説明した機能に加えて、以下の機能を有してよい。 Further, the local terminal 10 according to this modification may have the following functions in addition to the functions described with reference to FIG.
 例えば、本変形例に係るローカル端末10の表示制御部135は、複数の撮像装置のうち、リモートユーザにより選択された1の撮像装置の位置に応じた実空間における位置に、リモートユーザのアバタがローカルユーザにより視認されるように、表示を制御する。 For example, the display control unit 135 of the local terminal 10 according to the present modification has the remote user's avatar at a position in the real space corresponding to the position of one imaging device selected by the remote user among the plurality of imaging devices. The display is controlled so as to be visually recognized by the local user.
 また、本変形例に係る表示制御部135は、リモートユーザの選択に応じてアバタが視認される位置が切り替えられる場合に、アバタをフェードアウトさせた後に新たな位置にフェードインするようにアバタを表示させてもよい。 In addition, when the position where the avatar is visually recognized is switched according to the selection of the remote user, the display control unit 135 according to this modification displays the avatar so that the avatar fades out and then fades in to a new position. You may let them.
 また、本変形例に係る表示制御部135は、リモートユーザの選択に応じてアバタが視認される位置が切り替えられる場合に、直前に選択されていた撮像装置に応じた位置から現在選択されている撮像装置に応じた位置へ、アバタが移動するアニメーションを表示させてもよい。なお、係るアニメーションにおいて、アバタが3次元的に移動する映像効果が加えられてもよい。 Further, the display control unit 135 according to the present modification is currently selected from the position corresponding to the imaging device selected immediately before when the position where the avatar is visually recognized is switched according to the selection of the remote user. You may display the animation which an avatar moves to the position according to an imaging device. In this animation, a video effect in which the avatar moves three-dimensionally may be added.
 図10は本変形例に係る表示制御部135が表示させるアニメーションの例を示す説明図である。図10では、撮像装置30Aが選択されていた状態の後に、撮像装置30BがリモートユーザRU1により選択された状態を示している。リモート端末20は、撮像装置30Bの撮像により取得された全天球画像から、リモートユーザRU1の顔向きに応じた領域を切出して生成した表示画像G20を表示している。また、ローカルユーザLU1の視界V10において、撮像装置30Aと重畳されて視認されていたアバタA1が、撮像装置30Bと重畳されて視認される位置に移動している。 FIG. 10 is an explanatory diagram showing an example of an animation displayed by the display control unit 135 according to this modification. FIG. 10 shows a state in which the imaging device 30B is selected by the remote user RU1 after the imaging device 30A has been selected. The remote terminal 20 displays a display image G20 generated by cutting out an area corresponding to the face direction of the remote user RU1 from the omnidirectional image acquired by imaging of the imaging device 30B. Further, in the field of view V10 of the local user LU1, the avatar A1 that has been visually recognized while being superimposed with the imaging device 30A has moved to a position that is visually recognized with being superimposed with the imaging device 30B.
 係る構成により、リモートユーザは、より多様な視点からローカル環境1000を観察することが可能であると共に、ローカルユーザはリモートユーザの現在の視点に相当する撮像装置を容易に把握することが可能となる。 With such a configuration, the remote user can observe the local environment 1000 from various viewpoints, and the local user can easily grasp the imaging device corresponding to the current viewpoint of the remote user. .
  <4-3.変形例3>
 上記実施形態では、ローカルユーザとリモートユーザが1対1でコミュニケーションを行う例を説明したが、本技術は係る例に限定されない。例えば、1のローカルユーザに対して、複数のリモートユーザが存在してもよい。係る例について、変形例3として図11、図12を参照して説明する。
<4-3. Modification 3>
In the above-described embodiment, an example in which a local user and a remote user perform one-to-one communication has been described. However, the present technology is not limited to such an example. For example, a plurality of remote users may exist for one local user. Such an example will be described as a third modification with reference to FIGS. 11 and 12.
 図11は本変形例に係るコミュニケーションシステム1-3の構成例を示すブロック図である。図11に示すコミュニケーションシステム1-3の構成は、一部において図3を参照して説明したコミュニケーションシステム1の構成と同様であるため、適宜省略ながら説明する。 FIG. 11 is a block diagram showing a configuration example of the communication system 1-3 according to this modification. The configuration of the communication system 1-3 illustrated in FIG. 11 is partially the same as the configuration of the communication system 1 described with reference to FIG.
 図11に示すように、本変形例に係るコミュニケーションシステム1-3は、2つのリモート端末20A、及びリモート端末20Bを備え、2つのセンサ装置60A、及びセンサ装置60Bを備える点で図3に示したコミュニケーションシステム1と異なる。また、リモート端末20A、及びセンサ装置60Aはリモート環境2000Aに存在し、リモート端末20B、及びセンサ装置60Bはリモート環境2000Bに存在する。なお、リモート環境2000Aとリモート環境2000Bとは同一(単一)の環境であってもよく、係る場合にはセンサ装置60Aとセンサ装置60Bとは同一(単一)の装置であってもよい。 As shown in FIG. 11, the communication system 1-3 according to this modification includes two remote terminals 20A and 20B, and is shown in FIG. 3 in that it includes two sensor devices 60A and 60B. Different from communication system 1. Further, the remote terminal 20A and the sensor device 60A exist in the remote environment 2000A, and the remote terminal 20B and the sensor device 60B exist in the remote environment 2000B. The remote environment 2000A and the remote environment 2000B may be the same (single) environment. In such a case, the sensor device 60A and the sensor device 60B may be the same (single) device.
 また、リモート端末20Aは第1のリモートユーザにより用いられる情報処理装置であり、リモート端末20Bは第1のリモートユーザとは異なる第2のリモートユーザにより用いられる情報処理装置であってもよい。なお、本変形例に係るリモート端末20A、及びリモート端末20Bの構成は、図6を参照して説明したリモート端末20の構成と実質的に同一であるため、説明を省略する。 Further, the remote terminal 20A may be an information processing device used by a first remote user, and the remote terminal 20B may be an information processing device used by a second remote user different from the first remote user. Note that the configurations of the remote terminal 20A and the remote terminal 20B according to this modification are substantially the same as the configuration of the remote terminal 20 described with reference to FIG.
 また、本変形例に係る配信サーバ50は、リモート端末20A、及びリモート端末20Bの両方に、撮像装置30の撮像により得られた全天球画像を配信(送信)する。 Also, the distribution server 50 according to this modification distributes (transmits) the omnidirectional image obtained by the imaging of the imaging device 30 to both the remote terminal 20A and the remote terminal 20B.
 また、本変形例に係るローカル端末10は、図4を参照して説明した機能に加えて、以下の機能を有してよい。例えば、本変形例に係るローカル端末10の表示制御部135は、複数のリモートユーザを示すオブジェクト(例えばアバタ)を表示させてもよい。 Further, the local terminal 10 according to this modification may have the following functions in addition to the functions described with reference to FIG. For example, the display control unit 135 of the local terminal 10 according to this modification may display an object (for example, an avatar) indicating a plurality of remote users.
 図12は本変形例に係る表示制御部135が複数のリモートユーザのアバタを表示させるアニメーションの例を示す説明図である。図12に示す例では、表示制御部135は、リモート端末20Aを装着した第1のリモートユーザRU1のアバタA1と、リモート端末20Bを装着した第2のリモートユーザRU2のアバタA2の両方を表示させている。 FIG. 12 is an explanatory diagram showing an example of an animation in which the display control unit 135 according to this modification displays a plurality of remote user avatars. In the example shown in FIG. 12, the display control unit 135 displays both the avatar A1 of the first remote user RU1 wearing the remote terminal 20A and the avatar A2 of the second remote user RU2 wearing the remote terminal 20B. ing.
 なお、撮像装置30と重畳して視認されるように複数のアバタを表示してしまうと、複数のアバタが重なってしまうため、図12に示すように表示制御部135は、他のアバタと重ならない撮像装置30の近傍に視認されるように、アバタを表示させてもよい。なお、表示制御部135は、アバタを縮小表示することで、複数のアバタが重ならないように、かつ、撮像装置30と重畳して視認されるように、表示を制御してもよい。 Note that if a plurality of avatars are displayed so as to be viewed in a superimposed manner with the imaging device 30, the plurality of avatars are overlapped. Therefore, as shown in FIG. 12, the display control unit 135 overlaps with other avatars. An avatar may be displayed so as to be visually recognized in the vicinity of the imaging device 30 that should not be. Note that the display control unit 135 may control the display so that a plurality of avatars do not overlap with each other and can be visually recognized while being superimposed on the imaging device 30 by displaying the avatar in a reduced size.
 ローカル端末10の表示制御部135は、上述したようにリモートユーザに応じたアバタを表示させる機能を有し、図12に示す例ではアバタA1とアバタA2とは異なるアバタである。表示制御部135は、各リモートユーザの姿勢情報に応じて、各リモートユーザのアバタの姿勢を制御してよい。図12に示す例ではリモートユーザRU1の顔向きに応じた向きでアバタA1の顔A11の顔が表示され、リモートユーザRU2の顔向きに応じた向きでアバタA2の顔A21の顔が表示されている。 The display control unit 135 of the local terminal 10 has a function of displaying an avatar corresponding to the remote user as described above, and in the example illustrated in FIG. 12, the avatar A1 and the avatar A2 are different avatars. The display control unit 135 may control the avatar posture of each remote user in accordance with the posture information of each remote user. In the example shown in FIG. 12, the face of the avatar A1's face A11 is displayed in a direction corresponding to the face direction of the remote user RU1, and the face of the avatar A2's face A21 is displayed in a direction corresponding to the face direction of the remote user RU2. Yes.
 リモート環境2000Aに存在するリモートユーザRU1は、リモート端末20Aを装着している。また、リモート端末20Aには、撮像装置30の撮像により得られた全天球画像からリモートユーザRU1の顔向きに応じた領域を切出して生成した表示画像G21が表示される。 The remote user RU1 existing in the remote environment 2000A is wearing the remote terminal 20A. Further, on the remote terminal 20A, a display image G21 generated by cutting out an area corresponding to the face orientation of the remote user RU1 from the omnidirectional image obtained by imaging by the imaging device 30 is displayed.
 リモート環境2000Bに存在するリモートユーザRU2は、リモート端末20Bを装着している。また、リモート端末20Bには、撮像装置30の撮像により得られた全天球画像からリモートユーザRU2の顔向きに応じた領域を切出して生成した表示画像G22が表示される。 The remote user RU2 existing in the remote environment 2000B is wearing the remote terminal 20B. In addition, the remote terminal 20B displays a display image G22 generated by cutting out an area corresponding to the face direction of the remote user RU2 from the omnidirectional image obtained by imaging by the imaging device 30.
 なお、ローカル環境1000における各リモートユーザのアバタの位置に基づいて、各リモート端末20が、他のリモートユーザのアバタを表示してもよい。例えば、リモート端末20Aは、リモートユーザRU1が左側を向いた場合に、全天球画像から顔向きに応じた領域を切出した画像にリモートユーザRU2のアバタが合成された画像を表示画像として表示してもよい。 Note that, based on the position of each remote user's avatar in the local environment 1000, each remote terminal 20 may display the avatar of another remote user. For example, when the remote user RU1 turns to the left, the remote terminal 20A displays, as a display image, an image obtained by combining the avatar of the remote user RU2 with an image obtained by cutting out an area corresponding to the face direction from the omnidirectional image. May be.
 係る場合、リモート端末20Aは、リモートユーザRU2のアバタの近傍にメッセージを表示させたり、当該リモートユーザRU2のアバタの位置からリモートユーザRU2により発信されたメッセージが聞こえるように音響出力を行ったりしてもよい。 In such a case, the remote terminal 20A displays a message in the vicinity of the avatar of the remote user RU2, or outputs a sound so that a message transmitted by the remote user RU2 can be heard from the position of the avatar of the remote user RU2. Also good.
 なお、本変形例は、上述した変形例2と組み合わせられてもよい。例えば複数のリモートユーザが、それぞれ複数の撮像装置から1の撮像装置を選択し、ローカル端末10の表示制御部135は各リモートユーザが選択した撮像装置の位置に応じた位置にリモートユーザのアバタが視認されるように表示を制御してもよい。 Note that this modification may be combined with Modification 2 described above. For example, a plurality of remote users select one imaging device from a plurality of imaging devices, and the display control unit 135 of the local terminal 10 displays the remote user's avatar at a position corresponding to the position of the imaging device selected by each remote user. The display may be controlled so as to be visually recognized.
 また、各リモート端末20は、当該リモート端末20を装着したリモートユーザ以外の他のリモートユーザが選択した撮像装置の位置に応じた位置に、当該他のリモートユーザのアバタを表示させてもよい。係る場合、リモート端末20は、当該他のリモートユーザのアバタの近傍にメッセージを表示させたり、当該リモートユーザのアバタの位置からリモートユーザにより発信されたメッセージが聞こえるように音響出力を行ったりしてもよい。 Further, each remote terminal 20 may display the avatar of the other remote user at a position corresponding to the position of the imaging device selected by the remote user other than the remote user wearing the remote terminal 20. In such a case, the remote terminal 20 displays a message in the vicinity of the avatar of the other remote user, or outputs a sound so that a message transmitted by the remote user can be heard from the position of the avatar of the remote user. Also good.
  <4-4.変形例4>
 上記実施形態は、ローカル端末10が透過型の表示部を有する眼鏡型デバイスであり、リモート端末20が没入型HMDであり、撮像装置30が全天球カメラである例を説明したが、本技術に係る実施形態は係る例に限定されず、多様な装置構成により実現され得る。
<4-4. Modification 4>
In the above embodiment, an example has been described in which the local terminal 10 is a glasses-type device having a transmissive display unit, the remote terminal 20 is an immersive HMD, and the imaging device 30 is an omnidirectional camera. The embodiment according to the invention is not limited to the example, and can be realized by various apparatus configurations.
 例えば、ローカル端末10が有する表示部14は透過型でなくてもよい。係る場合、ローカル端末10が備えるセンサ部11に含まれ、ローカルユーザの視界(実空間)を撮影するカメラで取得された実空間画像に、アバタ(リモートユーザを示すオブジェクトの一例)が合成されて表示されてもよい。係る構成により、表示部14が透過型でない場合であっても、ローカル端末10の表示制御部135は、撮像装置30の位置に応じた実空間における位置にアバタがローカルユーザにより視認されるように、表示を制御することが可能である。 For example, the display unit 14 included in the local terminal 10 may not be a transmissive type. In such a case, an avatar (an example of an object indicating a remote user) is combined with a real space image that is included in the sensor unit 11 included in the local terminal 10 and acquired by a camera that captures the field of view (real space) of the local user. May be displayed. With such a configuration, even when the display unit 14 is not a transmissive type, the display control unit 135 of the local terminal 10 allows the local user to visually recognize the avatar at a position in the real space according to the position of the imaging device 30. It is possible to control the display.
 または、ローカル端末10は、スマートフォンやタブレット端末などであってもよい。係る場合であっても、センサ部11に含まれるカメラで取得された実空間画像、アバタが合成されて表示されてもよい。 Alternatively, the local terminal 10 may be a smartphone or a tablet terminal. Even in such a case, a real space image and an avatar acquired by a camera included in the sensor unit 11 may be combined and displayed.
 もしくは、ローカル端末10の表示部14はプロジェクタであってもよい。係る場合、ローカル端末10は、撮像装置30の位置に応じた実空間における位置にアバタの画像を投影してもよい。係る構成により、表示部14がプロジェクタである場合であっても、ローカル端末10の表示制御部135は、当該プロジェクタを制御することで、撮像装置30の位置に応じた実空間における位置にアバタがローカルユーザにより視認されるように、表示を制御することが可能である。 Alternatively, the display unit 14 of the local terminal 10 may be a projector. In such a case, the local terminal 10 may project an avatar image at a position in the real space according to the position of the imaging device 30. With such a configuration, even when the display unit 14 is a projector, the display control unit 135 of the local terminal 10 controls the projector so that the avatar is positioned at a position in the real space according to the position of the imaging device 30. The display can be controlled to be viewed by a local user.
 また、リモート端末20の表示部24は平面ディスプレイであってもよい。係る場合、リモートユーザの視界に相当する領域は、リモートコントローラ―等の入力装置を介した操作や、センサ部21により取得されるジェスチャ操作に応じて特定されてもよい。 Further, the display unit 24 of the remote terminal 20 may be a flat display. In such a case, the area corresponding to the field of view of the remote user may be specified according to an operation via an input device such as a remote controller or a gesture operation acquired by the sensor unit 21.
 また、リモート端末20は、スマートフォンやタブレット端末などであってもよい。係る場合であっても、リモートユーザの視界に相当する領域は、タッチ操作や、センサ部21により取得されるリモート端末20の姿勢に応じて特定されてもよい。 Further, the remote terminal 20 may be a smartphone or a tablet terminal. Even in such a case, the region corresponding to the visual field of the remote user may be specified according to the touch operation or the posture of the remote terminal 20 acquired by the sensor unit 21.
 また、撮像装置30は全天球カメラではなくてもよく、例えば撮像画角が180度のカメラであってもよい。そして、ローカル端末10の表示制御部135は、表示させるアバタの顔向きを撮像装置30の撮像画角に応じて限定してもよい。係る構成により、アバタの顔向きと、実際にリモートユーザが見る視界との不整合を抑制することが可能である。 Further, the imaging device 30 may not be a spherical camera, and may be a camera having an imaging angle of view of 180 degrees, for example. Then, the display control unit 135 of the local terminal 10 may limit the face direction of the avatar to be displayed according to the imaging angle of view of the imaging device 30. With such a configuration, it is possible to suppress inconsistency between the avatar's face direction and the field of view actually seen by the remote user.
 <<5.ハードウェア構成>>
 以上、本開示の実施形態を説明した。最後に、図13を参照して、本実施形態に係る情報処理装置のハードウェア構成について説明する。図13は、本実施形態に係る情報処理装置のハードウェア構成の一例を示すブロック図である。なお、図13に示す情報処理装置900は、例えば、ローカル端末10、リモート端末20、及び配信サーバ50を実現し得る。本実施形態に係るローカル端末10、リモート端末20、及び配信サーバ50による情報処理は、ソフトウェアと、以下に説明するハードウェアとの協働により実現される。
<< 5. Hardware configuration >>
The embodiment of the present disclosure has been described above. Finally, the hardware configuration of the information processing apparatus according to the present embodiment will be described with reference to FIG. FIG. 13 is a block diagram illustrating an example of a hardware configuration of the information processing apparatus according to the present embodiment. Note that the information processing apparatus 900 illustrated in FIG. 13 can implement the local terminal 10, the remote terminal 20, and the distribution server 50, for example. Information processing by the local terminal 10, the remote terminal 20, and the distribution server 50 according to the present embodiment is realized by cooperation of software and hardware described below.
 図13に示すように、情報処理装置900は、CPU(Central Processing Unit)901、ROM(Read Only Memory)902、RAM(Random Access Memory)903及びホストバス904aを備える。また、情報処理装置900は、ブリッジ904、外部バス904b、インタフェース905、入力装置906、出力装置907、ストレージ装置908、ドライブ909、接続ポート911、通信装置913、及びセンサ915を備える。情報処理装置900は、CPU901に代えて、又はこれとともに、DSP若しくはASIC等の処理回路を有してもよい。 As shown in FIG. 13, the information processing apparatus 900 includes a CPU (Central Processing Unit) 901, a ROM (Read Only Memory) 902, a RAM (Random Access Memory) 903, and a host bus 904a. The information processing apparatus 900 includes a bridge 904, an external bus 904b, an interface 905, an input device 906, an output device 907, a storage device 908, a drive 909, a connection port 911, a communication device 913, and a sensor 915. The information processing apparatus 900 may include a processing circuit such as a DSP or an ASIC in place of or in addition to the CPU 901.
 CPU901は、演算処理装置および制御装置として機能し、各種プログラムに従って情報処理装置900内の動作全般を制御する。また、CPU901は、マイクロプロセッサであってもよい。ROM902は、CPU901が使用するプログラムや演算パラメータ等を記憶する。RAM903は、CPU901の実行において使用するプログラムや、その実行において適宜変化するパラメータ等を一時記憶する。CPU901は、例えば、制御部13、制御部23を形成し得る。 The CPU 901 functions as an arithmetic processing unit and a control unit, and controls the overall operation in the information processing apparatus 900 according to various programs. Further, the CPU 901 may be a microprocessor. The ROM 902 stores programs used by the CPU 901, calculation parameters, and the like. The RAM 903 temporarily stores programs used in the execution of the CPU 901, parameters that change as appropriate during the execution, and the like. For example, the CPU 901 can form the control unit 13 and the control unit 23.
 CPU901、ROM902及びRAM903は、CPUバスなどを含むホストバス904aにより相互に接続されている。ホストバス904aは、ブリッジ904を介して、PCI(Peripheral Component Interconnect/Interface)バスなどの外部バス904bに接続されている。なお、必ずしもホストバス904a、ブリッジ904および外部バス904bを分離構成する必要はなく、1つのバスにこれらの機能を実装してもよい。 The CPU 901, ROM 902, and RAM 903 are connected to each other by a host bus 904a including a CPU bus. The host bus 904 a is connected to an external bus 904 b such as a PCI (Peripheral Component Interconnect / Interface) bus via a bridge 904. Note that the host bus 904a, the bridge 904, and the external bus 904b do not necessarily have to be configured separately, and these functions may be mounted on one bus.
 入力装置906は、例えば、マウス、キーボード、タッチパネル、ボタン、マイクロフォン、スイッチ及びレバー等、ユーザによって情報が入力される装置によって実現される。また、入力装置906は、例えば、赤外線やその他の電波を利用したリモートコントロール装置であってもよいし、情報処理装置900の操作に対応した携帯電話やPDA等の外部接続機器であってもよい。さらに、入力装置906は、例えば、上記の入力手段を用いてユーザにより入力された情報に基づいて入力信号を生成し、CPU901に出力する入力制御回路などを含んでいてもよい。情報処理装置900のユーザは、この入力装置906を操作することにより、情報処理装置900に対して各種のデータを入力したり処理動作を指示したりすることができる。 The input device 906 is realized by a device in which information is input by the user, such as a mouse, a keyboard, a touch panel, a button, a microphone, a switch, and a lever. The input device 906 may be, for example, a remote control device using infrared rays or other radio waves, or may be an external connection device such as a mobile phone or a PDA that supports the operation of the information processing device 900. . Furthermore, the input device 906 may include, for example, an input control circuit that generates an input signal based on information input by the user using the above-described input means and outputs the input signal to the CPU 901. A user of the information processing apparatus 900 can input various data and instruct a processing operation to the information processing apparatus 900 by operating the input device 906.
 出力装置907は、取得した情報をユーザに対して視覚的又は聴覚的に通知することが可能な装置で形成される。このような装置として、CRTディスプレイ装置、液晶ディスプレイ装置、プラズマディスプレイ装置、ELディスプレイ装置及びランプ等の表示装置や、スピーカ及びヘッドホン等の音声出力装置や、プリンタ装置等がある。出力装置907は、例えば、情報処理装置900が行った各種処理により得られた結果を出力する。具体的には、表示装置は、情報処理装置900が行った各種処理により得られた結果を、テキスト、イメージ、表、グラフ等、様々な形式で視覚的に表示する。他方、音声出力装置は、再生された音声データや音響データ等からなるオーディオ信号をアナログ信号に変換して聴覚的に出力する。出力装置907は、例えば表示部14、音響出力部15、表示部24、音響出力部25を形成し得る。 The output device 907 is formed of a device that can notify the user of the acquired information visually or audibly. Examples of such devices include CRT display devices, liquid crystal display devices, plasma display devices, EL display devices, display devices such as lamps, audio output devices such as speakers and headphones, printer devices, and the like. For example, the output device 907 outputs results obtained by various processes performed by the information processing device 900. Specifically, the display device visually displays results obtained by various processes performed by the information processing device 900 in various formats such as text, images, tables, and graphs. On the other hand, the audio output device converts an audio signal composed of reproduced audio data, acoustic data, and the like into an analog signal and outputs it aurally. The output device 907 can form, for example, the display unit 14, the sound output unit 15, the display unit 24, and the sound output unit 25.
 ストレージ装置908は、情報処理装置900の記憶部の一例として形成されたデータ格納用の装置である。ストレージ装置908は、例えば、HDD等の磁気記憶部デバイス、半導体記憶デバイス、光記憶デバイス又は光磁気記憶デバイス等により実現される。ストレージ装置908は、記憶媒体、記憶媒体にデータを記録する記録装置、記憶媒体からデータを読み出す読出し装置および記憶媒体に記録されたデータを削除する削除装置などを含んでもよい。このストレージ装置908は、CPU901が実行するプログラムや各種データ及び外部から取得した各種のデータ等を格納する。上記ストレージ装置908は、例えば、記憶部16、記憶部26を形成し得る。 The storage device 908 is a data storage device formed as an example of a storage unit of the information processing device 900. The storage apparatus 908 is realized by, for example, a magnetic storage device such as an HDD, a semiconductor storage device, an optical storage device, a magneto-optical storage device, or the like. The storage device 908 may include a storage medium, a recording device that records data on the storage medium, a reading device that reads data from the storage medium, a deletion device that deletes data recorded on the storage medium, and the like. The storage device 908 stores programs executed by the CPU 901, various data, various data acquired from the outside, and the like. The storage device 908 can form the storage unit 16 and the storage unit 26, for example.
 ドライブ909は、記憶媒体用リーダライタであり、情報処理装置900に内蔵、あるいは外付けされる。ドライブ909は、装着されている磁気ディスク、光ディスク、光磁気ディスク、または半導体メモリ等のリムーバブル記憶媒体に記録されている情報を読み出して、RAM903に出力する。また、ドライブ909は、リムーバブル記憶媒体に情報を書き込むこともできる。 The drive 909 is a storage medium reader / writer, and is built in or externally attached to the information processing apparatus 900. The drive 909 reads information recorded on a removable storage medium such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory, and outputs the information to the RAM 903. The drive 909 can also write information to a removable storage medium.
 接続ポート911は、外部機器と接続されるインタフェースであって、例えばUSB(Universal Serial Bus)などによりデータ伝送可能な外部機器との接続口である。 The connection port 911 is an interface connected to an external device, and is a connection port with an external device capable of transmitting data by USB (Universal Serial Bus), for example.
 通信装置913は、例えば、ネットワーク920に接続するための通信デバイス等で形成された通信インタフェースである。通信装置913は、例えば、有線若しくは無線LAN(Local Area Network)、LTE(Long Term Evolution)、Bluetooth(登録商標)又はWUSB(Wireless USB)用の通信カード等である。また、通信装置913は、光通信用のルータ、ADSL(Asymmetric Digital Subscriber Line)用のルータ又は各種通信用のモデム等であってもよい。この通信装置913は、例えば、インターネットや他の通信機器との間で、例えばTCP/IP等の所定のプロトコルに則して信号等を送受信することができる。通信装置913は、例えば、通信部12、通信部22を形成し得る。 The communication device 913 is a communication interface formed by a communication device or the like for connecting to the network 920, for example. The communication device 913 is, for example, a communication card for wired or wireless LAN (Local Area Network), LTE (Long Term Evolution), Bluetooth (registered trademark), or WUSB (Wireless USB). The communication device 913 may be a router for optical communication, a router for ADSL (Asymmetric Digital Subscriber Line), a modem for various communication, or the like. The communication device 913 can transmit and receive signals and the like according to a predetermined protocol such as TCP / IP, for example, with the Internet and other communication devices. The communication device 913 can form the communication unit 12 and the communication unit 22, for example.
 センサ915は、例えば、加速度センサ、ジャイロセンサ、地磁気センサ、光センサ、音センサ、測距センサ、力センサ等の各種のセンサである。センサ915は、情報処理装置900の姿勢、移動速度等、情報処理装置900自身の状態に関する情報や、情報処理装置900の周辺の明るさや騒音等、情報処理装置900の周辺環境に関する情報を取得する。また、センサ915は、GPS信号を受信して装置の緯度、経度及び高度を測定するGPSセンサを含んでもよい。センサ915は、例えば、センサ部11、センサ部21を形成し得る。 The sensor 915 is various sensors such as an acceleration sensor, a gyro sensor, a geomagnetic sensor, an optical sensor, a sound sensor, a distance measuring sensor, and a force sensor. The sensor 915 acquires information on the state of the information processing apparatus 900 itself, such as the posture and movement speed of the information processing apparatus 900, and information on the surrounding environment of the information processing apparatus 900, such as brightness and noise around the information processing apparatus 900. . Sensor 915 may also include a GPS sensor that receives GPS signals and measures the latitude, longitude, and altitude of the device. The sensor 915 can form the sensor part 11 and the sensor part 21, for example.
 なお、ネットワーク920は、ネットワーク920に接続されている装置から送信される情報の有線、または無線の伝送路である。例えば、ネットワーク920は、インターネット、電話回線網、衛星通信網などの公衆回線網や、Ethernet(登録商標)を含む各種のLAN(Local Area Network)、WAN(Wide Area Network)などを含んでもよい。また、ネットワーク920は、IP-VPN(Internet Protocol-Virtual Private Network)などの専用回線網を含んでもよい。 Note that the network 920 is a wired or wireless transmission path for information transmitted from a device connected to the network 920. For example, the network 920 may include a public line network such as the Internet, a telephone line network, and a satellite communication network, various LANs including the Ethernet (registered trademark), a wide area network (WAN), and the like. Further, the network 920 may include a dedicated line network such as an IP-VPN (Internet Protocol-Virtual Private Network).
 以上、本実施形態に係る情報処理装置900の機能を実現可能なハードウェア構成の一例を示した。上記の各構成要素は、汎用的な部材を用いて実現されていてもよいし、各構成要素の機能に特化したハードウェアにより実現されていてもよい。従って、本実施形態を実施する時々の技術レベルに応じて、適宜、利用するハードウェア構成を変更することが可能である。 Heretofore, an example of the hardware configuration capable of realizing the functions of the information processing apparatus 900 according to the present embodiment has been shown. Each of the above components may be realized using a general-purpose member, or may be realized by hardware specialized for the function of each component. Therefore, it is possible to change the hardware configuration to be used as appropriate according to the technical level at the time of carrying out this embodiment.
 なお、上述のような本実施形態に係る情報処理装置900の各機能を実現するためのコンピュータプログラムを作製し、PC等に実装することが可能である。また、このようなコンピュータプログラムが格納された、コンピュータで読み取り可能な記録媒体も提供することができる。記録媒体は、例えば、磁気ディスク、光ディスク、光磁気ディスク、フラッシュメモリ等である。また、上記のコンピュータプログラムは、記録媒体を用いずに、例えばネットワークを介して配信されてもよい。 It should be noted that a computer program for realizing each function of the information processing apparatus 900 according to the present embodiment as described above can be produced and mounted on a PC or the like. In addition, a computer-readable recording medium storing such a computer program can be provided. The recording medium is, for example, a magnetic disk, an optical disk, a magneto-optical disk, a flash memory, or the like. Further, the above computer program may be distributed via a network, for example, without using a recording medium.
 <<6.むすび>>
 以上説明したように、本開示の実施形態によれば、リモートコミュニケーションにおいて、より円滑なコミュニケーションを実現することが可能である。
<< 6. Conclusion >>
As described above, according to the embodiment of the present disclosure, it is possible to realize smoother communication in remote communication.
 以上、添付図面を参照しながら本開示の好適な実施形態について詳細に説明したが、本開示の技術的範囲はかかる例に限定されない。本開示の技術分野における通常の知識を有する者であれば、請求の範囲に記載された技術的思想の範疇内において、各種の変更例または修正例に想到し得ることは明らかであり、これらについても、当然に本開示の技術的範囲に属するものと了解される。 The preferred embodiments of the present disclosure have been described in detail above with reference to the accompanying drawings, but the technical scope of the present disclosure is not limited to such examples. It is obvious that a person having ordinary knowledge in the technical field of the present disclosure can come up with various changes or modifications within the scope of the technical idea described in the claims. Of course, it is understood that it belongs to the technical scope of the present disclosure.
 上記実施形態では、ローカル端末10がリモートユーザを示すオブジェクトとして、仮想的なオブジェクトであるアバタを表示させる例を説明したが、本技術は係る例に限定されない。例えば、ローカル端末10は、リモートユーザを示すオブジェクトとして、センサ装置60が含むカメラによりリモートユーザを撮影して得られた画像を表示させてもよい。 In the above embodiment, the example in which the local terminal 10 displays the avatar that is a virtual object as the object indicating the remote user has been described, but the present technology is not limited to the example. For example, the local terminal 10 may display an image obtained by photographing the remote user with a camera included in the sensor device 60 as an object indicating the remote user.
 また、本明細書に記載された効果は、あくまで説明的または例示的なものであって限定的ではない。つまり、本開示に係る技術は、上記の効果とともに、または上記の効果に代えて、本明細書の記載から当業者には明らかな他の効果を奏しうる。 In addition, the effects described in this specification are merely illustrative or illustrative, and are not limited. That is, the technology according to the present disclosure can exhibit other effects that are apparent to those skilled in the art from the description of the present specification in addition to or instead of the above effects.
 なお、以下のような構成も本開示の技術的範囲に属する。
(1)
 実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、表示を制御する表示制御部を備える、情報処理装置。
(2)
 前記表示制御部は、前記撮像装置と前記オブジェクトとが重畳されて視認されるように前記オブジェクトを表示させる、前記(1)に記載の情報処理装置。
(3)
 前記表示制御部は、前記撮像装置の撮像部と前記オブジェクトに含まれる目とが重畳されて視認されるように前記オブジェクトを表示させる、前記(2)に記載の情報処理装置。
(4)
 前記表示制御部は、前記撮像装置の近傍に前記オブジェクトが視認されるように前記オブジェクトを表示させる、前記(1)に記載の情報処理装置。
(5)
 前記表示制御部は、前記リモートユーザの姿勢情報に基づいて、前記オブジェクトの姿勢を制御する、前記(1)~(4)のいずれか一項に記載の情報処理装置。
(6)
 前記表示制御部は、前記姿勢情報に含まれる前記リモートユーザの顔向きに関する情報に応じて、前記オブジェクトに含まれる顔の向きを制御する、前記(5)に記載の情報処理装置。
(7)
 前記表示制御部は、前記リモートユーザに応じた前記オブジェクトを表示させる、前記(1)~(6)のいずれか一項に記載の情報処理装置。
(8)
 前記表示制御部は、前記リモートユーザにより発信されるメッセージを前記オブジェクトの近傍に表示させる、前記(1)~(7)のいずれか一項に記載の情報処理装置。
(9)
 前記表示制御部は、前記リモートユーザの状態に応じて表示を制御する、前記(1)~(8)のいずれか一項に記載の情報処理装置。
(10)
 前記表示制御部は、前記リモートユーザの状態を示すインジケータを前記オブジェクトの近傍に表示させる、前記(9)に記載の情報処理装置。
(11)
 前記表示制御部は、前記リモートユーザの状態に応じて前記オブジェクトの表示に関するパラメータを制御する、前記(10)に記載の情報処理装置。
(12)
 前記表示制御部は、複数の撮像装置の中から選択された1の撮像装置の位置に応じた前記実空間における位置に前記オブジェクトが視認されるように、表示を制御する、前記(1)~(11)のいずれか一項に記載の情報処理装置。
(13)
 前記表示制御部は、選択に応じて前記オブジェクトが視認される位置が切り替えられる場合に、直前に選択されていた撮像装置に応じた位置から現在選択されている前記撮像装置に応じた位置へ、前記オブジェクトが移動するアニメーションを表示させる、前記(12)に記載の情報処理装置。
(14)
 前記表示制御部は、複数のリモートユーザを示すオブジェクトを表示させる、前記(1)~(13)のいずれか一項に記載の情報処理装置。
(15)
 前記リモートユーザにより発信されるメッセージを音響出力させる音響出力制御部をさらに備え、前記音響出力制御部は、前記オブジェクトが視認される前記実空間における位置から前記メッセージが聞こえるように音響出力を制御する、前記(1)~(14)のいずれか一項に記載の情報処理装置。
(16)
 前記撮像装置は全天球カメラである、前記(1)~(15)のいずれか一項に記載の情報処理装置。
(17)
 前記表示制御部は、透過型の表示部の表示を制御する、前記(1)~(16)のいずれか一項に記載の情報処理装置。
(18)
 前記表示制御部は、プロジェクタを制御する、前記(1)~(16)のいずれか一項に記載の情報処理装置。
(19)
 実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、プロセッサが表示を制御すること、を含む情報処理方法。
(20)
 コンピュータに、
 実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、表示を制御する機能を実現させるための、プログラム。
The following configurations also belong to the technical scope of the present disclosure.
(1)
Based on relative position information indicating a relative positional relationship with the imaging device existing in the real space, an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device. An information processing apparatus comprising a display control unit that controls display.
(2)
The information processing apparatus according to (1), wherein the display control unit displays the object so that the imaging apparatus and the object are superimposed and viewed.
(3)
The information processing apparatus according to (2), wherein the display control unit displays the object so that an imaging unit of the imaging apparatus and eyes included in the object are superimposed and viewed.
(4)
The information processing apparatus according to (1), wherein the display control unit displays the object so that the object is visually recognized in the vicinity of the imaging apparatus.
(5)
The information processing apparatus according to any one of (1) to (4), wherein the display control unit controls the posture of the object based on posture information of the remote user.
(6)
The information processing apparatus according to (5), wherein the display control unit controls a face orientation included in the object according to information related to the face orientation of the remote user included in the posture information.
(7)
The information processing apparatus according to any one of (1) to (6), wherein the display control unit displays the object according to the remote user.
(8)
The information processing apparatus according to any one of (1) to (7), wherein the display control unit displays a message transmitted by the remote user in the vicinity of the object.
(9)
The information processing apparatus according to any one of (1) to (8), wherein the display control unit controls display according to a state of the remote user.
(10)
The information processing apparatus according to (9), wherein the display control unit displays an indicator indicating the state of the remote user in the vicinity of the object.
(11)
The information processing apparatus according to (10), wherein the display control unit controls a parameter related to display of the object according to a state of the remote user.
(12)
The display control unit controls display so that the object is visually recognized at a position in the real space according to the position of one imaging device selected from a plurality of imaging devices. The information processing apparatus according to any one of (11).
(13)
When the position where the object is visually recognized is switched according to the selection, the display control unit changes from the position corresponding to the imaging device selected immediately before to the position corresponding to the currently selected imaging device. The information processing apparatus according to (12), wherein an animation of moving the object is displayed.
(14)
The information processing apparatus according to any one of (1) to (13), wherein the display control unit displays an object indicating a plurality of remote users.
(15)
An acoustic output control unit that acoustically outputs a message transmitted by the remote user, and the acoustic output control unit controls the acoustic output so that the message can be heard from a position in the real space where the object is visually recognized. The information processing apparatus according to any one of (1) to (14).
(16)
The information processing apparatus according to any one of (1) to (15), wherein the imaging apparatus is an omnidirectional camera.
(17)
The information processing apparatus according to any one of (1) to (16), wherein the display control unit controls display of a transmissive display unit.
(18)
The information processing apparatus according to any one of (1) to (16), wherein the display control unit controls a projector.
(19)
Based on relative position information indicating a relative positional relationship with the imaging device existing in the real space, an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device. And an information processing method including: a processor controlling display.
(20)
On the computer,
Based on relative position information indicating a relative positional relationship with the imaging device existing in the real space, an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device. Program for realizing the function to control the display.
 1 コミュニケーションシステム
 10 ローカル端末
 11 センサ部
 12 通信部
 13 制御部
 14 表示部
 15 音響出力部
 16 記憶部
 20 リモート端末
 21 センサ部
 22 通信部
 23 制御部
 24 表示部
 25 音響出力部
 26 記憶部
 30 撮像装置
 40、60 センサ装置
 50 配信サーバ
 60 センサ装置
 70 通信網
 131 通信制御部
 133 相対位置取得部
 135 表示制御部
 137 音響出力制御部
 231 通信制御部
 235 表示制御部
 237 音響出力制御部
 1000 ローカル環境
 2000 リモート環境
DESCRIPTION OF SYMBOLS 1 Communication system 10 Local terminal 11 Sensor part 12 Communication part 13 Control part 14 Display part 15 Sound output part 16 Storage part 20 Remote terminal 21 Sensor part 22 Communication part 23 Control part 24 Display part 25 Sound output part 26 Storage part 30 Imaging device 40, 60 Sensor device 50 Distribution server 60 Sensor device 70 Communication network 131 Communication control unit 133 Relative position acquisition unit 135 Display control unit 137 Sound output control unit 231 Communication control unit 235 Display control unit 237 Sound output control unit 1000 Local environment 2000 Remote environment

Claims (20)

  1.  実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、表示を制御する表示制御部を備える、情報処理装置。 Based on relative position information indicating a relative positional relationship with the imaging device existing in the real space, an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device. An information processing apparatus comprising a display control unit that controls display.
  2.  前記表示制御部は、前記撮像装置と前記オブジェクトとが重畳されて視認されるように前記オブジェクトを表示させる、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit displays the object so that the imaging apparatus and the object are superimposed and viewed.
  3.  前記表示制御部は、前記撮像装置の撮像部と前記オブジェクトに含まれる目とが重畳されて視認されるように前記オブジェクトを表示させる、請求項2に記載の情報処理装置。 The information processing apparatus according to claim 2, wherein the display control unit displays the object so that the imaging unit of the imaging apparatus and the eyes included in the object are superimposed and viewed.
  4.  前記表示制御部は、前記撮像装置の近傍に前記オブジェクトが視認されるように前記オブジェクトを表示させる、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit displays the object so that the object is visually recognized in the vicinity of the imaging apparatus.
  5.  前記表示制御部は、前記リモートユーザの姿勢情報に基づいて、前記オブジェクトの姿勢を制御する、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit controls the posture of the object based on posture information of the remote user.
  6.  前記表示制御部は、前記姿勢情報に含まれる前記リモートユーザの顔向きに関する情報に応じて、前記オブジェクトに含まれる顔の向きを制御する、請求項5に記載の情報処理装置。 The information processing apparatus according to claim 5, wherein the display control unit controls a face orientation included in the object in accordance with information related to the face orientation of the remote user included in the posture information.
  7.  前記表示制御部は、前記リモートユーザに応じた前記オブジェクトを表示させる、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit displays the object corresponding to the remote user.
  8.  前記表示制御部は、前記リモートユーザにより発信されるメッセージを前記オブジェクトの近傍に表示させる、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit displays a message transmitted by the remote user in the vicinity of the object.
  9.  前記表示制御部は、前記リモートユーザの状態に応じて表示を制御する、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit controls display according to a state of the remote user.
  10.  前記表示制御部は、前記リモートユーザの状態を示すインジケータを前記オブジェクトの近傍に表示させる、請求項9に記載の情報処理装置。 The information processing apparatus according to claim 9, wherein the display control unit displays an indicator indicating a state of the remote user in the vicinity of the object.
  11.  前記表示制御部は、前記リモートユーザの状態に応じて前記オブジェクトの表示に関するパラメータを制御する、請求項10に記載の情報処理装置。 The information processing apparatus according to claim 10, wherein the display control unit controls a parameter related to display of the object according to a state of the remote user.
  12.  前記表示制御部は、複数の撮像装置の中から選択された1の撮像装置の位置に応じた前記実空間における位置に前記オブジェクトが視認されるように、表示を制御する、請求項1に記載の情報処理装置。 2. The display control unit according to claim 1, wherein the display control unit controls display so that the object is visually recognized at a position in the real space according to a position of one imaging apparatus selected from a plurality of imaging apparatuses. Information processing device.
  13.  前記表示制御部は、選択に応じて前記オブジェクトが視認される位置が切り替えられる場合に、直前に選択されていた撮像装置に応じた位置から現在選択されている前記撮像装置に応じた位置へ、前記オブジェクトが移動するアニメーションを表示させる、請求項12に記載の情報処理装置。 When the position where the object is visually recognized is switched according to the selection, the display control unit changes from the position corresponding to the imaging device selected immediately before to the position corresponding to the currently selected imaging device. The information processing apparatus according to claim 12, wherein an animation of moving the object is displayed.
  14.  前記表示制御部は、複数のリモートユーザを示すオブジェクトを表示させる、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit displays an object indicating a plurality of remote users.
  15.  前記リモートユーザにより発信されるメッセージを音響出力させる音響出力制御部をさらに備え、前記音響出力制御部は、前記オブジェクトが視認される前記実空間における位置から前記メッセージが聞こえるように音響出力を制御する、請求項1に記載の情報処理装置。 An acoustic output control unit that acoustically outputs a message transmitted by the remote user, and the acoustic output control unit controls the acoustic output so that the message can be heard from a position in the real space where the object is visually recognized. The information processing apparatus according to claim 1.
  16.  前記撮像装置は全天球カメラである、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the imaging apparatus is an omnidirectional camera.
  17.  前記表示制御部は、透過型の表示部の表示を制御する、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit controls display of a transmissive display unit.
  18.  前記表示制御部は、プロジェクタを制御する、請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the display control unit controls a projector.
  19.  実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、プロセッサが表示を制御すること、を含む情報処理方法。 Based on relative position information indicating a relative positional relationship with the imaging device existing in the real space, an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device. And an information processing method including: a processor controlling display.
  20.  コンピュータに、
     実空間に存在する撮像装置との相対的な位置関係を示す相対位置情報に基づいて、前記撮像装置の位置に応じた前記実空間における位置にリモートユーザを示すオブジェクトがローカルユーザにより視認されるように、表示を制御する機能を実現させるための、プログラム。
    On the computer,
    Based on relative position information indicating a relative positional relationship with the imaging device existing in the real space, an object indicating the remote user is visually recognized by the local user at a position in the real space corresponding to the position of the imaging device. Program for realizing the function to control the display.
PCT/JP2018/010433 2017-05-24 2018-03-16 Information processing device, information processing method, and program WO2018216327A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2017102488 2017-05-24
JP2017-102488 2017-05-24

Publications (1)

Publication Number Publication Date
WO2018216327A1 true WO2018216327A1 (en) 2018-11-29

Family

ID=64396463

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/010433 WO2018216327A1 (en) 2017-05-24 2018-03-16 Information processing device, information processing method, and program

Country Status (1)

Country Link
WO (1) WO2018216327A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000244886A (en) * 1999-01-20 2000-09-08 Canon Inc Computer conference system, computer processor, method for computer conference, processing method of computer processor, video conferencing system, method for video conferencing, and headphones
JP2010092304A (en) * 2008-10-08 2010-04-22 Sony Computer Entertainment Inc Information processing apparatus and method
JP2010219989A (en) * 2009-03-18 2010-09-30 Oki Electric Ind Co Ltd Communication support system, display control apparatus, and display control method
JP2015172883A (en) * 2014-03-12 2015-10-01 株式会社コナミデジタルエンタテインメント Terminal equipment, information communication method, and information communication program
WO2015185793A1 (en) * 2014-06-02 2015-12-10 Nokia Technologies Oy Method and apparatus for eye-line augmentation during a video conference

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000244886A (en) * 1999-01-20 2000-09-08 Canon Inc Computer conference system, computer processor, method for computer conference, processing method of computer processor, video conferencing system, method for video conferencing, and headphones
JP2010092304A (en) * 2008-10-08 2010-04-22 Sony Computer Entertainment Inc Information processing apparatus and method
JP2010219989A (en) * 2009-03-18 2010-09-30 Oki Electric Ind Co Ltd Communication support system, display control apparatus, and display control method
JP2015172883A (en) * 2014-03-12 2015-10-01 株式会社コナミデジタルエンタテインメント Terminal equipment, information communication method, and information communication program
WO2015185793A1 (en) * 2014-06-02 2015-12-10 Nokia Technologies Oy Method and apparatus for eye-line augmentation during a video conference

Similar Documents

Publication Publication Date Title
WO2016203792A1 (en) Information processing device, information processing method, and program
JP6822410B2 (en) Information processing system and information processing method
JP7005753B2 (en) Privacy screen
US11061466B2 (en) Apparatus and associated methods for presenting sensory scenes
CA2981208A1 (en) Method and system for implementing a multi-user virtual environment
US11361497B2 (en) Information processing device and information processing method
US10515481B2 (en) Method for assisting movement in virtual space and system executing the method
CN111386517A (en) Apparatus, and associated method, for communication between users experiencing virtual reality
US11151804B2 (en) Information processing device, information processing method, and program
KR20200096901A (en) Method and apparatus for navigating virtual content displayed by virtual reality (VR) device
CN109791436B (en) Apparatus and method for providing virtual scene
JPWO2017191700A1 (en) Image generating apparatus, image generating method, and program
JP2018094086A (en) Information processing device and image formation method
JP6563592B2 (en) Display control apparatus, display control method, and program
WO2018216327A1 (en) Information processing device, information processing method, and program
JP6159455B1 (en) Method, program, and recording medium for providing virtual space
WO2017203818A1 (en) Information processing device, information processing method, and program
US10940387B2 (en) Synchronized augmented reality gameplay across multiple gaming environments
WO2024004398A1 (en) Information processing device, program, and information processing system
WO2023087005A1 (en) Systems, methods, and media for controlling shared extended reality presentations
AU2022386387A1 (en) Systems, methods, and media for controlling shared extended reality presentations
CN117666769A (en) Virtual scene interaction method and device, storage medium and equipment
JP2018033938A (en) Method, program and recording medium for providing virtual space

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18804986

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18804986

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP