WO2022124047A1 - Information processing device, information processing system, and information processing method - Google Patents

Information processing device, information processing system, and information processing method Download PDF

Info

Publication number
WO2022124047A1
WO2022124047A1 PCT/JP2021/042524 JP2021042524W WO2022124047A1 WO 2022124047 A1 WO2022124047 A1 WO 2022124047A1 JP 2021042524 W JP2021042524 W JP 2021042524W WO 2022124047 A1 WO2022124047 A1 WO 2022124047A1
Authority
WO
WIPO (PCT)
Prior art keywords
viewpoint
information
line
sight
image
Prior art date
Application number
PCT/JP2021/042524
Other languages
French (fr)
Japanese (ja)
Inventor
健一郎 細川
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2022124047A1 publication Critical patent/WO2022124047A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]

Definitions

  • This disclosure relates to an information processing device, an information processing system, and an information processing method.
  • a technique of presenting an image rendered by a rendering device such as a game image or a 3D image to a user via a network.
  • a rendering device such as a game image or a 3D image
  • a game image of a viewpoint corresponding to a user's operation is rendered by a rendering device and presented to the user.
  • a technique of displaying an image rendered by a rendering device using AR (Augment Reality) or VR (Virtual Reality) on a head-mounted display worn by a user for example.
  • the rendered image When the image rendered by the rendering device is presented to the user via the network, for example, the rendered image may be delayed and presented to the user depending on the traffic situation of the network. Therefore, there is known a technique of predicting the posture of the user after the lapse of delay from the posture of the user and rendering an image based on the predicted posture. In such a technique, the rendered image is corrected according to the difference between the predicted posture and the actual posture and presented to the user.
  • an information processing device includes a generation unit and a transmission unit.
  • the generation unit acquires the first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image.
  • the generation unit generates line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number.
  • the generation unit generates the second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
  • the transmission unit transmits the line-of-sight information and the second rendered image to the terminal device.
  • Each of one or more embodiments (including examples and modifications) described below can be implemented independently. On the other hand, at least a part of the plurality of embodiments described below may be carried out in combination with at least a part of other embodiments as appropriate. These plurality of embodiments may contain novel features that differ from each other. Therefore, these plurality of embodiments may contribute to solving different purposes or problems, and may have different effects.
  • FIG. 1 is a diagram for explaining an outline of the proposed technique of the present disclosure.
  • the proposed technique according to the present disclosure is implemented in the information processing system shown in FIG.
  • the information processing system 10 includes a server device 100 and a client device 200.
  • the server device 100 and the client device 200 are connected via a network N such as the Internet.
  • the server device 100 is, for example, a rendering server, and transmits a rendered image (hereinafter, also referred to as a rendered image) to the client device 200.
  • the server device 100 is a game server, and provides a game service to the client device 200 by transmitting a rendered image which is a game image to the client device 200.
  • the server device 100 may be a device that provides AR or VR to the user via the client device 200. In this case, the server device 100 generates, for example, a rendered image according to the user's viewpoint and transmits it to the client device 200.
  • the client device 200 is an information processing device such as a smartphone, a PC, or a game device. Further, the client device 200 may be an information processing device worn on the user's head, such as a head-mounted display (HMD: Head Mounted Display) or smart glasses. The client device 200 displays the rendered image acquired from the server device 100 on the display unit of the own device or the display device connected to the own device.
  • HMD Head Mounted Display
  • the display of the rendered image on the client device 200 may be delayed due to the communication delay generated in the network N.
  • the communication delay generated in the network N.
  • the user's viewpoint moves while the communication delay occurs.
  • the user may feel VR sickness due to the deviation between the viewpoint of the displayed image and the viewpoint of the user.
  • the transmission time depends on the traffic of the Internet line. Therefore, depending on the amount of traffic on the Internet line, the quality of services provided to users such as games may be significantly deteriorated.
  • the server device 100 converts the rendered image into a multi-viewpoint image, and transmits the converted multi-viewpoint image to the client device 200. At this time, the server device 100 generates a multi-viewpoint image with high accuracy by using the information used when generating the rendered image (hereinafter, also referred to as rendering incidental information).
  • the client device 200 generates a display image based on the received multi-viewpoint image according to the current user's viewpoint or the operation status of the game by the user, and presents the display image to the user.
  • the server device 100 acquires the rendered image M0 and the rendering incidental information (step S1).
  • the rendered image is an image generated by the rendering unit (not shown) of the server device 100 based on the operation status of the game and the viewpoint information of the user transmitted from the client device 200.
  • the rendering incidental information is information used by the rendering unit to generate a rendered image, and includes, for example, depth information.
  • the rendering unit uses the rendering incidental information to generate a rendered image based on a predetermined number of viewpoint positions and line-of-sight directions.
  • the server device 100 acquires a rendered image M0 viewed from one viewpoint position and line-of-sight direction from a plurality of images in a viewpoint position and a line-of-sight direction that can be generated by the rendering unit.
  • a plurality of rendered images that can be generated by the rendering unit are arranged and shown in a matrix.
  • the rendered image with hatching is an image that is not actually generated by the rendering unit, and the rendering unit generates the rendered image M0 without hatching.
  • the server device 100 generates the viewpoint buffer information (step S2).
  • the viewpoint buffer information is information used to generate a multi-viewpoint image, and includes a plurality of viewpoint positions and line-of-sight directions.
  • the server device 100 generates viewpoint buffer information based on a game status, a communication delay status (communication status), and the like.
  • the server device 100 generates a viewpoint buffer using the viewpoint buffer information and the rendering incidental information based on the rendered image (step S3).
  • the viewpoint buffer includes a plurality of viewpoint positions included in the viewpoint buffer information and a plurality of images viewed from the line-of-sight direction (hereinafter, also referred to as multi-viewpoint images).
  • FIG. 1 shows a plurality of images that can be viewpoint-converted by the server device 100 arranged in a matrix.
  • the hatched images are images that are not actually converted by the server device 100, and the server device 100 generates the unhatched images M1 to M5 by the viewpoint conversion.
  • the server device 100 can generate a highly accurate multi-viewpoint image by using the rendering incidental information.
  • the server device 100 transmits the viewpoint buffer information and the viewpoint buffer to the client device 200 via the network N (step S4).
  • the client device 200 receives the viewpoint buffer information and the viewpoint buffer from the server device 100 (step S5).
  • the client device 200 generates a display image from the multi-viewpoint image included in the viewpoint buffer based on the current viewpoint of the user (step S6), and presents the display image to the user.
  • the server device 100 transmits a plurality of multi-viewpoint images generated from the rendered image to the client device 200.
  • the server device 100 can generate a high-quality multi-viewpoint image by using the rendering incidental information.
  • the client device 200 generates a display image according to the viewpoint of the current user from a plurality of multi-viewpoint images.
  • the information processing system 10 can present the user with a high-quality display image according to the viewpoint of the current user even if a communication delay occurs in the network N, and the deterioration of the provided service due to the communication delay is deteriorated. It can be suppressed.
  • the information processing system 10 may be an AR system or a VR system.
  • the information processing system 10 may be any system that provides the user with an image corresponding to the movement of the viewpoint, and is not limited to the system that provides the game service.
  • FIG. 2 is a diagram showing an example of the configuration of the information processing system 10 according to the embodiment of the present disclosure.
  • the information processing system includes a server device 100, a client device 200, and a peripheral device 300.
  • the server device 100 distributes a rendered game image (for example, a rendered image).
  • the client device 200 receives the video of the game through the network N, which is a communication line such as the Internet, and reproduces it, for example, on the display unit (not shown) of the peripheral device 300.
  • the peripheral device 300 includes, for example, a display, an HMD, a controller, and the like.
  • the client device 200 detects, for example, command information an operation or operation performed by the user in response to the reproduced video.
  • the client device 200 detects the operation on the controller as command information.
  • the client device 200 detects the movement of the user's head as the movement of the HMD.
  • the client device 200 transmits the detected command information to the server device 100 via the network N.
  • the server device 100 renders an image of the game at the next time according to the received command information and distributes it to the client device 200.
  • the server device 100 is an information processing device that provides an application function to the client device 200.
  • the server device 100 is, for example, a game server (Game Server) that provides a game service as an application function, and transmits a rendered image to be displayed as a game screen to the client device 200.
  • Game Server Game Server
  • the server device 100 receives the command information from the client device 200.
  • command information includes information that can identify the user's viewpoint, such as information on the movement of the user's head, and information on operations by the user with respect to the controller.
  • the server device 100 includes a rendering unit 110, a multi-viewpoint conversion unit 120, an encoding unit 130, a transmission / reception unit 140, and a command information acquisition unit 150.
  • the rendering unit 110 renders the game image.
  • the rendering unit 110 determines, for example, the camera position in the game (for example, the user's viewpoint position), the progress of the game, and the like based on the command information acquired from the client device 200, and produces a game image according to the determination result. Generate.
  • the rendering unit 110 is an application that has a Game Logic function for controlling the progress of the game and provides a game service.
  • the rendering unit 110 generates n1 viewpoint positions and line-of-sight directions (hereinafter, also simply referred to as viewpoints or viewpoint information) rendered images.
  • the rendering unit 110 shall generate a rendered image viewed from one viewpoint.
  • the rendering unit 110 generates a rendered image using the rendering incidental information based on the command information.
  • the rendering incidental information includes, for example, depth information and normal information of objects included in the game. The details of the rendering incidental information will be described later.
  • the time when the rendering unit 110 generates the rendered image is referred to as a rendering time t1.
  • the rendering unit 110 outputs the rendered image, the rendering incidental information, and the information about the game to the multi-viewpoint conversion unit 120.
  • the information about the game includes, for example, information about the title of the game, information about the type of the game such as a race game, a simulation game, and an RPG game, and information such as a game play status.
  • the multi-viewpoint conversion unit 120 generates a viewpoint buffer including a plurality of viewpoint conversion images from the rendered image generated by the rendering unit 110.
  • a plurality of viewpoint conversion images are collectively referred to as a multi-viewpoint image.
  • the multi-viewpoint conversion unit 120 acquires information on the network traffic (delay) status from the transmission / reception unit 140 in addition to the rendered image acquired from the rendering unit 110, rendering incidental information, and information on the game.
  • the multi-viewpoint conversion unit 120 generates viewpoint buffer information from the acquired information and converts the rendered image into a multi-viewpoint image viewed from n2 viewpoints. It is assumed that the number of viewpoints n2 of the multi-viewpoint image is larger than the number of viewpoints n1 of the rendered image (n1 ⁇ n2). Further, the viewpoint buffer information includes the coordinate information of the viewpoint of the multi-view image included in the viewpoint buffer. The details of the viewpoint buffer and the viewpoint buffer information generated by the multi-viewpoint conversion unit 120 will be described later.
  • the multi-viewpoint conversion unit 120 outputs the generated viewpoint buffer and viewpoint buffer information to the encoding unit 130.
  • the multi-viewpoint conversion unit 120 generates in advance an image viewed from a viewpoint around the viewpoint of the rendered image (hereinafter, also referred to as a rendering viewpoint) as a multi-viewpoint image based on the rendered image rendered by the rendering unit 110.
  • the multi-viewpoint conversion unit 120 builds a buffer for images of a plurality of viewpoints as a viewpoint buffer in advance.
  • the server device 100 transmits a game image having such a viewpoint buffer to the client device 200.
  • the GAP between the rendering time t1 and the command time t2 causes a deviation between the viewpoint of the game image and the viewpoint of the image expected by the user.
  • the client device 200 can generate a game image in which the deviation is interpolated using the viewpoint buffer.
  • the client device 200 is based on the viewpoint conversion image of the viewpoint included in the viewpoint buffer and close to the viewpoint of the image expected by the user.
  • a game image can be generated by performing viewpoint conversion. In this way, even if the deviation is large, by generating the game image from the viewpoint conversion image included in the viewpoint buffer, the deviation of the viewpoint can be reduced as compared with the case where the game image is generated from the rendered image. Deterioration of the quality of the game screen can be further suppressed.
  • the encoding unit 130 compresses the viewpoint buffer and the viewpoint buffer information generated by the multi-viewpoint conversion unit 120 into a small capacity, and outputs the compressed bit stream (hereinafter, also referred to as compressed data) to the transmission / reception unit 140. do.
  • the transmission / reception unit 140 is a communication interface (I / F) that communicates with an external device.
  • the transmission / reception unit 140 is realized by, for example, a NIC (Network Interface Card) or the like.
  • the transmission / reception unit 140 converts the compressed data encoded by the encoding unit 130 into a transmission signal and transmits it to the client device 200. Further, the transmission / reception unit 140 receives command information from the client device 200 and notifies the viewpoint information acquisition unit 150 of the reception result.
  • the command information acquisition unit 150 acquires command information via the transmission / reception unit 140.
  • the command information acquisition unit 150 outputs the acquired command information to the rendering unit 110.
  • the client device 200 is an information processing device that receives application functions from the server device 100.
  • the client device 200 receives, for example, a game service as an application function.
  • the client device 200 receives an image (viewpoint buffer) related to the game transmitted from the server device 100, and displays the game image based on the received viewpoint buffer.
  • the client device 200 transmits command information according to the operation or operation of the game by the user to the server device 100.
  • the client device 200 has a transmission / reception unit 210, a decoding unit 220, and an input information acquisition unit 230.
  • the transmission / reception unit 210 is a communication interface (I / F) that communicates with an external device.
  • the transmission / reception unit 210 is realized by, for example, a NIC (Network Interface Card) or the like.
  • the transmission / reception unit 210 receives a transmission signal from the server device 100 as a reception signal, and converts the received reception signal into compressed data. Further, the transmission / reception unit 210 transmits command information to the server device 100.
  • the decoding unit 220 decodes the compressed data received by the transmission / reception unit 210.
  • the decoding unit 220 generates a game image to be presented to the user from the multi-viewpoint image included in the viewpoint buffer based on the user's viewpoint information and the viewpoint buffer information acquired by the input information acquisition unit 230.
  • the user's viewpoint information acquired by the input information acquisition unit 230 is the latest viewpoint information in the current game, that is, information regarding the viewpoint of the game image expected by the user, and is simply "current". Also referred to as "user's viewpoint information”.
  • the decoding unit 220 displays the generated game image on the display unit (not shown) of the peripheral device 300.
  • the input information acquisition unit 230 is a user interface (UI: User Interface) that accepts operations of the peripheral device 300 by the user.
  • the input information acquisition unit 230 converts the operation or operation performed by the user in response to the game image displayed on the peripheral device 300 into command information that can be interpreted by the server device 100. Further, the input information acquisition unit 230 generates viewpoint information (viewpoint information of the current user) of the game image to be displayed on the peripheral device 300.
  • the input information acquisition unit 230 outputs the generated viewpoint information to the decoding unit 220. Further, the input information acquisition unit 230 includes the viewpoint information in the command information and transmits the viewpoint information to the server device 100 via the transmission / reception unit 210.
  • the peripheral device 300 is a device such as a display, smart glasses, HMD, controller, etc. that displays a game image output by the client device 200 and accepts operations by the user.
  • the peripheral device 300 is a display
  • the peripheral device 300 presents a game image to the user.
  • the peripheral device 300 is a controller
  • the peripheral device 300 accepts an operation performed by the user in response to a game image displayed on a display device (not shown).
  • the peripheral device 300 is an HMD
  • the peripheral device 300 displays a game image on a display unit (not shown) and detects the movement of the user's head as the movement of the HMD.
  • the peripheral device 300 is a device different from the client device 200, but the present invention is not limited to this.
  • the peripheral device 300 may be realized as a function of the client device 200.
  • the server device 100 uses the rendering incidental information to generate the viewpoint buffer. Further, the client device 200 generates a game image (display image) according to the viewpoint information of the current user based on the viewpoint buffer. As a result, the information processing system 10 can display the game image as seen from the viewpoint of the current user regardless of the delay of the network N.
  • the viewpoint buffer is generated by the server device 100.
  • the server device 100 uses the rendering incidental information to perform viewpoint conversion of the rendered image.
  • the client device 200 performs viewpoint conversion of the rendered image based on the viewpoint information of the current user.
  • the client device 200 it is difficult for the client device 200 to perform viewpoint conversion using rendering incidental information. This is because the rendering incidental information is inferior in coding efficiency as compared with the multi-viewpoint image, so that the communication amount becomes large.
  • the rendering incidental information includes not only image information but also various information such as information about objects in the game. Therefore, the rendering incidental information has a lower coding efficiency than the viewpoint buffer composed of image information having different viewpoints.
  • the information processing system 10 according to the embodiment of the present disclosure further reduces the image delay due to the viewpoint movement while suppressing the increase in network traffic, as compared with the case where the viewpoint conversion of the rendered image is performed by the client device 200. be able to. As described above, the information processing system 10 according to the embodiment of the present disclosure can provide a higher quality service.
  • FIG. 3 is a block diagram showing a configuration example of the multi-viewpoint conversion unit 120 according to the embodiment of the present disclosure. As shown in FIG. 3, the multi-viewpoint conversion unit 120 includes a viewpoint buffer control unit 121 and a multi-viewpoint generation unit 122.
  • the viewpoint buffer control unit 121 determines the viewpoint information of the multi-view image generated by the multi-view generation unit 122.
  • the viewpoint buffer control unit 121 is based on at least one of the viewpoint information of the rendered image generated by the rendering unit 110, the information about the game such as the game title and the play status, and the traffic information (communication status) of the network N. To decide.
  • FIGS. 4 to 7 are diagrams for explaining an example of viewpoint information determined by the viewpoint buffer control unit 121 according to the embodiment of the present disclosure.
  • FIGS. 4 to 7 show a plurality of rendered images that can be generated by the rendering unit 110 and examples of the multi-viewpoint images that can be generated by the multi-viewpoint generation unit 122 side by side in a matrix.
  • the viewpoint buffer control unit 121 determines the viewpoint information of the multi-view image generated by the multi-view generation unit 122 from a predetermined range including the viewpoint information of the rendered image generated by the rendering unit 110 (hereinafter, also referred to as rendering viewpoint information). do.
  • the viewpoint buffer control unit 121 determines the viewpoint information indicated by a cross from the circular range centered on the rendering viewpoint information as the viewpoint information of the viewpoint converted image (hereinafter, also referred to as converted viewpoint information). do.
  • FIG. 4 shows an example in which the viewpoint buffer control unit 121 determines 12 conversion viewpoint information.
  • the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the situation of the game or the network N.
  • the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the delay status (communication status) of the network N. More specifically, the viewpoint buffer control unit 121 determines the conversion viewpoint information from a larger predetermined range as the delay of the network N increases.
  • the viewpoint buffer control unit 121 determines, for example, 12 conversion viewpoint information from the range R1 shown in FIG. Further, when the delay of the network N is large, as shown in FIG. 5, for example, 12 conversion viewpoint information is determined from the range R2 wider than the range R1 (see FIG. 4).
  • the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the delay of the network N. As a result, even when the delay of the network N is large, the game image seen from the viewpoint of the current user can be generated on the client device 200 side, and the deterioration of the quality of the service provided to the user can be further suppressed.
  • the viewpoint buffer control unit 121 evenly determines the conversion viewpoint information from the ranges R1 and R2. Therefore, the viewpoint buffer control unit 121 determines the conversion viewpoint information more densely as the delay of the network N is smaller, and determines the conversion viewpoint information sparsely as the delay is larger.
  • the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the traffic (delay) status of the network N, but the present invention is not limited to this.
  • the viewpoint buffer control unit 121 may determine the conversion viewpoint information according to the information about the game such as the game title and the play situation (scene).
  • the viewpoint buffer control unit 121 can determine the conversion viewpoint information according to the game title and the type of the game. For example, in the case of a game in which the viewpoint moves quickly, such as a racing game, the viewpoint buffer control unit 121 can determine the converted viewpoint information from a larger range. On the other hand, for example, in the case of a game in which the viewpoint movement is small, the viewpoint buffer control unit 121 can determine the converted viewpoint information from a smaller range.
  • the viewpoint buffer control unit 121 may determine the conversion viewpoint information according to the play situation. For example, the viewpoint buffer control unit 121 determines the conversion viewpoint information from a larger range when playing in a space with a wide movement range such as outdoors, and when playing in a space with a narrow movement range such as indoors. The conversion viewpoint information may be determined from a narrower range.
  • the viewpoint buffer control unit 121 can determine the conversion viewpoint information according to the information about the game such as the game title and the play status, in other words, the content including the rendered image and the scene of the rendered image.
  • the viewpoint buffer control unit 121 can determine the converted viewpoint information according to the information related to the speed of the viewpoint movement among the information related to the game.
  • the viewpoint buffer control unit 121 evenly determines the conversion viewpoint information from the predetermined range, but the present invention is not limited to this.
  • the viewpoint buffer control unit 121 is densely located in the vicinity of the rendering viewpoint information (center of the range R3) in the circular range R3 centered on the rendering viewpoint information, and around the range R3.
  • the conversion viewpoint information may be determined so as to be sparse.
  • the viewpoint buffer control unit 121 can determine the conversion viewpoint information at the density shown in FIG.
  • the viewpoint buffer control unit 121 may determine the conversion viewpoint information according to the rendering viewpoint information, particularly the history of the rendering viewpoint information. In this case, the viewpoint buffer control unit 121 predicts the viewpoint information of the future time (for example, the command generation time t2) from the change (history) of the past rendering viewpoint information. The viewpoint buffer control unit 121 determines the converted viewpoint information from the range including the predicted viewpoint information.
  • the viewpoint buffer control unit 121 predicts the viewpoint information in a predetermined direction (P1 in FIG. 7), and determines the conversion viewpoint information from the long range R4 in the predicted direction (upper right direction in FIG. 7).
  • the number of converted viewpoint information (number of viewpoints) determined by the viewpoint buffer control unit 121 may be constant or variable.
  • the server device 100 may change the number of conversion viewpoint information according to the situation. This makes it possible to improve the efficiency of the calculation cost of the viewpoint conversion by the server device 100.
  • the server device 100 determines a certain number of conversion viewpoint information. May be good.
  • FIG. 8 is a diagram for explaining an example of viewpoint information according to the embodiment of the present disclosure.
  • the viewpoint information includes the viewpoint position and the line-of-sight direction.
  • the viewpoint position is represented by the viewpoint (camera) coordinates (x wk , y wk , z wk ) representing the position of the camera (or the user's head) that saw the image to be rendered.
  • the coordinate system of the viewpoint coordinates is, for example, a world coordinate system in which game objects are arranged.
  • the line-of-sight direction is represented by a line-of-sight unit vector (x vk , y vk , x vk ) indicating the direction of the camera (or the user's head).
  • the viewpoint buffer control unit 121 determines a plurality of sets of the viewpoint position (viewpoint coordinates) and the line-of-sight direction (line-of-sight unit vector) as a plurality of converted viewpoint information.
  • FIG. 9 is a diagram showing an example of conversion viewpoint information determined by the viewpoint buffer control unit 121 according to the embodiment of the present disclosure.
  • FIG. 9 shows a case where the viewpoint buffer control unit 121 determines N conversion viewpoint information.
  • the viewpoint buffer control unit 121 sets ⁇ (x w1 , y w1 , z w1 ), (x v1 , y v1 , x v1 ) ⁇ , ⁇ (x w1, y w1, z w1) ⁇ , ⁇ (x w1, y w1, x v1) ⁇ , ⁇ (x w1, y w1, z w1) ⁇ , ⁇ (x w1, y v1, x v1) ⁇ , ⁇ (x w1, y w1, z w1), x w2 , y w2 , z w2 ), (x v2 , y v2 , x v2 ) ⁇ ... ⁇ (x w2 , y w2 , z w2 ), (x v2 , y v2 , x v2 ) ⁇ ... ⁇ (x w2 , y w2
  • the viewpoint buffer control unit 121 outputs the determined set of viewpoint coordinates and line-of-sight unit vector to the multi-viewpoint generation unit 122 (see FIG. 3).
  • the converted viewpoint information shown in FIGS. 8 and 9 is an example, and the present invention is not limited to this.
  • the converted viewpoint information may be expressed in any format as long as the multi-viewpoint generation unit 122 can convert the viewpoint of the rendered image based on the converted viewpoint information.
  • the viewpoint buffer control unit 121 may output, for example, the conversion viewpoint information stored in the storage unit (not shown) to the multi-viewpoint generation unit 122 as the determined conversion viewpoint information, or may be determined using a function. You may.
  • the viewpoint buffer control unit 121 determines the conversion viewpoint information using a function will be described with reference to FIG. 10, but the method for determining the conversion viewpoint information is not limited to this.
  • FIG. 10 is a block diagram showing a configuration example of the viewpoint buffer control unit 121 according to the embodiment of the present disclosure.
  • the viewpoint buffer control unit 121 includes a viewpoint function generation unit 1211, a line-of-sight function generation unit 1213, a sampling unit 1212, and a line-of-sight vector generation unit 1214.
  • the viewpoint buffer control unit 121 first determines a set of a viewpoint density function and a line-of-sight function in a three-dimensional space.
  • the viewpoint buffer control unit 121 determines a predetermined viewpoint density function and line-of-sight function.
  • the viewpoint buffer control unit 121 may determine the viewpoint density function and the line-of-sight function according to information about the game, information about network traffic, rendering viewpoint information, and the like.
  • the viewpoint function generation unit 1211 determines, for example, an arbitrary probability density function P (x, y, z) as the viewpoint density function P. For example, the viewpoint function generation unit 1211 determines the viewpoint density function P according to information about the game, information about network traffic, rendering viewpoint information, etc. from a plurality of probability density functions stored in advance in the storage unit (not shown). do.
  • the viewpoint function generation unit 1211 may change the parameters of the viewpoint density function P according to the information about the game, the information about the network traffic, the rendering viewpoint information, and the like.
  • the viewpoint function generation unit 1211 outputs the determined viewpoint density function P to the sampling unit 1212.
  • the line-of-sight function generation unit 1213 determines a function that inputs an arbitrary viewpoint coordinate and outputs a line-of-sight unit vector as a line-of-sight function V. For example, the line-of-sight function generation unit 1213 determines the line-of-sight function V according to information about a game, information about network traffic, rendering viewpoint information, and the like from a plurality of line-of-sight functions stored in advance in a storage unit (not shown).
  • the line-of-sight function generation unit 1213 may change the parameters of the line-of-sight function V according to information about the game, information about network traffic, rendering viewpoint information, and the like.
  • the line-of-sight function generation unit 1213 outputs the determined line-of-sight function V to the line-of-sight vector generation unit 1214.
  • the sampling unit 1212 determines, for example, N viewpoint coordinates to be used for the viewpoint conversion from the viewpoint density function P by using various sampling methods. Examples of the sampling method include an inverse function method, rejection sampling, and Markov chain Monte Carlo method. The sampling unit 1212 outputs the determined viewpoint coordinates to the line-of-sight vector generation unit 1214.
  • the line-of-sight vector generation unit 1214 determines the line-of-sight unit vector by inputting the viewpoint coordinates determined by the sampling unit 1212 into the line-of-sight function V.
  • the line-of-sight vector generation unit 1214 outputs the determined N viewpoint coordinates and line-of-sight unit vectors to the multi-viewpoint generation unit 122 (see FIG. 3) as viewpoint buffer information.
  • the line-of-sight vector generation unit 1214 outputs the viewpoint coordinates and the line-of-sight unit vector as the conversion viewpoint information, but the present invention is not limited to this.
  • the viewpoint buffer control unit 121 may output the viewpoint density function P and the line-of-sight function V to the multi-viewpoint generation unit 122 as viewpoint buffer information.
  • the line-of-sight vector generation unit 1214 may determine the parameters of the predetermined viewpoint density function P and the predetermined line-of-sight function V, and output the determined parameters to the multi-viewpoint generation unit 122 as the viewpoint buffer information.
  • the viewpoint buffer control unit 121 may output the viewpoint density function P and the line-of-sight function V as the viewpoint buffer information, or information (for example, parameters, etc.) for uniquely determining the function.
  • the sampling unit 1212 and the line-of-sight vector generation unit 1214 of the viewpoint buffer control unit 121 may be omitted.
  • the information on the sampling method by the sampling unit 1212 and the function to be used is shared by the multi-viewpoint generation unit 122, the decoding unit 220 of the client device 200 (see FIG. 2), and the like.
  • the viewpoint buffer control unit 121 generates the viewpoint buffer information according to the communication status of the network N and the game information, but the present invention is not limited to this.
  • the multi-viewpoint conversion unit 120 may generate a viewpoint conversion image based on a predetermined constant viewpoint buffer information regardless of the above-mentioned information.
  • the server device 100 may omit the viewpoint buffer control unit 121 by storing the viewpoint buffer information in a storage unit (not shown) in advance, for example.
  • the multi-viewpoint generation unit 122 Based on the rendered image, the multi-viewpoint generation unit 122 generates a viewpoint conversion image in which the line-of-sight unit vector direction is viewed from the viewpoint coordinates included in the conversion viewpoint information.
  • FIG. 3 shows a plurality of rendered images that can be generated by the rendering unit 110 arranged in a matrix.
  • the hatched image is an image that is not actually rendered by the rendering unit 110, and the rendering unit 110 transfers the rendered image M0 without hatching to the multi-viewpoint generation unit 122. Output.
  • a plurality of viewpoint conversion images that can be generated by the multi-viewpoint generation unit 122 are arranged and shown in a matrix.
  • the viewpoint conversion image with hatching is an image that is not actually generated by the multi-viewpoint generation unit 122
  • the multi-viewpoint generation unit 122 is a viewpoint conversion image without hatching.
  • the multi-viewpoint generation unit 122 generates a viewpoint conversion image using the rendering incidental information.
  • the rendering incidental information includes, for example, metadata generated each time a game image is rendered (hereinafter referred to as run-time metadata) and metadata created during game production (hereinafter referred to as production-time metadata). Is included.
  • run-time metadata is the normal information of objects in the game.
  • run-time metadata include information related to light reflection, such as object specular, roughness, albedo, and transparency.
  • One example of run-time metadata is information about the speed of objects in the game.
  • An example of run-time metadata is information that represents the material of an object, such as the material ID of the object in the game.
  • An example of run-time metadata is the Depth information of objects in the game.
  • run-time metadata is metadata created depending on the player's viewpoint, but the run-time metadata also includes metadata created independently of the player's viewpoint.
  • run-time metadata include an in-game user interface (UI) showing the player's status, the progress of the game, and the Reflection map.
  • UI in-game user interface
  • production metadata there is data that is asynchronous to the frame of the game loop and does not depend on the player's viewpoint, such as mesh (3D model), texture, animation pattern of objects in the game, and layout data. Be done.
  • the multi-viewpoint generation unit 122 generates a viewpoint conversion image from the rendered image using the above-mentioned rendering incidental information.
  • the multi-viewpoint generation unit 122 may generate a viewpoint-converted image from a rendered image by using, for example, a model-based method, or may generate a viewpoint-converted image by using a DNN-based method. In this way, the viewpoint conversion image can be generated by using various existing methods.
  • the multi-viewpoint generation unit 122 may perform generation processing specialized for each type of game or each title of the game. For example, when the multi-viewpoint generation unit 122 executes a viewpoint conversion image generation process using a DNN-based method, the multi-viewpoint generation unit 122 can be generated by preparing a coefficient database according to the type of game or the like. It is possible to perform generation processing according to the type of game. Further, when the multi-viewpoint generation unit 122 executes the viewpoint conversion image generation process using the model-based method, the multi-viewpoint generation unit 122 can be used as a game by preparing a model according to the type of the game or the like. It is possible to perform generation processing according to the type of.
  • the multi-viewpoint generation unit 122 can perform viewpoint conversion with high accuracy by executing the generation process according to the type of the game, the title of the game, in other words, the rendering incidental information.
  • the multi-viewpoint generation unit 122 may change the generation process for all game types or game titles.
  • the multi-viewpoint generation unit 122 may execute a specific generation process in a specific game.
  • the multi-view generation unit 122 When the viewpoint buffer information includes N conversion viewpoint information, the multi-view generation unit 122 generates N viewpoint conversion images.
  • the multi-viewpoint generation unit 122 outputs the generated N viewpoint conversion images to the encoding unit 130 as a viewpoint buffer (multi-viewpoint image). At this time, the multi-viewpoint generation unit 122 outputs the viewpoint buffer information to the encoding unit 130 in addition to the viewpoint buffer.
  • the viewpoint buffer control unit 121 generates the viewpoint buffer information, but the present invention is not limited to this.
  • the multi-viewpoint generation unit 122 may perform viewpoint conversion based on a certain number of conversion viewpoint information within a certain range regardless of information on a game, information on network traffic, and the like.
  • the multi-viewpoint generation unit 122 may generate the viewpoint conversion image based on the viewpoint buffer information stored in the storage unit (not shown), and the viewpoint buffer control unit 121 may be omitted.
  • the encoding unit 130 Upon receiving the viewpoint buffer information and the viewpoint buffer from the multi-viewpoint conversion unit 120, the encoding unit 130 encodes the viewpoint buffer information and the viewpoint buffer for transmission to the client device 200. The encoding unit 130 encodes the viewpoint buffer information so that it can be independently decoded prior to the viewpoint buffer.
  • the encoding method performed by the encoding unit 130 may be a model-based method or a DNN-based method.
  • the encoding unit 130 may or may not perform encoding in consideration of the viewpoint buffer information and the viewpoint buffer in the time direction. That is, the coding performed by the encoding unit 130 may be intra-coding or inter-coding.
  • the multi-viewpoint conversion unit 120 generates a multi-viewpoint image with a slightly deviated viewpoint. From this, it is considered that each viewpoint-converted image included in the multi-viewpoint image does not change significantly, and the viewpoint-converted image has high redundancy. Therefore, the encoding unit 130 can compress the multi-viewpoint image (viewpoint buffer) at a high compression rate.
  • the decoding unit 220 first decodes the viewpoint buffer information from the compressed data received by the transmission / reception unit 210, and decodes the viewpoint conversion image included in the viewpoint buffer from the decoded viewpoint buffer information and the current user's viewpoint information.
  • the decoding unit 220 generates a display image (game image) based on the current user's viewpoint information and the decoded viewpoint conversion image.
  • FIG. 11 is a block diagram showing a configuration example of the decoding unit 220 according to the embodiment of the present disclosure.
  • the decoding unit 220 shown in FIG. 11 includes a buffer information decoding unit 221, a decoding control unit 222, a viewpoint buffer decoding unit 223, and an image generation unit 224.
  • the buffer information decoding unit 221 receives compressed data as received data from the transmission / reception unit 210.
  • the buffer information decoding unit 221 decodes the viewpoint buffer information from the compressed data and outputs it to the decoding control unit 222.
  • the decode control unit 222 acquires the viewpoint information of the current user from the input information acquisition unit 230.
  • the current user's viewpoint information is viewpoint information determined by command information such as a user's operation or operation, and indicates a viewpoint position and a line-of-sight direction to be rendered as a display image.
  • the decode control unit 222 coordinates one coordinate system to match the other coordinate system. Perform the conversion.
  • the decoding control unit 222 performs processing in a unified manner in the coordinate system of the conversion viewpoint information.
  • the decoding control unit 222 determines the viewpoint information (hereinafter referred to as decoding viewpoint information) of the viewpoint conversion image to be decoded by the viewpoint buffer decoding unit 223 based on the viewpoint information and the viewpoint buffer information of the current user.
  • FIG. 12 is a diagram for explaining the decoding viewpoint information determined by the decoding control unit 222 according to the embodiment of the present disclosure.
  • FIG. 12 shows a plurality of viewpoint conversion images that can be generated by the multi-viewpoint generation unit 122 side by side in a matrix. Further, in FIG. 12, the viewpoint information of the viewpoint converted image actually included in the viewpoint buffer, in other words, the converted viewpoint information included in the viewpoint buffer information is shown by a cross. Further, "x" in FIG. 12 indicates the viewpoint information VP0 of the current user.
  • the decoding control unit 222 determines M conversion viewpoint information from the conversion viewpoint information VP1 to VP12 as the decoding viewpoint information based on the current user viewpoint information VP0. For example, the decoding control unit 222 selects the upper M conversion viewpoint information as the decoding viewpoint information in order of increasing distance from the current user viewpoint information VP0.
  • the conversion viewpoint information VP4, VP7, VP8, and VP11 selected by the decoding control unit 222 are indicated by a circled cross.
  • the method and number of decoding viewpoint information selected by the decoding control unit 222 is not limited to the example shown in FIG.
  • the conversion viewpoint information that can be generated by interpolating the display image corresponding to the current user's viewpoint information VP0 by the image generation unit 224 in the subsequent stage may be selected.
  • the selection method and the number may be appropriately changed depending on the processing by the image generation unit 224.
  • the decoding control unit 222 outputs the selected decoding viewpoint information to the viewpoint buffer decoding unit 223. Further, the decoding control unit 222 outputs the current user's viewpoint information VP0 to the image generation unit 224.
  • the viewpoint buffer decoding unit 223 decodes the viewpoint conversion image corresponding to the decoding viewpoint information from the received data acquired from the transmission / reception unit 210.
  • the viewpoint buffer decoding unit 223 acquires the conversion viewpoint information VP4, VP7, VP8, and VP11 as the decoding viewpoint information.
  • the viewpoint buffer decoding unit 223 decodes the viewpoint conversion image corresponding to the conversion viewpoint information VP4, VP7, VP8, and VP11.
  • the viewpoint buffer decoding unit 223 outputs the decoded M viewpoint conversion images and the decoded viewpoint information to the image generation unit 224.
  • the image generation unit 224 generates game images corresponding to the current user's viewpoint information acquired from the decoding control unit 222 based on the M viewpoint conversion images acquired from the viewpoint buffer decoding unit 223 and the decoding viewpoint information. ..
  • the image generation unit 224 interpolates and generates a game image from, for example, M viewpoint conversion images.
  • the image generation unit 224 can interpolate and generate a game image by using various methods.
  • the image generation unit 224 can interpolate and generate a game image using, for example, a model-based method or a DNN-based method.
  • the image generation unit 224 presents the game image to the user by outputting the generated game image to the display unit (not shown) of the peripheral device 300.
  • the decoding unit 220 decodes the number of viewpoint conversion images used for generating the game image based on the viewpoint information VP0 of the current user. As a result, the decoding unit 220 can reduce the processing load as compared with the case of decoding all the viewpoint conversion images included in the viewpoint buffer.
  • the input information acquisition unit 230 detects the operation of the peripheral device 300 (for example, the controller) by the user and the operation of the user by the peripheral device 300 (for example, HMD).
  • the input information acquisition unit 230 converts the detection result into command information that can be interpreted by, for example, the rendering unit 110 of the server device 100, and generates the viewpoint information of the current user.
  • the input information acquisition unit 230 outputs the viewpoint information of the current user to the decoding unit 220. Further, the input information acquisition unit 230 includes the viewpoint information of the current user in the command information and transmits it to the server device 100 via the transmission / reception unit 210.
  • FIG. 13 is a flowchart showing the flow of the multi-viewpoint conversion process executed by the server device 100 according to the embodiment of the present disclosure. While providing the service to the client device 200, the server device 100 repeatedly executes the multi-viewpoint conversion process shown in FIG. 13, for example, at a predetermined cycle.
  • the server device 100 acquires a rendered image viewed from the viewpoint position and the line-of-sight direction of the first number (for example, one) (step S101). Further, the server device 100 acquires the rendering incidental information used for generating the rendered image (step S102).
  • the server device 100 generates viewpoint buffer information including a second number (for example, 12) of viewpoint positions and line-of-sight directions, which is larger than the first number (step S103).
  • the server device 100 generates a viewpoint buffer including the viewpoint position and the viewpoint conversion image seen from the line-of-sight direction included in the viewpoint buffer information (step S104).
  • the server device 100 generates a viewpoint buffer based on the rendered image and the rendering incidental information.
  • the server device 100 encodes the viewpoint buffer information and the viewpoint buffer (step S105), and transmits the obtained transmission data to the client device 200 (step S106).
  • FIG. 14 is a flowchart showing a flow of display image generation processing executed by the client device 200 according to the embodiment of the present disclosure.
  • the client device 200 receives a signal from the server device 100, the client device 200 executes the display image generation process shown in FIG.
  • the client device 200 receives the received signal (step S201).
  • the client device 200 first decodes the viewpoint buffer information from the received signal (step S202).
  • the client device 200 selects the converted viewpoint information corresponding to the viewpoint converted image to be decoded from the converted viewpoint information included in the viewpoint buffer information as the decoded viewpoint information (step S203).
  • the client device 200 decodes the viewpoint buffer and acquires a viewpoint conversion image corresponding to the decoded viewpoint information (step S204).
  • the client device 200 interpolates and generates a display image (game image) viewed from the current user's viewpoint using the decoded viewpoint conversion image (step S205).
  • the control device for controlling the server device 100 or the client device 200 of the present embodiment may be realized by a dedicated computer system or a general-purpose computer system.
  • a program for executing the above operation is stored and distributed in a computer-readable recording medium such as an optical disk, a semiconductor memory, a magnetic tape, or a flexible disk.
  • the control device is configured by installing the program in a computer and executing the above-mentioned processing.
  • the control device may be an external device (for example, a personal computer) of the server device 100 or the client device 200. Further, the control device may be an internal device of the server device 100 or the client device 200.
  • the above communication program may be stored in a disk device provided in a server device on a network such as the Internet so that it can be downloaded to a computer or the like.
  • the above-mentioned functions may be realized by cooperation between the OS (Operating System) and the application software.
  • the part other than the OS may be stored in a medium and distributed, or the part other than the OS may be stored in the server device so that it can be downloaded to a computer or the like.
  • each component of each device shown in the figure is a functional concept, and does not necessarily have to be physically configured as shown in the figure. That is, the specific form of distribution / integration of each device is not limited to the one shown in the figure, and all or part of them may be functionally or physically distributed / physically in arbitrary units according to various loads and usage conditions. Can be integrated and configured.
  • the present embodiment includes a device or any configuration constituting the system, for example, a processor as a system LSI (Large Scale Integration), a module using a plurality of processors, a unit using a plurality of modules, and a unit. It can also be implemented as a set or the like (that is, a configuration of a part of the device) to which other functions are added.
  • a processor as a system LSI (Large Scale Integration)
  • a module using a plurality of processors a unit using a plurality of modules
  • a unit that is, a configuration of a part of the device
  • the system means a set of a plurality of components (devices, modules (parts), etc.), and it does not matter whether all the components are in the same housing. Therefore, a plurality of devices housed in separate housings and connected via a network, and a device in which a plurality of modules are housed in one housing are both systems. ..
  • the present embodiment can have a cloud computing configuration in which one function is shared by a plurality of devices via a network and jointly processed.
  • the server device 100 (an example of an information processing device) according to the embodiment of the present disclosure includes a multi-viewpoint conversion unit 120 (an example of a generation unit) and a transmission / reception unit 140 (an example of a transmission unit). Be prepared.
  • the multi-viewpoint conversion unit 120 includes a rendered image (an example of the first rendered image) and rendering as seen from the viewpoint position and the line-of-sight direction (an example of the first number of viewpoint positions and the line-of-sight direction) included in the rendered viewpoint information.
  • Acquires incidental information (an example of rendering information used for rendering a rendered image).
  • the multi-viewpoint conversion unit 120 generates viewpoint buffer information (an example of line-of-sight information regarding a second number of viewpoint positions and line-of-sight directions that is larger than the first number).
  • the multi-viewpoint conversion unit 120 generates a viewpoint conversion image (an example of the second rendered image) seen from the viewpoint position and the line-of-sight direction included in the decoded viewpoint information based on the rendered image and the rendering incidental information.
  • the multi-viewpoint conversion unit 120 transmits the viewpoint buffer including the viewpoint buffer information and the viewpoint conversion image to the client device 200 (an example of the terminal device).
  • the server device 100 can present the user with a high-quality image with further reduced delay, and can further suppress the deterioration of the quality of the service provided to the user.
  • the multi-viewpoint conversion unit 120 of the server device 100 determines the viewpoint position to be included in the viewpoint buffer information according to the delay situation generated in the communication with the client device 200.
  • the server device 100 can generate a viewpoint conversion image according to the delay situation, and can present the user with a high-quality image with further reduced delay.
  • the multi-viewpoint conversion unit 120 of the server device 100 determines the viewpoint position to be included in the viewpoint buffer based on the history of the rendering viewpoint information (an example of the viewpoint position of the rendered image acquired in the past). ..
  • the server device 100 can generate a viewpoint conversion image according to the user's operation or movement, and can present the user with a high-quality image with further reduced delay.
  • the multi-viewpoint conversion unit 120 of the server device 100 determines the viewpoint position to be included in the viewpoint buffer information according to the type and title of the game (an example of the content including the rendered image).
  • the server device 100 can generate a viewpoint conversion image according to the user's operation or movement peculiar to the content, and can present the user with a high-quality image with further reduced delay.
  • the multi-viewpoint conversion unit 120 of the server device 100 determines the viewpoint position to be included in the viewpoint buffer information according to the game play situation (an example of the scene of the rendered image).
  • the server device 100 can generate a viewpoint conversion image that can correspond to the user's operation and movement according to the scene of the content (for example, the play situation), and the user can obtain a high-quality image with further reduced delay. Can be presented to.
  • the rendering information according to the embodiment of the present disclosure is run-time metadata (an example of information generated when rendering a rendered image).
  • the server device 100 can generate a viewpoint conversion image with higher image quality by generating the viewpoint conversion image using the runtime metadata.
  • the rendering information according to the embodiment of the present disclosure is production-time metadata (an example of information preset for rendering a rendered image).
  • the server device 100 can generate a viewpoint-converted image with higher image quality by generating the viewpoint-converted image using the metadata at the time of production.
  • the multi-viewpoint conversion unit 120 of the server device 100 calculates the viewpoint density function (an example of the viewpoint function representing the distribution of the viewpoint position), the line-of-sight function, and the calculated viewpoint density function.
  • the viewpoint buffer information is generated using the line-of-sight function.
  • the transmission / reception unit 140 of the server device 100 transmits information regarding the viewpoint density function and the line-of-sight function as viewpoint buffer information.
  • the server device 100 can reduce the amount of information of the viewpoint buffer information.
  • the client device 200 (an example of an information processing device) according to the embodiment of the present disclosure includes a transmission / reception unit 210 (an example of a reception unit) and a decoding unit 220 (an example of a generation unit).
  • the receiving unit 210 is a viewpoint converted image generated based on the rendered image seen from the viewpoint position and the line-of-sight direction included in the rendered viewpoint information and the rendering incidental information, and the viewpoint position and the line of sight included in the converted viewpoint information.
  • the viewpoint buffer including the viewpoint conversion image viewed from the direction and the viewpoint buffer information including the conversion viewpoint information are received.
  • the decoding unit 220 generates a display image based on the viewpoint conversion image and the viewpoint buffer information.
  • the client device 200 can present the user with a high-quality image with further reduced delay, and can further suppress the deterioration of the quality of the service provided to the user.
  • the decoding unit 220 of the client device 200 selects and selects at least one viewpoint position and line-of-sight direction included in the viewpoint buffer information based on the user's current viewpoint position and line-of-sight direction.
  • the display image is generated based on the viewpoint conversion image viewed from the viewpoint position and the line-of-sight direction.
  • the client device 200 can present a high-quality image to the user while reducing the number of viewpoint-converted images to be decoded.
  • the information processing system 10 includes a server device 100 and a client device 200.
  • the server device 100 includes a multi-viewpoint conversion unit 120 and a transmission / reception unit 140.
  • the multi-viewpoint conversion unit 120 acquires the rendered image as seen from the viewpoint position and the line-of-sight direction included in the rendered viewpoint information, and the rendered image incidental information.
  • the multi-viewpoint conversion unit 120 generates viewpoint buffer information.
  • the multi-viewpoint conversion unit 120 generates a viewpoint conversion image seen from the viewpoint position and the line-of-sight direction included in the decoded viewpoint information based on the rendered image and the rendering incidental information.
  • the transmission / reception unit 140 transmits the viewpoint buffer including the viewpoint buffer information and the viewpoint conversion image to the client device 200.
  • the client device 200 includes a transmission / reception unit 210 and a decoding unit 220.
  • the transmission / reception unit 210 receives the viewpoint conversion image and the viewpoint buffer information.
  • the decoding unit 220 generates a display image based on the viewpoint conversion image and the viewpoint buffer information.
  • the information processing system 10 can present the user with a high-quality image with further reduced delay, and can further suppress the deterioration of the quality of the service provided to the user.
  • the present technology can also have the following configurations.
  • (1) The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
  • the line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
  • a generation unit that generates a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
  • a transmission unit that transmits the line-of-sight information and the second rendered image to the terminal device, and Information processing device equipped with.
  • the information processing apparatus according to any one of (1) to (5), wherein the rendering information is information generated when the first rendered image is rendered.
  • the rendering information is information preset for rendering the first rendered image.
  • the generation unit calculates a viewpoint function representing the distribution of the viewpoint position and a line-of-sight function representing the line-of-sight direction according to the viewpoint position, and uses the calculated viewpoint function and the line-of-sight function to obtain the second line-of-sight function. Generate the viewpoint position and the line-of-sight direction of the number, The transmitting unit transmits information about the viewpoint function and the line-of-sight function as the line-of-sight information.
  • the information processing apparatus according to any one of (1) to (7).
  • (9) The first rendered image viewed from the viewpoint position and the line-of-sight direction of the first number, and the second rendered image generated based on the rendering information used for rendering the first rendered image. Reception of receiving a second rendered image viewed from the viewpoint position and the line-of-sight direction of a second number larger than the first number, and line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number.
  • Department and A generation unit that generates a display image based on the second rendered image and the line-of-sight information, Information processing device equipped with.
  • the generation unit selects at least one of the viewpoint positions and the line-of-sight directions included in the line-of-sight information based on the user's current viewpoint position and the line-of-sight direction, and views from the selected viewpoint position and the line-of-sight direction.
  • the information processing apparatus according to (9), which generates the display image based on the second rendered image.
  • An information processing system including an information processing device and a terminal device.
  • the information processing device is The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
  • the line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
  • a generation unit that generates a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
  • a transmission unit that transmits the line-of-sight information and the second rendered image, and Equipped with The terminal device is A receiving unit that receives the second rendered image and the line-of-sight information.
  • a generation unit that generates a display image based on the second rendered image and the line-of-sight information, Information processing system equipped with.
  • the first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
  • the line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
  • a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number is generated.
  • the line-of-sight information and the second rendered image are transmitted to the terminal device. Information processing method.
  • Information processing system 100 Server equipment 110 Rendering unit 120 Multi-viewpoint conversion unit 130 Encoding unit 140, 210 Transmission / reception unit 150 Command information acquisition unit 200 Client equipment 220 Decoding unit 230 Input information acquisition unit 300 Peripheral devices

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Image Generation (AREA)
  • Processing Or Creating Images (AREA)

Abstract

An information processing device (100) is provided with a generation unit (120) and a transmission unit (140). The generation unit (120) acquires a first rendering image viewed from each of a first number of viewpoint positions and visual line directions, and rendering information used in rendering the first rendering image. The generation unit (120) generates visual line information relating to a second number of viewpoint positions and visual line directions, the second number being greater than the first number . The generation unit (120) generates a second rendering image viewed from each of the second number of viewpoint positions and visual line directions, on the basis of the first rendering image and the rendering information. The transmission unit (140) transmits, to a terminal device (200), the visual line information and the second rendering image.

Description

情報処理装置、情報処理システム及び情報処理方法Information processing equipment, information processing system and information processing method
 本開示は、情報処理装置、情報処理システム及び情報処理方法に関する。 This disclosure relates to an information processing device, an information processing system, and an information processing method.
 従来、ゲーム画像や3D画像などレンダリング装置でレンダリングした画像を、ネットワークを介してユーザに提示する技術が存在する。例えば、ゲームストリーミングでは、ユーザの操作に応じた視点のゲーム画像がレンダリング装置でレンダリングされ、ユーザに提示される。また、AR(Augment Reality;拡張現実)や、VR(Virtual Reality;仮想現実)を用いて、レンダリング装置でレンダリングした画像を、例えばユーザが装着したヘッドマウントディスプレイに表示する技術が知られている。 Conventionally, there is a technique of presenting an image rendered by a rendering device such as a game image or a 3D image to a user via a network. For example, in game streaming, a game image of a viewpoint corresponding to a user's operation is rendered by a rendering device and presented to the user. Further, there is known a technique of displaying an image rendered by a rendering device using AR (Augment Reality) or VR (Virtual Reality) on a head-mounted display worn by a user, for example.
 レンダリング装置でレンダリングした画像を、ネットワークを介してユーザに提示する場合、例えばネットワークのトラフィック状況により、レンダリング画像が遅延してユーザに提示される場合がある。そこで、ユーザの姿勢から遅延経過後のユーザの姿勢を予測し、予測した姿勢に基づいて画像をレンダリングする技術が知られている。かかる技術では、予測した姿勢と実際の姿勢との差分に応じてレンダリングした画像を補正してユーザに提示する。 When the image rendered by the rendering device is presented to the user via the network, for example, the rendered image may be delayed and presented to the user depending on the traffic situation of the network. Therefore, there is known a technique of predicting the posture of the user after the lapse of delay from the posture of the user and rendering an image based on the predicted posture. In such a technique, the rendered image is corrected according to the difference between the predicted posture and the actual posture and presented to the user.
特開2007-79664号公報Japanese Unexamined Patent Publication No. 2007-79664
 上記技術では、例えば、ユーザが大きく動いた場合など、予測した姿勢と実際の姿勢との差分が大きいと補正量も大きくなるためレンダリング画像の画質が劣化し、提供するサービスの品質が劣化する恐れがある。 In the above technique, for example, when the user moves a lot, if the difference between the predicted posture and the actual posture is large, the correction amount becomes large, so that the image quality of the rendered image deteriorates and the quality of the provided service may deteriorate. There is.
 そこで、本開示ではユーザに提供するサービスの品質劣化をより抑制することができる情報処理装置、情報処理システム及び情報処理方法を提案する。 Therefore, in this disclosure, we propose an information processing device, an information processing system, and an information processing method that can further suppress the deterioration of the quality of the service provided to the user.
 なお、上記課題又は目的は、本明細書に開示される複数の実施形態が解決し得、又は達成し得る複数の課題又は目的の1つに過ぎない。 It should be noted that the above-mentioned problem or purpose is only one of a plurality of problems or purposes that can be solved or achieved by the plurality of embodiments disclosed in the present specification.
 本開示によれば、情報処理装置が提供される。情報処理装置は、生成部と、送信部と、を備える。生成部は、第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得する。生成部は、前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成する。生成部は、前記第1のレンダリング画像及び前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成する。送信部は、前記視線情報及び前記第2のレンダリング画像を端末装置に送信する。 According to the present disclosure, an information processing device is provided. The information processing device includes a generation unit and a transmission unit. The generation unit acquires the first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image. The generation unit generates line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number. The generation unit generates the second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information. The transmission unit transmits the line-of-sight information and the second rendered image to the terminal device.
本開示の提案技術の概要を説明するための図である。It is a figure for demonstrating the outline of the proposed technique of this disclosure. 本開示の実施形態に係る情報処理システムの構成の一例を示す図である。It is a figure which shows an example of the structure of the information processing system which concerns on embodiment of this disclosure. 本開示の実施形態に係る多視点変換部の構成例を示すブロック図である。It is a block diagram which shows the structural example of the multi-viewpoint conversion part which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点バッファ制御部が決定する視点情報の一例を説明するための図である。It is a figure for demonstrating an example of viewpoint information determined by the viewpoint buffer control unit which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点バッファ制御部が決定する視点情報の一例を説明するための図である。It is a figure for demonstrating an example of viewpoint information determined by the viewpoint buffer control unit which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点バッファ制御部が決定する視点情報の一例を説明するための図である。It is a figure for demonstrating an example of viewpoint information determined by the viewpoint buffer control unit which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点バッファ制御部が決定する視点情報の一例を説明するための図である。It is a figure for demonstrating an example of viewpoint information determined by the viewpoint buffer control unit which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点情報の一例について説明するための図である。It is a figure for demonstrating an example of viewpoint information which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点バッファ制御部が決定する変換視点情報の一例を示す図である。It is a figure which shows an example of the conversion viewpoint information determined by the viewpoint buffer control unit which concerns on embodiment of this disclosure. 本開示の実施形態に係る視点バッファ制御部の構成例を示すブロック図である。It is a block diagram which shows the structural example of the viewpoint buffer control part which concerns on embodiment of this disclosure. 本開示の実施形態に係るデコード部の構成例を示すブロック図である。It is a block diagram which shows the structural example of the decoding part which concerns on embodiment of this disclosure. 本開示の実施形態に係るデコード制御部が決定するデコード視点情報について説明するための図である。It is a figure for demonstrating the decoding viewpoint information determined by the decoding control unit which concerns on embodiment of this disclosure. 本開示の実施形態に係るサーバ装置が実行する多視点変換処理の流れを示すフローチャートである。It is a flowchart which shows the flow of the multi-viewpoint conversion process executed by the server apparatus which concerns on embodiment of this disclosure. 本開示の実施形態に係るクライアント装置が実行する表示画像生成処理の流れを示すフローチャートである。It is a flowchart which shows the flow of the display image generation processing executed by the client apparatus which concerns on embodiment of this disclosure.
 以下に添付図面を参照しながら、本開示の実施形態について詳細に説明する。なお、本明細書及び図面において、実質的に同一の機能構成を有する構成要素については、同一の符号を付することにより重複説明を省略する。 The embodiments of the present disclosure will be described in detail with reference to the accompanying drawings below. In the present specification and the drawings, components having substantially the same functional configuration are designated by the same reference numerals, so that duplicate description will be omitted.
 また、本明細書及び図面において、実施形態の類似する構成要素については、同一の符号の後に異なるアルファベットを付して区別する場合がある。ただし、類似する構成要素の各々を特に区別する必要がない場合、同一符号のみを付する。 Further, in the present specification and the drawings, similar components of the embodiment may be distinguished by adding different alphabets after the same reference numerals. However, if it is not necessary to distinguish each of the similar components, only the same reference numerals are given.
 以下に説明される1又は複数の実施形態(実施例、変形例を含む)は、各々が独立に実施されることが可能である。一方で、以下に説明される複数の実施形態は少なくとも一部が他の実施形態の少なくとも一部と適宜組み合わせて実施されてもよい。これら複数の実施形態は、互いに異なる新規な特徴を含み得る。したがって、これら複数の実施形態は、互いに異なる目的又は課題を解決することに寄与し得、互いに異なる効果を奏し得る。 Each of one or more embodiments (including examples and modifications) described below can be implemented independently. On the other hand, at least a part of the plurality of embodiments described below may be carried out in combination with at least a part of other embodiments as appropriate. These plurality of embodiments may contain novel features that differ from each other. Therefore, these plurality of embodiments may contribute to solving different purposes or problems, and may have different effects.
 なお、説明は以下の順序で行うものとする。
  1.はじめに
   1.1.提案技術の概要
  2.情報処理システムの構成例
   2.1.情報処理システム
   2.2.サーバ装置
   2.3.クライアント装置
  3.情報処理
   3.1.多視点変換処理
   3.2.サーバ装置
  4.その他の実施形態
  5.まとめ
The explanations will be given in the following order.
1. 1. Introduction 1.1. Outline of the proposed technology 2. Configuration example of information processing system 2.1. Information processing system 2.2. Server device 2.3. Client device 3. Information processing 3.1. Multi-viewpoint conversion process 3.2. Server device 4. Other embodiments 5. summary
 <<1.はじめに>>
 <1.1.提案技術の概要>
 まず、本開示に係る提案技術の概要について説明する。図1は、本開示の提案技術の概要を説明するための図である。本開示に係る提案技術は、図1に示す情報処理システムで実施される。図1に示すように、情報処理システム10は、サーバ装置100と、クライアント装置200と、を有する。サーバ装置100及びクライアント装置200は、例えばインターネットのようなネットワークNを介して接続される。
<< 1. Introduction >>
<1.1. Outline of proposed technology >
First, the outline of the proposed technique according to the present disclosure will be described. FIG. 1 is a diagram for explaining an outline of the proposed technique of the present disclosure. The proposed technique according to the present disclosure is implemented in the information processing system shown in FIG. As shown in FIG. 1, the information processing system 10 includes a server device 100 and a client device 200. The server device 100 and the client device 200 are connected via a network N such as the Internet.
 サーバ装置100は、例えばレンダリングサーバであり、クライアント装置200にレンダリングした画像(以下、レンダリング画像とも称する)を送信する。例えば、サーバ装置100は、ゲームサーバであり、クライアント装置200にゲーム画像であるレンダリング画像を送信することで、クライアント装置200にゲームサービスを提供する。あるいは、サーバ装置100は、ARやVRを、クライアント装置200を介してユーザに提供する装置であってもよい。この場合、サーバ装置100は、例えばユーザの視点に応じたレンダリング画像を生成し、クライアント装置200に送信する。 The server device 100 is, for example, a rendering server, and transmits a rendered image (hereinafter, also referred to as a rendered image) to the client device 200. For example, the server device 100 is a game server, and provides a game service to the client device 200 by transmitting a rendered image which is a game image to the client device 200. Alternatively, the server device 100 may be a device that provides AR or VR to the user via the client device 200. In this case, the server device 100 generates, for example, a rendered image according to the user's viewpoint and transmits it to the client device 200.
 クライアント装置200は、例えばスマートフォンやPC、あるいはゲーム機器等の情報処理装置である。また、クライアント装置200は、ヘッドマウントディスプレイ(HMD:Head Mounted Display)やスマートグラスのようにユーザの頭部に装着する情報処理装置であってもよい。クライアント装置200は、サーバ装置100から取得したレンダリング画像を、自装置の表示部又は自装置と接続する表示装置に表示する。 The client device 200 is an information processing device such as a smartphone, a PC, or a game device. Further, the client device 200 may be an information processing device worn on the user's head, such as a head-mounted display (HMD: Head Mounted Display) or smart glasses. The client device 200 displays the rendered image acquired from the server device 100 on the display unit of the own device or the display device connected to the own device.
 サーバ装置100がネットワークNを介してレンダリング画像をクライアント装置200に送信する場合、ネットワークNで発生する通信遅延により、クライアント装置200でのレンダリング画像の表示に遅れが生じる場合がある。例えば情報処理システム10がゲームサービスを提供するゲームストリーミングの場合、かかる遅れが、ユーザに与える良好なゲーム体験の妨げとなる恐れがある。特に、仮想カメラをユーザが制御する類のゲームストリーミングや、AR、VR技術を用いたゲームストリーミングなど、視点移動を伴うゲームでは、通信遅延が発生している間にユーザの視点が移動してしまい、表示画像の視点と実際のユーザの視点とに大きなずれが発生する恐れがある。VRでは表示画像の視点とユーザの視点とのずれによってユーザがVR酔いを感じてしまう恐れがある。 When the server device 100 transmits a rendered image to the client device 200 via the network N, the display of the rendered image on the client device 200 may be delayed due to the communication delay generated in the network N. For example, in the case of game streaming in which the information processing system 10 provides a game service, such a delay may hinder a good game experience given to the user. In particular, in games involving viewpoint movement such as game streaming in which the user controls the virtual camera and game streaming using AR and VR technologies, the user's viewpoint moves while the communication delay occurs. , There is a possibility that a large difference may occur between the viewpoint of the displayed image and the viewpoint of the actual user. In VR, the user may feel VR sickness due to the deviation between the viewpoint of the displayed image and the viewpoint of the user.
 このように、例えばクラウド上に配置されたサーバ装置100でレンダリングした画像をクライアント装置200に送信するためには伝送時間がかかる。特に、ネットワークNが一般的なインターネット回線の場合、伝送時間はインターネット回線のトラフィックによって左右される。そのため、インターネット回線のトラフィック量によっては、例えばゲームなどユーザに提供するサービスの品質が非常に悪化する恐れがある。 As described above, for example, it takes a transmission time to transmit the image rendered by the server device 100 arranged on the cloud to the client device 200. In particular, when the network N is a general Internet line, the transmission time depends on the traffic of the Internet line. Therefore, depending on the amount of traffic on the Internet line, the quality of services provided to users such as games may be significantly deteriorated.
 そこで、本開示の技術では、サーバ装置100がレンダリング画像を多視点画像に変換し、変換した多視点画像をクライアント装置200に送信する。このとき、サーバ装置100が、レンダリング画像を生成する際に使用した情報(以下、レンダリング付帯情報とも言う)を使用して高精度に多視点画像を生成する。 Therefore, in the technique of the present disclosure, the server device 100 converts the rendered image into a multi-viewpoint image, and transmits the converted multi-viewpoint image to the client device 200. At this time, the server device 100 generates a multi-viewpoint image with high accuracy by using the information used when generating the rendered image (hereinafter, also referred to as rendering incidental information).
 クライアント装置200は、受信した多視点画像に基づき、現在のユーザの視点、あるいはユーザによるゲームの操作状況に応じて表示画像を生成し、ユーザに提示する。 The client device 200 generates a display image based on the received multi-viewpoint image according to the current user's viewpoint or the operation status of the game by the user, and presents the display image to the user.
 ここで、図1を用いて、本開示の技術に係る情報処理の概要について説明する。 Here, with reference to FIG. 1, an outline of information processing related to the technique of the present disclosure will be described.
 まず、サーバ装置100は、レンダリング画像M0及びレンダリング付帯情報を取得する(ステップS1)。ここで、レンダリング画像は、クライアント装置200から送信されるゲームの操作状況やユーザの視点情報に基づき、サーバ装置100のレンダリング部(図示省略)によって生成される画像である。また、レンダリング付帯情報は、レンダリング部がレンダリング画像を生成するために使用する情報であり、例えばデプス情報などが含まれる。レンダリング部は、レンダリング付帯情報を使用して、所定数の視点位置及び視線方向に基づくレンダリング画像を生成する。 First, the server device 100 acquires the rendered image M0 and the rendering incidental information (step S1). Here, the rendered image is an image generated by the rendering unit (not shown) of the server device 100 based on the operation status of the game and the viewpoint information of the user transmitted from the client device 200. Further, the rendering incidental information is information used by the rendering unit to generate a rendered image, and includes, for example, depth information. The rendering unit uses the rendering incidental information to generate a rendered image based on a predetermined number of viewpoint positions and line-of-sight directions.
 図1の例では、サーバ装置100は、レンダリング部が生成し得る複数の視点位置及び視線方向の画像の中から、1つの視点位置及び視線方向から見たレンダリング画像M0を取得する。図1では、レンダリング部が生成し得る複数のレンダリング画像をマトリクス状に配置して示している。図示したマトリクス状のレンダリング画像のうち、ハッチングを付したレンダリング画像は実際にはレンダリング部によって生成されない画像であり、レンダリング部は、ハッチングを付していないレンダリング画像M0を生成する。 In the example of FIG. 1, the server device 100 acquires a rendered image M0 viewed from one viewpoint position and line-of-sight direction from a plurality of images in a viewpoint position and a line-of-sight direction that can be generated by the rendering unit. In FIG. 1, a plurality of rendered images that can be generated by the rendering unit are arranged and shown in a matrix. Among the illustrated matrix-shaped rendered images, the rendered image with hatching is an image that is not actually generated by the rendering unit, and the rendering unit generates the rendered image M0 without hatching.
 次に、サーバ装置100は、視点バッファ情報を生成する(ステップS2)。視点バッファ情報は、多視点画像を生成するために使用される情報であり、複数の視点位置及び視線方向を含む。サーバ装置100は、ゲームの状況や通信遅延の状況(通信状況)等に基づき、視点バッファ情報を生成する。 Next, the server device 100 generates the viewpoint buffer information (step S2). The viewpoint buffer information is information used to generate a multi-viewpoint image, and includes a plurality of viewpoint positions and line-of-sight directions. The server device 100 generates viewpoint buffer information based on a game status, a communication delay status (communication status), and the like.
 サーバ装置100は、レンダリング画像に基づき、視点バッファ情報及びレンダリング付帯情報を使用して視点バッファを生成する(ステップS3)。視点バッファには、視点バッファ情報に含まれる複数の視点位置及び視線方向から見た複数の画像(以下、多視点画像とも言う)が含まれる。 The server device 100 generates a viewpoint buffer using the viewpoint buffer information and the rendering incidental information based on the rendered image (step S3). The viewpoint buffer includes a plurality of viewpoint positions included in the viewpoint buffer information and a plurality of images viewed from the line-of-sight direction (hereinafter, also referred to as multi-viewpoint images).
 図1では、サーバ装置100が視点変換し得る複数の画像をマトリクス状に配置して示している。図示したマトリクス状の画像のうち、ハッチングを付した画像は実際にはサーバ装置100によって変換されない画像であり、サーバ装置100は、視点変換によってハッチングを付していない画像M1~M5を生成する。 FIG. 1 shows a plurality of images that can be viewpoint-converted by the server device 100 arranged in a matrix. Among the illustrated matrix-shaped images, the hatched images are images that are not actually converted by the server device 100, and the server device 100 generates the unhatched images M1 to M5 by the viewpoint conversion.
 サーバ装置100は、レンダリング付帯情報を使用することで、高精度な多視点画像を生成することができる。 The server device 100 can generate a highly accurate multi-viewpoint image by using the rendering incidental information.
 サーバ装置100は、視点バッファ情報及び視点バッファを、ネットワークNを介してクライアント装置200に送信する(ステップS4)。 The server device 100 transmits the viewpoint buffer information and the viewpoint buffer to the client device 200 via the network N (step S4).
 クライアント装置200は、視点バッファ情報及び視点バッファをサーバ装置100から受信する(ステップS5)。 The client device 200 receives the viewpoint buffer information and the viewpoint buffer from the server device 100 (step S5).
 クライアント装置200は、現在のユーザの視点に基づき、視点バッファに含まれる多視点画像から表示画像を生成し(ステップS6)、ユーザに提示する。 The client device 200 generates a display image from the multi-viewpoint image included in the viewpoint buffer based on the current viewpoint of the user (step S6), and presents the display image to the user.
 このように、本開示の一実施形態に係る情報処理では、サーバ装置100がレンダリング画像から生成した複数の多視点画像をクライアント装置200に送信する。このとき、サーバ装置100は、レンダリング付帯情報を使用することで、高画質の多視点画像を生成することができる。また、クライアント装置200は、現在のユーザの視点に応じた表示画像を複数の多視点画像から生成する。これにより、情報処理システム10は、ネットワークNで通信遅延が発生した場合でも、現在のユーザの視点に応じた高画質の表示画像をユーザに提示することができ、通信遅延による提供サービスの劣化を抑制することができる。 As described above, in the information processing according to the embodiment of the present disclosure, the server device 100 transmits a plurality of multi-viewpoint images generated from the rendered image to the client device 200. At this time, the server device 100 can generate a high-quality multi-viewpoint image by using the rendering incidental information. Further, the client device 200 generates a display image according to the viewpoint of the current user from a plurality of multi-viewpoint images. As a result, the information processing system 10 can present the user with a high-quality display image according to the viewpoint of the current user even if a communication delay occurs in the network N, and the deterioration of the provided service due to the communication delay is deteriorated. It can be suppressed.
 なお、ここでは、本開示の技術の適用例として、情報処理システム10がゲームサービスを提供する場合について説明するが、情報処理システム10が提供するサービスはゲームサービスに限定されない。例えば、情報処理システム10がARシステムやVRシステムであってもよい。このように、情報処理システム10は、視点移動に応じた画像をユーザに提供するシステムであればよく、ゲームサービスを提供するシステムに限定されない。 Here, as an application example of the technique of the present disclosure, a case where the information system 10 provides a game service will be described, but the service provided by the information system 10 is not limited to the game service. For example, the information processing system 10 may be an AR system or a VR system. As described above, the information processing system 10 may be any system that provides the user with an image corresponding to the movement of the viewpoint, and is not limited to the system that provides the game service.
 <<2.情報処理システムの構成例>>
 <2.1.情報処理システム>
 図2は、本開示の実施形態に係る情報処理システム10の構成の一例を示す図である。情報処理システムは、サーバ装置100と、クライアント装置200と、周辺機器300と、を有する。
<< 2. Information processing system configuration example >>
<2.1. Information processing system >
FIG. 2 is a diagram showing an example of the configuration of the information processing system 10 according to the embodiment of the present disclosure. The information processing system includes a server device 100, a client device 200, and a peripheral device 300.
 本開示の実施形態に係る情報処理システムでは、サーバ装置100は、レンダリングしたゲームの映像(例えばレンダリング画像)を配信する。クライアント装置200は、インターネットなどの通信回線であるネットワークNを通じて、ゲームの映像を受信して、例えば周辺機器300の表示部(図示省略)に再生する。なお、周辺機器300は、例えばディスプレイやHMD、コントローラ等が含まれる。 In the information system according to the embodiment of the present disclosure, the server device 100 distributes a rendered game image (for example, a rendered image). The client device 200 receives the video of the game through the network N, which is a communication line such as the Internet, and reproduces it, for example, on the display unit (not shown) of the peripheral device 300. The peripheral device 300 includes, for example, a display, an HMD, a controller, and the like.
 クライアント装置200は、再生した映像に反応してユーザが行う操作や動作を例えばコマンド情報として検出する。ユーザがコントローラを操作する場合、クライアント装置200は、コントローラに対する操作をコマンド情報として検出する。また、ユーザが例えばHMDを装着している場合、クライアント装置200は、ユーザの頭部の動きをHMDの動きとして検出する。クライアント装置200は、検出したコマンド情報を、ネットワークNを介してサーバ装置100に送信する。 The client device 200 detects, for example, command information an operation or operation performed by the user in response to the reproduced video. When the user operates the controller, the client device 200 detects the operation on the controller as command information. Further, when the user wears the HMD, for example, the client device 200 detects the movement of the user's head as the movement of the HMD. The client device 200 transmits the detected command information to the server device 100 via the network N.
 サーバ装置100は、受信したコマンド情報に応じて次の時刻のゲームの画像をレンダリングしてクライアント装置200に配信する。 The server device 100 renders an image of the game at the next time according to the received command information and distributes it to the client device 200.
 [サーバ装置100]
 図2に示すように、サーバ装置100は、クライアント装置200にアプリケーション機能を提供する情報処理装置である。サーバ装置100は、例えば、アプリケーション機能としてゲームサービスを提供するゲームサーバ(Game Server)であり、クライアント装置200に対してゲーム画面として表示するレンダリング画像を送信する。また、サーバ装置100は、コマンド情報をクライアント装置200から受信する。かかるコマンド情報には、ユーザの頭部の動きに関する情報など、ユーザの視点を特定し得る情報や、ユーザによるコントローラに対する操作情報が含まれる。
[Server device 100]
As shown in FIG. 2, the server device 100 is an information processing device that provides an application function to the client device 200. The server device 100 is, for example, a game server (Game Server) that provides a game service as an application function, and transmits a rendered image to be displayed as a game screen to the client device 200. Further, the server device 100 receives the command information from the client device 200. Such command information includes information that can identify the user's viewpoint, such as information on the movement of the user's head, and information on operations by the user with respect to the controller.
 サーバ装置100は、レンダリング部110と、多視点変換部120と、エンコード部130と、送受信部140と、コマンド情報取得部150と、を備える。 The server device 100 includes a rendering unit 110, a multi-viewpoint conversion unit 120, an encoding unit 130, a transmission / reception unit 140, and a command information acquisition unit 150.
 (レンダリング部110)
 レンダリング部110は、ゲーム画像のレンダリングを行う。レンダリング部110は、例えば、クライアント装置200から取得したコマンド情報に基づいて、ゲーム内のカメラ位置(例えば、ユーザの視点位置)やゲームの進行状況などを判定し、判定結果に応じたゲーム画像を生成する。このように、レンダリング部110は、ゲームの進行を制御するGame Logicの機能を有し、ゲームサービスの提供を行うアプリケーションである。レンダリング部110は、n1個の視点位置及び視線方向(以下、単に視点、あるいは、視点情報とも記載する)のレンダリング画像を生成する。
(Rendering unit 110)
The rendering unit 110 renders the game image. The rendering unit 110 determines, for example, the camera position in the game (for example, the user's viewpoint position), the progress of the game, and the like based on the command information acquired from the client device 200, and produces a game image according to the determination result. Generate. As described above, the rendering unit 110 is an application that has a Game Logic function for controlling the progress of the game and provides a game service. The rendering unit 110 generates n1 viewpoint positions and line-of-sight directions (hereinafter, also simply referred to as viewpoints or viewpoint information) rendered images.
 一般的に、レンダリング部110は、計算量等の制約から1視点(n1=1)など少数のレンダリング画像を生成する。以下、説明を簡略化するために、レンダリング部110は1個の視点から見たレンダリング画像を生成するものとする。換言すると、レンダリング部110は、コマンド情報に基づき、レンダリング付帯情報を使用してレンダリング画像を生成する。レンダリング付帯情報には、例えば、ゲーム内に含まれるオブジェクトのデプス情報や法線情報等が含まれる。なお、レンダリング付帯情報の詳細について後述する。また、レンダリング部110がレンダリング画像を生成した時刻をレンダリング時刻t1と称する。 Generally, the rendering unit 110 generates a small number of rendered images such as one viewpoint (n1 = 1) due to restrictions such as the amount of calculation. Hereinafter, in order to simplify the explanation, the rendering unit 110 shall generate a rendered image viewed from one viewpoint. In other words, the rendering unit 110 generates a rendered image using the rendering incidental information based on the command information. The rendering incidental information includes, for example, depth information and normal information of objects included in the game. The details of the rendering incidental information will be described later. Further, the time when the rendering unit 110 generates the rendered image is referred to as a rendering time t1.
 レンダリング部110は、レンダリング画像、レンダリング付帯情報、及び、ゲームに関する情報を多視点変換部120に出力する。ここで、ゲームに関する情報には、例えば、ゲームのタイトル、あるいは、レースゲーム、シミュレーションゲーム、RPGゲーム等ゲームの種別に関する情報や、ゲームのプレイ状況などの情報が含まれる。 The rendering unit 110 outputs the rendered image, the rendering incidental information, and the information about the game to the multi-viewpoint conversion unit 120. Here, the information about the game includes, for example, information about the title of the game, information about the type of the game such as a race game, a simulation game, and an RPG game, and information such as a game play status.
 (多視点変換部120)
 多視点変換部120は、レンダリング部110が生成するレンダリング画像から複数の視点変換画像を含む視点バッファを生成する。なお、複数の視点変換画像をまとめて多視点画像とも称する。
(Multi-viewpoint conversion unit 120)
The multi-viewpoint conversion unit 120 generates a viewpoint buffer including a plurality of viewpoint conversion images from the rendered image generated by the rendering unit 110. In addition, a plurality of viewpoint conversion images are collectively referred to as a multi-viewpoint image.
 多視点変換部120は、レンダリング部110から取得したレンダリング画像、レンダリング付帯情報、及び、ゲームに関する情報に加え、送受信部140からネットワークトラフィック(遅延)状況に関する情報を取得する。多視点変換部120は、取得した情報から、視点バッファ情報を生成し、レンダリング画像をn2個の視点からみた多視点画像に変換する。なお、多視点画像の視点数n2は、レンダリング画像の視点数n1より多い(n1<n2)ものとする。また、視点バッファ情報には、視点バッファに含まれる多視点画像の視点の座標情報が含まれる。多視点変換部120が生成する視点バッファ及び視点バッファ情報の詳細については後述する。 The multi-viewpoint conversion unit 120 acquires information on the network traffic (delay) status from the transmission / reception unit 140 in addition to the rendered image acquired from the rendering unit 110, rendering incidental information, and information on the game. The multi-viewpoint conversion unit 120 generates viewpoint buffer information from the acquired information and converts the rendered image into a multi-viewpoint image viewed from n2 viewpoints. It is assumed that the number of viewpoints n2 of the multi-viewpoint image is larger than the number of viewpoints n1 of the rendered image (n1 <n2). Further, the viewpoint buffer information includes the coordinate information of the viewpoint of the multi-view image included in the viewpoint buffer. The details of the viewpoint buffer and the viewpoint buffer information generated by the multi-viewpoint conversion unit 120 will be described later.
 多視点変換部120は、生成した視点バッファ及び視点バッファ情報をエンコード部130に出力する。 The multi-viewpoint conversion unit 120 outputs the generated viewpoint buffer and viewpoint buffer information to the encoding unit 130.
 ここで、レンダリング画像がネットワークNを介してユーザの元で再生されるまでには、ネットワークNのトラフィックに応じて相応の遅延が発生する。理想的には、ユーザによって時々刻々クライアント装置200に入力されるコマンド情報を即座に反映して、適切な視点でゲーム画像をレンダリングしてユーザに提示することが望まれる。しかしながら、上述した遅延によって、レンダリング画像が生成されるレンダリング時刻t1と、コマンド情報が入力される時刻(以下、コマンド生成時刻t2とも言う)との間にGAPが発生する。そのため、ユーザが期待する映像とサーバ装置100がレンダリングする映像とでは、視点が異なる場合がある。 Here, there is a corresponding delay depending on the traffic of the network N until the rendered image is reproduced by the user via the network N. Ideally, it is desired that the command information input to the client device 200 by the user from moment to moment is immediately reflected, and the game image is rendered and presented to the user from an appropriate viewpoint. However, due to the delay described above, GAP is generated between the rendering time t1 at which the rendered image is generated and the time at which the command information is input (hereinafter, also referred to as command generation time t2). Therefore, the viewpoint may be different between the image expected by the user and the image rendered by the server device 100.
 そこで、多視点変換部120は、レンダリング部110がレンダリングしたレンダリング画像に基づき、レンダリング画像の視点(以下、レンダリング視点とも言う)周辺の視点から見た画像を多視点画像としてあらかじめ生成しておく。換言すると、多視点変換部120は、複数視点の画像のバッファを視点バッファとしてあらかじめ構築しておく。 Therefore, the multi-viewpoint conversion unit 120 generates in advance an image viewed from a viewpoint around the viewpoint of the rendered image (hereinafter, also referred to as a rendering viewpoint) as a multi-viewpoint image based on the rendered image rendered by the rendering unit 110. In other words, the multi-viewpoint conversion unit 120 builds a buffer for images of a plurality of viewpoints as a viewpoint buffer in advance.
 サーバ装置100は、このような視点バッファを有するゲーム画像をクライアント装置200に送信する。上述したように、レンダリング時刻t1とコマンド時刻t2との間のGAPにより、ゲーム画像の視点とユーザの期待する画像の視点とにずれが生じたとする。 The server device 100 transmits a game image having such a viewpoint buffer to the client device 200. As described above, it is assumed that the GAP between the rendering time t1 and the command time t2 causes a deviation between the viewpoint of the game image and the viewpoint of the image expected by the user.
 この場合であっても、当該ずれが視点バッファの範囲に収まっていれば、クライアント装置200は、視点バッファを用いてずれを補間したゲーム画像を生成することができる。 Even in this case, if the deviation is within the range of the viewpoint buffer, the client device 200 can generate a game image in which the deviation is interpolated using the viewpoint buffer.
 また、当該ずれが大きく、視点バッファの範囲から外れてしまった場合でも、クライアント装置200は、視点バッファに含まれる視点であって、ユーザが期待する画像の視点に近い視点の視点変換画像に基づいて視点変換を行うことでゲーム画像を生成することができる。このように、ずれが大きくても、視点バッファに含まれる視点変換画像からゲーム画像を生成することで、レンダリング画像からゲーム画像を生成する場合に比べて、視点のずれを小さくすることができ、ゲーム画面の品質の劣化をより抑制することができる。 Further, even if the deviation is large and deviates from the range of the viewpoint buffer, the client device 200 is based on the viewpoint conversion image of the viewpoint included in the viewpoint buffer and close to the viewpoint of the image expected by the user. A game image can be generated by performing viewpoint conversion. In this way, even if the deviation is large, by generating the game image from the viewpoint conversion image included in the viewpoint buffer, the deviation of the viewpoint can be reduced as compared with the case where the game image is generated from the rendered image. Deterioration of the quality of the game screen can be further suppressed.
 (エンコード部130)
 エンコード部130は、多視点変換部120が生成した視点バッファ及び視点バッファ情報を符号化することで小容量に圧縮し、圧縮したビットストリーム(以下、圧縮データとも記載する)を送受信部140に出力する。
(Encoding unit 130)
The encoding unit 130 compresses the viewpoint buffer and the viewpoint buffer information generated by the multi-viewpoint conversion unit 120 into a small capacity, and outputs the compressed bit stream (hereinafter, also referred to as compressed data) to the transmission / reception unit 140. do.
 (送受信部140)
 送受信部140は、外部装置との通信を行う通信インタフェース(I/F)である。送受信部140は、例えば、NIC(Network Interface Card)等によって実現される。例えば、送受信部140は、エンコード部130が符号化した圧縮データを送信信号に変換してクライアント装置200に送信する。また、送受信部140は、クライアント装置200からコマンド情報を受信し、受信結果を視点情報取得部150に通知する。
(Transmission / reception unit 140)
The transmission / reception unit 140 is a communication interface (I / F) that communicates with an external device. The transmission / reception unit 140 is realized by, for example, a NIC (Network Interface Card) or the like. For example, the transmission / reception unit 140 converts the compressed data encoded by the encoding unit 130 into a transmission signal and transmits it to the client device 200. Further, the transmission / reception unit 140 receives command information from the client device 200 and notifies the viewpoint information acquisition unit 150 of the reception result.
 (コマンド情報取得部150)
 コマンド情報取得部150は、送受信部140を介してコマンド情報を取得する。コマンド情報取得部150は、取得したコマンド情報をレンダリング部110に出力する。
(Command information acquisition unit 150)
The command information acquisition unit 150 acquires command information via the transmission / reception unit 140. The command information acquisition unit 150 outputs the acquired command information to the rendering unit 110.
 [クライアント装置200]
 クライアント装置200は、サーバ装置100からアプリケーション機能の提供を受ける情報処理装置である。クライアント装置200は、例えばアプリケーション機能としてゲームサービスの提供を受ける。
[Client device 200]
The client device 200 is an information processing device that receives application functions from the server device 100. The client device 200 receives, for example, a game service as an application function.
 クライアント装置200は、サーバ装置100から送信されるゲームに関する画像(視点バッファ)を受信し、受信した視点バッファに基づいてゲーム画像を表示する。クライアント装置200は、ユーザによるゲームに対する操作や動作に応じたコマンド情報をサーバ装置100に送信する。 The client device 200 receives an image (viewpoint buffer) related to the game transmitted from the server device 100, and displays the game image based on the received viewpoint buffer. The client device 200 transmits command information according to the operation or operation of the game by the user to the server device 100.
 クライアント装置200は、送受信部210と、デコード部220と、入力情報取得部230と、を有する。 The client device 200 has a transmission / reception unit 210, a decoding unit 220, and an input information acquisition unit 230.
 (送受信部210)
 送受信部210は、外部装置との通信を行う通信インタフェース(I/F)である。送受信部210は、例えば、NIC(Network Interface Card)等によって実現される。例えば、送受信部210は、サーバ装置100から送信信号を受信信号として受信し、受信した受信信号を圧縮データに変換する。また、送受信部210は、サーバ装置100にコマンド情報を送信する。
(Transmission / reception unit 210)
The transmission / reception unit 210 is a communication interface (I / F) that communicates with an external device. The transmission / reception unit 210 is realized by, for example, a NIC (Network Interface Card) or the like. For example, the transmission / reception unit 210 receives a transmission signal from the server device 100 as a reception signal, and converts the received reception signal into compressed data. Further, the transmission / reception unit 210 transmits command information to the server device 100.
 (デコード部220)
 デコード部220は、送受信部210が受信した圧縮データを復号する。デコード部220は、入力情報取得部230が取得するユーザの視点情報と視点バッファ情報とに基づき、視点バッファに含まれる多視点画像からユーザに提示するゲーム画像を生成する。なお、入力情報取得部230が取得するユーザの視点情報は、現在のゲーム内での最新の視点情報であり、すなわち、ユーザが期待するゲーム画像の視点に関する情報であり、以下、単に「現在のユーザの視点情報」とも称する。
(Decoding unit 220)
The decoding unit 220 decodes the compressed data received by the transmission / reception unit 210. The decoding unit 220 generates a game image to be presented to the user from the multi-viewpoint image included in the viewpoint buffer based on the user's viewpoint information and the viewpoint buffer information acquired by the input information acquisition unit 230. The user's viewpoint information acquired by the input information acquisition unit 230 is the latest viewpoint information in the current game, that is, information regarding the viewpoint of the game image expected by the user, and is simply "current". Also referred to as "user's viewpoint information".
 デコード部220は、生成したゲーム画像を、周辺機器300の表示部(図示省略)に表示する。 The decoding unit 220 displays the generated game image on the display unit (not shown) of the peripheral device 300.
 (入力情報取得部230)
 入力情報取得部230は、ユーザによる周辺機器300の操作を受け付けるユーザインタフェース(UI:User Interface)である。入力情報取得部230は、周辺機器300に表示されたゲーム画像に反応する形でユーザが行う操作や動作を、サーバ装置100が解釈し得るコマンド情報に変換する。また、入力情報取得部230は、周辺機器300に表示するゲーム画像の視点情報(現在のユーザの視点情報)を生成する。入力情報取得部230は、生成した視点情報をデコード部220に出力する。また、入力情報取得部230は、視点情報をコマンド情報に含めて、送受信部210を介してサーバ装置100に送信する。
(Input information acquisition unit 230)
The input information acquisition unit 230 is a user interface (UI: User Interface) that accepts operations of the peripheral device 300 by the user. The input information acquisition unit 230 converts the operation or operation performed by the user in response to the game image displayed on the peripheral device 300 into command information that can be interpreted by the server device 100. Further, the input information acquisition unit 230 generates viewpoint information (viewpoint information of the current user) of the game image to be displayed on the peripheral device 300. The input information acquisition unit 230 outputs the generated viewpoint information to the decoding unit 220. Further, the input information acquisition unit 230 includes the viewpoint information in the command information and transmits the viewpoint information to the server device 100 via the transmission / reception unit 210.
 [周辺機器300]
 周辺機器300は、例えばディスプレイや、スマートグラス、HMD、コントローラ等、クライアント装置200が出力するゲーム画像を表示したり、ユーザによる操作を受け付けたりする装置である。周辺機器300がディスプレイの場合、周辺機器300は、ゲーム画像をユーザに提示する。また、周辺機器300がコントローラの場合、周辺機器300は、表示装置(図示省略)に表示されるゲーム画像に反応してユーザが行う操作を受け付ける。また、周辺機器300がHMDである場合、周辺機器300は、表示部(図示省略)にゲーム画像を表示するとともに、ユーザの頭部の動きをHMDの動きとして検出する。
[Peripheral device 300]
The peripheral device 300 is a device such as a display, smart glasses, HMD, controller, etc. that displays a game image output by the client device 200 and accepts operations by the user. When the peripheral device 300 is a display, the peripheral device 300 presents a game image to the user. When the peripheral device 300 is a controller, the peripheral device 300 accepts an operation performed by the user in response to a game image displayed on a display device (not shown). When the peripheral device 300 is an HMD, the peripheral device 300 displays a game image on a display unit (not shown) and detects the movement of the user's head as the movement of the HMD.
 なお、ここでは、周辺機器300がクライアント装置200とは異なる装置であるとしたが、これに限定されない。例えば、周辺機器300がクライアント装置200の機能として実現されてもよい。 Here, it is assumed that the peripheral device 300 is a device different from the client device 200, but the present invention is not limited to this. For example, the peripheral device 300 may be realized as a function of the client device 200.
 上述したように、本開示の実施形態では、サーバ装置100がレンダリング付帯情報を使用して視点バッファを生成する。また、クライアント装置200が視点バッファに基づき、現在のユーザの視点情報に応じたゲーム画像(表示画像)を生成する。これにより、情報処理システム10は、ネットワークNの遅延によらず、現在のユーザの視点から見たゲーム画像を表示することができる。 As described above, in the embodiment of the present disclosure, the server device 100 uses the rendering incidental information to generate the viewpoint buffer. Further, the client device 200 generates a game image (display image) according to the viewpoint information of the current user based on the viewpoint buffer. As a result, the information processing system 10 can display the game image as seen from the viewpoint of the current user regardless of the delay of the network N.
 このように、情報処理システム10では、視点バッファをサーバ装置100が生成する。換言すると、サーバ装置100がレンダリング付帯情報を使用してレンダリング画像の視点変換を行う。このように、レンダリング画像のみから視点変換を行うより、レンダリング付帯情報を使用してレンダリング画像の視点変換を行う方が、より高画質な視点変換画像が得られる。 In this way, in the information processing system 10, the viewpoint buffer is generated by the server device 100. In other words, the server device 100 uses the rendering incidental information to perform viewpoint conversion of the rendered image. As described above, it is possible to obtain a higher image quality viewpoint conversion image by performing the viewpoint conversion of the rendered image using the rendering incidental information than by performing the viewpoint conversion only from the rendered image.
 ここで、クライアント装置200が現在のユーザの視点情報に基づき、レンダリング画像の視点変換を行うことも考えられる。 Here, it is conceivable that the client device 200 performs viewpoint conversion of the rendered image based on the viewpoint information of the current user.
 しかしながら、この場合、クライアント装置200は、レンダリング付帯情報を使用した視点変換を行うことが難しい。これは、レンダリング付帯情報は、多視点画像と比較して符号化効率化が劣るため、通信量が大きくなってしまうからである。後述するように、レンダリング付帯情報には、画像情報だけでなく、ゲーム内のオブジェクトに関する情報など種々の情報が含まれる。そのため、レンダリング付帯情報は、視点が異なる画像情報で構成される視点バッファと比較して、符号化効率が悪くなってしまう。 However, in this case, it is difficult for the client device 200 to perform viewpoint conversion using rendering incidental information. This is because the rendering incidental information is inferior in coding efficiency as compared with the multi-viewpoint image, so that the communication amount becomes large. As will be described later, the rendering incidental information includes not only image information but also various information such as information about objects in the game. Therefore, the rendering incidental information has a lower coding efficiency than the viewpoint buffer composed of image information having different viewpoints.
 そのため、本開示の実施形態に係る情報処理システム10は、クライアント装置200でレンダリング画像の視点変換を行う場合と比較して、ネットワークトラフィックの増加を抑制しつつ、視点移動による画像遅延をより低減させることができる。このように、本開示の実施形態に係る情報処理システム10は、より高品質なサービスを提供することができる。 Therefore, the information processing system 10 according to the embodiment of the present disclosure further reduces the image delay due to the viewpoint movement while suppressing the increase in network traffic, as compared with the case where the viewpoint conversion of the rendered image is performed by the client device 200. be able to. As described above, the information processing system 10 according to the embodiment of the present disclosure can provide a higher quality service.
 <2.2.サーバ装置>
 次に、サーバ装置100の各部の詳細について説明する。
<2.2. Server device>
Next, the details of each part of the server device 100 will be described.
 [多視点変換部120]
 図3は、本開示の実施形態に係る多視点変換部120の構成例を示すブロック図である。図3に示すように、多視点変換部120は、視点バッファ制御部121と、多視点生成部122と、を備える。
[Multi-viewpoint conversion unit 120]
FIG. 3 is a block diagram showing a configuration example of the multi-viewpoint conversion unit 120 according to the embodiment of the present disclosure. As shown in FIG. 3, the multi-viewpoint conversion unit 120 includes a viewpoint buffer control unit 121 and a multi-viewpoint generation unit 122.
 (視点バッファ制御部121)
 視点バッファ制御部121は、多視点生成部122で生成する多視点画像の視点情報を決定する。視点バッファ制御部121は、レンダリング部110が生成するレンダリング画像の視点情報、ゲームタイトルやプレイ状況などのゲームに関する情報、及び、ネットワークNのトラフィック情報(通信状況)の少なくとも1つに基づき、視点情報を決定する。
(Viewpoint buffer control unit 121)
The viewpoint buffer control unit 121 determines the viewpoint information of the multi-view image generated by the multi-view generation unit 122. The viewpoint buffer control unit 121 is based on at least one of the viewpoint information of the rendered image generated by the rendering unit 110, the information about the game such as the game title and the play status, and the traffic information (communication status) of the network N. To decide.
 図4~図7を用いて、視点バッファ制御部121が決定する視点情報の一例について説明する。図4~図7は、本開示の実施形態に係る視点バッファ制御部121が決定する視点情報の一例を説明するための図である。 An example of viewpoint information determined by the viewpoint buffer control unit 121 will be described with reference to FIGS. 4 to 7. 4 to 7 are diagrams for explaining an example of viewpoint information determined by the viewpoint buffer control unit 121 according to the embodiment of the present disclosure.
 なお、図4~図7では、レンダリング部110が生成し得る複数のレンダリング画像、及び、多視点生成部122で生成し得る多視点画像の一例をマトリクス状に並べて示している。 Note that FIGS. 4 to 7 show a plurality of rendered images that can be generated by the rendering unit 110 and examples of the multi-viewpoint images that can be generated by the multi-viewpoint generation unit 122 side by side in a matrix.
 視点バッファ制御部121は、レンダリング部110が生成したレンダリング画像の視点情報(以下、レンダリング視点情報とも言う)を含む所定範囲の中から多視点生成部122で生成する多視点画像の視点情報を決定する。 The viewpoint buffer control unit 121 determines the viewpoint information of the multi-view image generated by the multi-view generation unit 122 from a predetermined range including the viewpoint information of the rendered image generated by the rendering unit 110 (hereinafter, also referred to as rendering viewpoint information). do.
 図4の例では、視点バッファ制御部121は、レンダリング視点情報を中心とした円形の範囲の中から十字で示す視点情報を、視点変換画像の視点情報(以下、変換視点情報とも言う)に決定する。図4では、視点バッファ制御部121が、12個の変換視点情報を決定する例を示している。 In the example of FIG. 4, the viewpoint buffer control unit 121 determines the viewpoint information indicated by a cross from the circular range centered on the rendering viewpoint information as the viewpoint information of the viewpoint converted image (hereinafter, also referred to as converted viewpoint information). do. FIG. 4 shows an example in which the viewpoint buffer control unit 121 determines 12 conversion viewpoint information.
 なお、視点バッファ制御部121は、ゲームやネットワークNの状況に応じて変換視点情報を決定する。 The viewpoint buffer control unit 121 determines the conversion viewpoint information according to the situation of the game or the network N.
 例えば、視点バッファ制御部121は、ネットワークNの遅延状況(通信状況)に応じて、変換視点情報を決定する。より具体的には、視点バッファ制御部121は、ネットワークNの遅延が大きいほど、より大きい所定範囲の中から変換視点情報を決定する。 For example, the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the delay status (communication status) of the network N. More specifically, the viewpoint buffer control unit 121 determines the conversion viewpoint information from a larger predetermined range as the delay of the network N increases.
 視点バッファ制御部121は、ネットワークNの遅延が小さい場合、例えば図4に示す範囲R1から12個の変換視点情報を決定する。また、ネットワークNの遅延が大きい場合、例えば図5に示すように、範囲R1(図4参照)より広い範囲R2の中から12個の変換視点情報を決定する。 When the delay of the network N is small, the viewpoint buffer control unit 121 determines, for example, 12 conversion viewpoint information from the range R1 shown in FIG. Further, when the delay of the network N is large, as shown in FIG. 5, for example, 12 conversion viewpoint information is determined from the range R2 wider than the range R1 (see FIG. 4).
 このように、視点バッファ制御部121は、ネットワークNの遅延に応じて変換視点情報を決定する。これにより、ネットワークNの遅延が大きい場合でも、クライアント装置200側で現在のユーザの視点から見たゲーム画像を生成することができ、ユーザに提供するサービスの品質劣化をより抑制することができる。 In this way, the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the delay of the network N. As a result, even when the delay of the network N is large, the game image seen from the viewpoint of the current user can be generated on the client device 200 side, and the deterioration of the quality of the service provided to the user can be further suppressed.
 図4及び図5では、視点バッファ制御部121は、範囲R1、R2の中から均等に変換視点情報を決定する。そのため、視点バッファ制御部121は、ネットワークNの遅延が小さいほど、密に変換視点情報を決定し、遅延が大きいほど、疎に変換視点情報を決定する。 In FIGS. 4 and 5, the viewpoint buffer control unit 121 evenly determines the conversion viewpoint information from the ranges R1 and R2. Therefore, the viewpoint buffer control unit 121 determines the conversion viewpoint information more densely as the delay of the network N is smaller, and determines the conversion viewpoint information sparsely as the delay is larger.
 ここでは、視点バッファ制御部121がネットワークNのトラフィック(遅延)状況に応じて変換視点情報を決定するとしたが、これに限定されない。視点バッファ制御部121は、ゲームタイトルやプレイ状況(シーン)などのゲームに関する情報に応じて変換視点情報を決定してもよい。 Here, the viewpoint buffer control unit 121 determines the conversion viewpoint information according to the traffic (delay) status of the network N, but the present invention is not limited to this. The viewpoint buffer control unit 121 may determine the conversion viewpoint information according to the information about the game such as the game title and the play situation (scene).
 例えば、視点バッファ制御部121は、ゲームタイトルやゲームの種別に応じて変換視点情報を決定し得る。例えば、レーシングゲームのように視点移動が速いゲームの場合、視点バッファ制御部121は、より大きい範囲の中から変換視点情報を決定し得る。一方、例えば、視点移動が少ないゲームの場合、視点バッファ制御部121は、より小さい範囲の中から変換視点情報を決定し得る。 For example, the viewpoint buffer control unit 121 can determine the conversion viewpoint information according to the game title and the type of the game. For example, in the case of a game in which the viewpoint moves quickly, such as a racing game, the viewpoint buffer control unit 121 can determine the converted viewpoint information from a larger range. On the other hand, for example, in the case of a game in which the viewpoint movement is small, the viewpoint buffer control unit 121 can determine the converted viewpoint information from a smaller range.
 あるいは、視点バッファ制御部121がプレイ状況に応じて変換視点情報を決定してもよい。例えば、視点バッファ制御部121は、屋外など移動範囲が広い空間でプレイしている場合はより大きい範囲の中から変換視点情報を決定し、屋内など移動範囲が狭い空間でプレイしている場合はより狭い範囲の中から変換視点情報を決定するようにしてもよい。 Alternatively, the viewpoint buffer control unit 121 may determine the conversion viewpoint information according to the play situation. For example, the viewpoint buffer control unit 121 determines the conversion viewpoint information from a larger range when playing in a space with a wide movement range such as outdoors, and when playing in a space with a narrow movement range such as indoors. The conversion viewpoint information may be determined from a narrower range.
 このように、視点バッファ制御部121は、ゲームタイトルやプレイ状況などのゲームに関する情報、換言すると、レンダリング画像を含むコンテンツやレンダリング画像のシーンに応じて、変換視点情報を決定し得る。特に、視点バッファ制御部121は、ゲームに関する情報のうち、視点移動の速さに関連する情報に応じて変換視点情報を決定し得る。 In this way, the viewpoint buffer control unit 121 can determine the conversion viewpoint information according to the information about the game such as the game title and the play status, in other words, the content including the rendered image and the scene of the rendered image. In particular, the viewpoint buffer control unit 121 can determine the converted viewpoint information according to the information related to the speed of the viewpoint movement among the information related to the game.
 また、上記例では、視点バッファ制御部121が所定範囲の中から均等に変換視点情報を決定するとしたが、これに限定されない。例えば、図6に示すように、視点バッファ制御部121は、レンダリング視点情報を中心とした円形の範囲R3の中から例えばレンダリング視点情報付近(範囲R3の中心)では密に、範囲R3の周辺では疎になるように変換視点情報を決定してもよい。例えば、サーバ装置100が視点移動の少ないゲームサービスを提供し、かつ、ネットワークNの遅延が大きい場合、視点バッファ制御部121は、図6に示すような密度で変換視点情報を決定し得る。 Further, in the above example, the viewpoint buffer control unit 121 evenly determines the conversion viewpoint information from the predetermined range, but the present invention is not limited to this. For example, as shown in FIG. 6, the viewpoint buffer control unit 121 is densely located in the vicinity of the rendering viewpoint information (center of the range R3) in the circular range R3 centered on the rendering viewpoint information, and around the range R3. The conversion viewpoint information may be determined so as to be sparse. For example, when the server device 100 provides a game service with less viewpoint movement and the delay of the network N is large, the viewpoint buffer control unit 121 can determine the conversion viewpoint information at the density shown in FIG.
 また、視点バッファ制御部121は、レンダリング視点情報、特に、レンダリング視点情報の履歴に応じて変換視点情報を決定するようにしてもよい。この場合、視点バッファ制御部121は、過去のレンダリング視点情報の変化(履歴)から、未来の時刻(例えばコマンド生成時刻t2)の視点情報を予測する。視点バッファ制御部121は、予測した視点情報を含む範囲から変換視点情報を決定する。 Further, the viewpoint buffer control unit 121 may determine the conversion viewpoint information according to the rendering viewpoint information, particularly the history of the rendering viewpoint information. In this case, the viewpoint buffer control unit 121 predicts the viewpoint information of the future time (for example, the command generation time t2) from the change (history) of the past rendering viewpoint information. The viewpoint buffer control unit 121 determines the converted viewpoint information from the range including the predicted viewpoint information.
 図7の例では、視点バッファ制御部121は、所定方向の視点情報(図7のP1)を予測し、予測した方向(図7の右上方向)に長い範囲R4から変換視点情報を決定する。 In the example of FIG. 7, the viewpoint buffer control unit 121 predicts the viewpoint information in a predetermined direction (P1 in FIG. 7), and determines the conversion viewpoint information from the long range R4 in the predicted direction (upper right direction in FIG. 7).
 なお、視点バッファ制御部121が決定する変換視点情報の数(視点数)は、一定であってもよく、可変であってもよい。例えば、サーバ装置100がクラウドシステムに構築され、サーバ装置100の計算リソースを柔軟に増減できる場合には、状況に応じてサーバ装置100が変換視点情報の数を変更するようにしてもよい。これにより、サーバ装置100による視点変換の計算コストを効率化することができる。また、例えばサーバ装置100を特定のハードウェアで構築する場合など、ハードウェアソリューションなどの制約のため演算量を一定に保ちたい場合には、サーバ装置100が一定数の変換視点情報を決定してもよい。 The number of converted viewpoint information (number of viewpoints) determined by the viewpoint buffer control unit 121 may be constant or variable. For example, when the server device 100 is constructed in a cloud system and the calculation resources of the server device 100 can be flexibly increased or decreased, the server device 100 may change the number of conversion viewpoint information according to the situation. This makes it possible to improve the efficiency of the calculation cost of the viewpoint conversion by the server device 100. Further, when it is desired to keep the amount of calculation constant due to restrictions such as a hardware solution, for example, when the server device 100 is constructed with specific hardware, the server device 100 determines a certain number of conversion viewpoint information. May be good.
 次に、視点バッファ制御部121が決定する変換視点情報について説明する。まず、図8を用いて、本開示の実施形態に係る視点情報の一例について説明する。図8は、本開示の実施形態に係る視点情報の一例について説明するための図である。 Next, the conversion viewpoint information determined by the viewpoint buffer control unit 121 will be described. First, an example of viewpoint information according to the embodiment of the present disclosure will be described with reference to FIG. FIG. 8 is a diagram for explaining an example of viewpoint information according to the embodiment of the present disclosure.
 上述したように視点情報には、視点位置及び視線方向が含まれる。図8に示すように視点位置は、レンダリングする画像を見たカメラ(又は、ユーザの頭部)の位置を表す視点(カメラ)座標(xwk、ywk、zwk)で表される。なお、視点座標の座標系は、例えばゲームのオブジェクトが配置されるワールド座標系である。また、視線方向は、カメラ(又は、ユーザの頭部)の向きを示す視線単位ベクトル(xvk、yvk、xvk)で表される。 As described above, the viewpoint information includes the viewpoint position and the line-of-sight direction. As shown in FIG. 8, the viewpoint position is represented by the viewpoint (camera) coordinates (x wk , y wk , z wk ) representing the position of the camera (or the user's head) that saw the image to be rendered. The coordinate system of the viewpoint coordinates is, for example, a world coordinate system in which game objects are arranged. Further, the line-of-sight direction is represented by a line-of-sight unit vector (x vk , y vk , x vk ) indicating the direction of the camera (or the user's head).
 視点バッファ制御部121は、複数の変換視点情報として、視点位置(視点座標)及び視線方向(視線単位ベクトル)の組を複数個決定する。図9は、本開示の実施形態に係る視点バッファ制御部121が決定する変換視点情報の一例を示す図である。図9では、視点バッファ制御部121がN個の変換視点情報を決定する場合について示している。 The viewpoint buffer control unit 121 determines a plurality of sets of the viewpoint position (viewpoint coordinates) and the line-of-sight direction (line-of-sight unit vector) as a plurality of converted viewpoint information. FIG. 9 is a diagram showing an example of conversion viewpoint information determined by the viewpoint buffer control unit 121 according to the embodiment of the present disclosure. FIG. 9 shows a case where the viewpoint buffer control unit 121 determines N conversion viewpoint information.
 図9に示すように、視点バッファ制御部121は、視点座標及び視線単位ベクトルの組として、{(xw1、yw1、zw1)、(xv1、yv1、xv1)}、{(xw2、yw2、zw2)、(xv2、yv2、xv2)}・・・{(xw2、yw2、zw2)、(xv2、yv2、xv2)}を決定する。 As shown in FIG. 9, the viewpoint buffer control unit 121 sets {(x w1 , y w1 , z w1 ), (x v1 , y v1 , x v1 )}, {(x w1, y w1, z w1)}, {(x w1, y w1, x v1)}, {(x w1, y w1, z w1)}, {(x w1, y v1, x v1)}, {(x w1, y w1, z w1), x w2 , y w2 , z w2 ), (x v2 , y v2 , x v2 )} ... {(x w2 , y w2 , z w2 ), (x v2 , y v2 , x v2 )} ..
 視点バッファ制御部121は、決定した視点座標及び視線単位ベクトルの組を多視点生成部122に出力する(図3参照)。 The viewpoint buffer control unit 121 outputs the determined set of viewpoint coordinates and line-of-sight unit vector to the multi-viewpoint generation unit 122 (see FIG. 3).
 なお、図8及び図9に示す変換視点情報は一例であり、これに限定されない。多視点生成部122が、変換視点情報に基づいてレンダリング画像の視点を変換できれば、変換視点情報はどのような形式で表現されてもよい。 Note that the conversion viewpoint information shown in FIGS. 8 and 9 is an example, and the present invention is not limited to this. The converted viewpoint information may be expressed in any format as long as the multi-viewpoint generation unit 122 can convert the viewpoint of the rendered image based on the converted viewpoint information.
 ここで、視点バッファ制御部121は、例えば記憶部(図示省略)に記憶する変換視点情報を決定した変換視点情報として多視点生成部122に出力してもよく、あるいは、関数を用いて決定してもよい。以下では、図10を用いて、視点バッファ制御部121が関数を用いて変換視点情報を決定する場合について説明するが、変換視点情報を決定する方法はこれに限定されるものではない。 Here, the viewpoint buffer control unit 121 may output, for example, the conversion viewpoint information stored in the storage unit (not shown) to the multi-viewpoint generation unit 122 as the determined conversion viewpoint information, or may be determined using a function. You may. Hereinafter, a case where the viewpoint buffer control unit 121 determines the conversion viewpoint information using a function will be described with reference to FIG. 10, but the method for determining the conversion viewpoint information is not limited to this.
 図10は、本開示の実施形態に係る視点バッファ制御部121の構成例を示すブロック図である。視点バッファ制御部121は、視点関数生成部1211と、視線関数生成部1213と、サンプリング部1212と、視線ベクトル生成部1214と、を備える。 FIG. 10 is a block diagram showing a configuration example of the viewpoint buffer control unit 121 according to the embodiment of the present disclosure. The viewpoint buffer control unit 121 includes a viewpoint function generation unit 1211, a line-of-sight function generation unit 1213, a sampling unit 1212, and a line-of-sight vector generation unit 1214.
 視点バッファ制御部121は、まず3次元空間上の視点密度関数及び視線関数の組を決定する。視点バッファ制御部121は、あらかじめ決められた視点密度関数及び視線関数を決定する。あるいは、視点バッファ制御部121が、ゲームに関する情報やネットワークトラフィックに関する情報、レンダリング視点情報等に応じて視点密度関数及び視線関数を決定してもよい。 The viewpoint buffer control unit 121 first determines a set of a viewpoint density function and a line-of-sight function in a three-dimensional space. The viewpoint buffer control unit 121 determines a predetermined viewpoint density function and line-of-sight function. Alternatively, the viewpoint buffer control unit 121 may determine the viewpoint density function and the line-of-sight function according to information about the game, information about network traffic, rendering viewpoint information, and the like.
 より具体的に、視点関数生成部1211は、例えば、任意の確率密度関数P(x、y、z)を視点密度関数Pに決定する。例えば、視点関数生成部1211は、あらかじめ記憶部(図示省略)に記憶される複数の確率密度関数の中からゲームに関する情報やネットワークトラフィックに関する情報、レンダリング視点情報等に応じて視点密度関数Pを決定する。 More specifically, the viewpoint function generation unit 1211 determines, for example, an arbitrary probability density function P (x, y, z) as the viewpoint density function P. For example, the viewpoint function generation unit 1211 determines the viewpoint density function P according to information about the game, information about network traffic, rendering viewpoint information, etc. from a plurality of probability density functions stored in advance in the storage unit (not shown). do.
 このとき、視点関数生成部1211は、ゲームに関する情報やネットワークトラフィックに関する情報、レンダリング視点情報等に応じて視点密度関数Pのパラメータを変更するようにしてもよい。視点関数生成部1211は、決定した視点密度関数Pをサンプリング部1212に出力する。 At this time, the viewpoint function generation unit 1211 may change the parameters of the viewpoint density function P according to the information about the game, the information about the network traffic, the rendering viewpoint information, and the like. The viewpoint function generation unit 1211 outputs the determined viewpoint density function P to the sampling unit 1212.
 また、視線関数生成部1213は、任意の視点座標を入力とし、視線単位ベクトルを出力とする関数を視線関数Vとして決定する。例えば、視線関数生成部1213は、あらかじめ記憶部(図示省略)に記憶される複数の視線関数の中からゲームに関する情報やネットワークトラフィックに関する情報、レンダリング視点情報等に応じて視線関数Vを決定する。 Further, the line-of-sight function generation unit 1213 determines a function that inputs an arbitrary viewpoint coordinate and outputs a line-of-sight unit vector as a line-of-sight function V. For example, the line-of-sight function generation unit 1213 determines the line-of-sight function V according to information about a game, information about network traffic, rendering viewpoint information, and the like from a plurality of line-of-sight functions stored in advance in a storage unit (not shown).
 このとき、視線関数生成部1213は、ゲームに関する情報やネットワークトラフィックに関する情報、レンダリング視点情報等に応じて視線関数Vのパラメータを変更するようにしてもよい。視線関数生成部1213は、決定した視線関数Vを視線ベクトル生成部1214に出力する。 At this time, the line-of-sight function generation unit 1213 may change the parameters of the line-of-sight function V according to information about the game, information about network traffic, rendering viewpoint information, and the like. The line-of-sight function generation unit 1213 outputs the determined line-of-sight function V to the line-of-sight vector generation unit 1214.
 サンプリング部1212は、例えば、視点密度関数Pから種々のサンプリング手法を用いて視点変換に使用するN個の視点座標を決定する。サンプリング手法としては、例えば、逆関数法、棄却サンプリング、マルコフ連鎖モンテカルロ法などが挙げられる。サンプリング部1212は、決定した視点座標を視線ベクトル生成部1214に出力する。 The sampling unit 1212 determines, for example, N viewpoint coordinates to be used for the viewpoint conversion from the viewpoint density function P by using various sampling methods. Examples of the sampling method include an inverse function method, rejection sampling, and Markov chain Monte Carlo method. The sampling unit 1212 outputs the determined viewpoint coordinates to the line-of-sight vector generation unit 1214.
 視線ベクトル生成部1214は、視線関数Vにサンプリング部1212が決定した視点座標を入力することで、視線単位ベクトルを決定する。視線ベクトル生成部1214は、決定したN個の視点座標及び視線単位ベクトルを視点バッファ情報として多視点生成部122(図3参照)に出力する。 The line-of-sight vector generation unit 1214 determines the line-of-sight unit vector by inputting the viewpoint coordinates determined by the sampling unit 1212 into the line-of-sight function V. The line-of-sight vector generation unit 1214 outputs the determined N viewpoint coordinates and line-of-sight unit vectors to the multi-viewpoint generation unit 122 (see FIG. 3) as viewpoint buffer information.
 なお、ここでは、視線ベクトル生成部1214が、変換視点情報として、視点座標及び視線単位ベクトルを出力するとしたが、これに限定されない。例えば、視点バッファ制御部121が視点バッファ情報として視点密度関数P及び視線関数Vを多視点生成部122に出力するようにしてもよい。あるいは、視線ベクトル生成部1214が所定の視点密度関数P、及び、所定の視線関数Vのパラメータを決定し、決定したパラメータを視点バッファ情報として多視点生成部122に出力するようにしてもよい。 Here, it is assumed that the line-of-sight vector generation unit 1214 outputs the viewpoint coordinates and the line-of-sight unit vector as the conversion viewpoint information, but the present invention is not limited to this. For example, the viewpoint buffer control unit 121 may output the viewpoint density function P and the line-of-sight function V to the multi-viewpoint generation unit 122 as viewpoint buffer information. Alternatively, the line-of-sight vector generation unit 1214 may determine the parameters of the predetermined viewpoint density function P and the predetermined line-of-sight function V, and output the determined parameters to the multi-viewpoint generation unit 122 as the viewpoint buffer information.
 このように、視点バッファ制御部121が、視点バッファ情報として視点密度関数P及び視線関数V、あるいは、当該関数を一意に決定するための情報(例えばパラメータ等)を出力するようにしてもよい。この場合、視点バッファ制御部121のサンプリング部1212及び視線ベクトル生成部1214は省略され得る。また、この場合サンプリング部1212によるサンプリング方法や使用する関数に関する情報を、多視点生成部122やクライアント装置200のデコード部220(図2参照)等が共有しているものとする。 In this way, the viewpoint buffer control unit 121 may output the viewpoint density function P and the line-of-sight function V as the viewpoint buffer information, or information (for example, parameters, etc.) for uniquely determining the function. In this case, the sampling unit 1212 and the line-of-sight vector generation unit 1214 of the viewpoint buffer control unit 121 may be omitted. Further, in this case, it is assumed that the information on the sampling method by the sampling unit 1212 and the function to be used is shared by the multi-viewpoint generation unit 122, the decoding unit 220 of the client device 200 (see FIG. 2), and the like.
 なお、ここでは、視点バッファ制御部121がネットワークNの通信状況やゲーム情報に応じて視点バッファ情報を生成するとしたが、これに限定されない。例えば、多視点変換部120が上述した情報によらずあらかじめ決められた一定の視点バッファ情報に基づき、視点変換画像を生成するようにしてもよい。この場合、サーバ装置100は、例えば、視点バッファ情報を記憶部(図示省略)にあらかじめ記憶させておくことで、視点バッファ制御部121を省略してもよい。 Here, it is assumed that the viewpoint buffer control unit 121 generates the viewpoint buffer information according to the communication status of the network N and the game information, but the present invention is not limited to this. For example, the multi-viewpoint conversion unit 120 may generate a viewpoint conversion image based on a predetermined constant viewpoint buffer information regardless of the above-mentioned information. In this case, the server device 100 may omit the viewpoint buffer control unit 121 by storing the viewpoint buffer information in a storage unit (not shown) in advance, for example.
 図3に戻る。多視点生成部122は、レンダリング画像に基づき、変換視点情報に含まれる視点座標から視線単位ベクトル方向を見た視点変換画像を生成する。 Return to Fig. 3. Based on the rendered image, the multi-viewpoint generation unit 122 generates a viewpoint conversion image in which the line-of-sight unit vector direction is viewed from the viewpoint coordinates included in the conversion viewpoint information.
 なお、図3では、レンダリング部110が生成し得る複数のレンダリング画像をマトリクス状に配置して示している。図示したマトリクス状のレンダリング画像のうち、ハッチングを付した画像は実際にはレンダリング部110によってレンダリングされない画像であり、レンダリング部110は、ハッチングを付していないレンダリング画像M0を多視点生成部122に出力する。 Note that FIG. 3 shows a plurality of rendered images that can be generated by the rendering unit 110 arranged in a matrix. Among the illustrated matrix-shaped rendered images, the hatched image is an image that is not actually rendered by the rendering unit 110, and the rendering unit 110 transfers the rendered image M0 without hatching to the multi-viewpoint generation unit 122. Output.
 また、図3では、多視点生成部122が生成し得る複数の視点変換画像をマトリクス状に配置して示している。図示したマトリクス状の視点変換画像のうち、ハッチングを付した視点変換画像は実際には多視点生成部122によって生成されない画像であり、多視点生成部122は、ハッチングを付していない視点変換画像M1~M5を生成する。 Further, in FIG. 3, a plurality of viewpoint conversion images that can be generated by the multi-viewpoint generation unit 122 are arranged and shown in a matrix. Of the matrix-shaped viewpoint conversion images shown in the figure, the viewpoint conversion image with hatching is an image that is not actually generated by the multi-viewpoint generation unit 122, and the multi-viewpoint generation unit 122 is a viewpoint conversion image without hatching. Generate M1 to M5.
 なお、多視点生成部122は、レンダリング付帯情報を使用して視点変換画像を生成する。 The multi-viewpoint generation unit 122 generates a viewpoint conversion image using the rendering incidental information.
 ここで、レンダリング付帯情報の一例について説明する。レンダリング付帯情報には、例えば、ゲーム画像のレンダリング時に都度生成されるメタデータ(以下、実行時メタデータと言う)や、ゲームの制作時に作成されるメタデータ(以下、制作時メタデータと言う)が含まれる。 Here, an example of rendering incidental information will be described. The rendering incidental information includes, for example, metadata generated each time a game image is rendered (hereinafter referred to as run-time metadata) and metadata created during game production (hereinafter referred to as production-time metadata). Is included.
 実行時メタデータの一例として、ゲーム内のオブジェクトの法線情報が挙げられる。実行時メタデータの一例として、オブジェクトのスペキュラー、ラフネス、アルベド、透明度など、光の反射に関連する情報が挙げられる。実行時メタデータの一例として、ゲーム内のオブジェクトの速度に関する情報が挙げられる。実行時メタデータの一例として、ゲーム内のオブジェクトの素材IDなど、オブジェクトの材質を表す情報が挙げられる。実行時メタデータの一例として、ゲーム内のオブジェクトのDepth情報が挙げられる。 An example of run-time metadata is the normal information of objects in the game. Examples of run-time metadata include information related to light reflection, such as object specular, roughness, albedo, and transparency. One example of run-time metadata is information about the speed of objects in the game. An example of run-time metadata is information that represents the material of an object, such as the material ID of the object in the game. An example of run-time metadata is the Depth information of objects in the game.
 上述した実行時メタデータは、プレイヤーの視点に依存して作成されるメタデータであるが、実行時メタデータには、プレイヤーの視点に依存せずに作成されるメタデータも含まれる。このような実行時メタデータの一例として、プレイヤーの状況やゲームの進行状況等を示すゲーム内のユーザインタフェース(UI)や、Reflection mapなどが挙げられる。 The above-mentioned run-time metadata is metadata created depending on the player's viewpoint, but the run-time metadata also includes metadata created independently of the player's viewpoint. Examples of such run-time metadata include an in-game user interface (UI) showing the player's status, the progress of the game, and the Reflection map.
 また、制作時メタデータの一例として、メッシュ(3Dモデル)やテクスチャ、ゲーム内のオブジェクトのアニメーションパターンやレイアウトデータのように、ゲームループのフレームに非同期であり、プレイヤーの視点に依存しないデータが挙げられる。 In addition, as an example of production metadata, there is data that is asynchronous to the frame of the game loop and does not depend on the player's viewpoint, such as mesh (3D model), texture, animation pattern of objects in the game, and layout data. Be done.
 多視点生成部122は、上述したレンダリング付帯情報を用いてレンダリング画像から視点変換画像を生成する。多視点生成部122は、例えばモデルベースの手法を用いて、レンダリング画像から視点変換画像を生成してもよく、あるいはDNNベースの手法を用いて視点変換画像を生成してもよい。このように、視点変換画像は、既存の種々の手法を用いて生成され得る。 The multi-viewpoint generation unit 122 generates a viewpoint conversion image from the rendered image using the above-mentioned rendering incidental information. The multi-viewpoint generation unit 122 may generate a viewpoint-converted image from a rendered image by using, for example, a model-based method, or may generate a viewpoint-converted image by using a DNN-based method. In this way, the viewpoint conversion image can be generated by using various existing methods.
 なお、上述したレンダリング付帯情報は、ゲームの種別、あるいは、ゲームのタイトルごとに異なり得る。そこで、多視点生成部122は、ゲームの種別、あるいは、ゲームのタイトルごとに特化した生成処理を行うようにしてもよい。例えば、多視点生成部122がDNNベースの手法を用いた視点変換画像の生成処理を実行する場合、ゲームの種別等に応じた係数データベースを用意しておくことで、多視点生成部122は、ゲームの種別に応じた生成処理を行うことができる。また、多視点生成部122がモデルベースの手法を用いた視点変換画像の生成処理を実行する場合、ゲームの種別等に応じたモデルを用意しておくことで、多視点生成部122は、ゲームの種別に応じた生成処理を行うことができる。 Note that the above-mentioned rendering incidental information may differ depending on the type of game or the title of the game. Therefore, the multi-viewpoint generation unit 122 may perform generation processing specialized for each type of game or each title of the game. For example, when the multi-viewpoint generation unit 122 executes a viewpoint conversion image generation process using a DNN-based method, the multi-viewpoint generation unit 122 can be generated by preparing a coefficient database according to the type of game or the like. It is possible to perform generation processing according to the type of game. Further, when the multi-viewpoint generation unit 122 executes the viewpoint conversion image generation process using the model-based method, the multi-viewpoint generation unit 122 can be used as a game by preparing a model according to the type of the game or the like. It is possible to perform generation processing according to the type of.
 このように、多視点生成部122が、ゲームの種別、あるいは、ゲームのタイトル、換言するとレンダリング付帯情報に応じた生成処理を実行することで、高精度に視点変換を行うことができる。 In this way, the multi-viewpoint generation unit 122 can perform viewpoint conversion with high accuracy by executing the generation process according to the type of the game, the title of the game, in other words, the rendering incidental information.
 なお、多視点生成部122が全てのゲームの種別、あるいは、ゲームのタイトルごとに生成処理を変更する必要はない。例えば、多視点生成部122が、特定のゲームにおいて特定の生成処理を実行するようにしてもよい。 It is not necessary for the multi-viewpoint generation unit 122 to change the generation process for all game types or game titles. For example, the multi-viewpoint generation unit 122 may execute a specific generation process in a specific game.
 視点バッファ情報にN個の変換視点情報が含まれる場合、多視点生成部122は、N個の視点変換画像を生成する。多視点生成部122は、生成したN個の視点変換画像を視点バッファ(多視点画像)としてエンコード部130に出力する。このとき、多視点生成部122は、視点バッファに加え、視点バッファ情報もエンコード部130に出力する。 When the viewpoint buffer information includes N conversion viewpoint information, the multi-view generation unit 122 generates N viewpoint conversion images. The multi-viewpoint generation unit 122 outputs the generated N viewpoint conversion images to the encoding unit 130 as a viewpoint buffer (multi-viewpoint image). At this time, the multi-viewpoint generation unit 122 outputs the viewpoint buffer information to the encoding unit 130 in addition to the viewpoint buffer.
 なお、ここでは、視点バッファ制御部121が視点バッファ情報を生成するとしたが、これに限定されない。例えば、多視点生成部122が、ゲームに関する情報、ネットワークトラフィックに関する情報等によらず、一定の範囲で一定数の変換視点情報に基づき視点変換を行ってもよい。この場合、例えば、多視点生成部122が、記憶部(図示省略)に記憶される視点バッファ情報に基づいて視点変換画像を生成するようにし、視点バッファ制御部121を省略してもよい。 Here, it is assumed that the viewpoint buffer control unit 121 generates the viewpoint buffer information, but the present invention is not limited to this. For example, the multi-viewpoint generation unit 122 may perform viewpoint conversion based on a certain number of conversion viewpoint information within a certain range regardless of information on a game, information on network traffic, and the like. In this case, for example, the multi-viewpoint generation unit 122 may generate the viewpoint conversion image based on the viewpoint buffer information stored in the storage unit (not shown), and the viewpoint buffer control unit 121 may be omitted.
 [エンコーダ部130]
 図2に戻る。エンコード部130は、多視点変換部120から視点バッファ情報及び視点バッファを受け取ると、クライアント装置200に伝送するためにエンコードする。エンコード部130は、視点バッファ情報を、視点バッファに先んじて単独でデコードできるようにエンコードを行う。エンコード部130が行うエンコードの手法は、モデルベースの手法であっても、DNNベースの手法であってもよい。
[Encoder unit 130]
Return to FIG. Upon receiving the viewpoint buffer information and the viewpoint buffer from the multi-viewpoint conversion unit 120, the encoding unit 130 encodes the viewpoint buffer information and the viewpoint buffer for transmission to the client device 200. The encoding unit 130 encodes the viewpoint buffer information so that it can be independently decoded prior to the viewpoint buffer. The encoding method performed by the encoding unit 130 may be a model-based method or a DNN-based method.
 また、エンコード部130は、時間方向の視点バッファ情報及び視点バッファを考慮してエンコードを行ってもよく、行わなくてもよい。すなわし、エンコード部130が行う符号化は、イントラ符号化であってもよく、インター符号化であってもよい。 Further, the encoding unit 130 may or may not perform encoding in consideration of the viewpoint buffer information and the viewpoint buffer in the time direction. That is, the coding performed by the encoding unit 130 may be intra-coding or inter-coding.
 なお、多視点変換部120は、視点が少しずれた多視点画像を生成する。このことから多視点画像に含まれる各視点変換画像は大きく変化せず、視点変換画像の冗長性が高いと考えられる。そのため、エンコード部130は、多視点画像(視点バッファ)を高い圧縮率で圧縮することができる。 The multi-viewpoint conversion unit 120 generates a multi-viewpoint image with a slightly deviated viewpoint. From this, it is considered that each viewpoint-converted image included in the multi-viewpoint image does not change significantly, and the viewpoint-converted image has high redundancy. Therefore, the encoding unit 130 can compress the multi-viewpoint image (viewpoint buffer) at a high compression rate.
 <2.3.クライアント装置>
 [デコード部220]
 次に、本開示の実施形態に係るクライアント装置200のデコード部220の一例について説明する。デコード部220は、送受信部210が受信した圧縮データから、まず視点バッファ情報をデコードし、デコードした視点バッファ情報及び現在のユーザの視点情報から、視点バッファに含まれる視点変換画像をデコードする。デコード部220は、現在のユーザの視点情報及びデコードした視点変換画像に基づき、表示画像(ゲーム画像)を生成する。
<2.3. Client device>
[Decoding unit 220]
Next, an example of the decoding unit 220 of the client device 200 according to the embodiment of the present disclosure will be described. The decoding unit 220 first decodes the viewpoint buffer information from the compressed data received by the transmission / reception unit 210, and decodes the viewpoint conversion image included in the viewpoint buffer from the decoded viewpoint buffer information and the current user's viewpoint information. The decoding unit 220 generates a display image (game image) based on the current user's viewpoint information and the decoded viewpoint conversion image.
 図11は、本開示の実施形態に係るデコード部220の構成例を示すブロック図である。図11に示すデコード部220は、バッファ情報デコード部221と、デコード制御部222と、視点バッファデコード部223と、画像生成部224と、を備える。 FIG. 11 is a block diagram showing a configuration example of the decoding unit 220 according to the embodiment of the present disclosure. The decoding unit 220 shown in FIG. 11 includes a buffer information decoding unit 221, a decoding control unit 222, a viewpoint buffer decoding unit 223, and an image generation unit 224.
 バッファ情報デコード部221は、送受信部210から受信データとして圧縮データを受け取る。バッファ情報デコード部221は、圧縮データの中から視点バッファ情報をデコードし、デコード制御部222に出力する。 The buffer information decoding unit 221 receives compressed data as received data from the transmission / reception unit 210. The buffer information decoding unit 221 decodes the viewpoint buffer information from the compressed data and outputs it to the decoding control unit 222.
 デコード制御部222は、入力情報取得部230から現在のユーザの視点情報を取得する。現在のユーザの視点情報は、ユーザの操作や動作などによるコマンド情報により決定される視点情報であり、表示画像としてレンダリングする視点位置及び視線方向を示すものである。 The decode control unit 222 acquires the viewpoint information of the current user from the input information acquisition unit 230. The current user's viewpoint information is viewpoint information determined by command information such as a user's operation or operation, and indicates a viewpoint position and a line-of-sight direction to be rendered as a display image.
 デコード制御部222は、視点バッファ情報に含まれる変換視点情報の座標系と、入力情報取得部230から取得する視点情報の座標系とが異なる場合、一方の座標系を他方の座標系にあわせる座標変換を行う。ここでは、デコード制御部222が、変換視点情報の座標系に統一して処理を行うものとする。 When the coordinate system of the converted viewpoint information included in the viewpoint buffer information and the coordinate system of the viewpoint information acquired from the input information acquisition unit 230 are different, the decode control unit 222 coordinates one coordinate system to match the other coordinate system. Perform the conversion. Here, it is assumed that the decoding control unit 222 performs processing in a unified manner in the coordinate system of the conversion viewpoint information.
 デコード制御部222は、現在のユーザの視点情報及び視点バッファ情報に基づき、視点バッファデコード部223でデコードする視点変換画像の視点情報(以下、デコード視点情報と言う)を決定する。 The decoding control unit 222 determines the viewpoint information (hereinafter referred to as decoding viewpoint information) of the viewpoint conversion image to be decoded by the viewpoint buffer decoding unit 223 based on the viewpoint information and the viewpoint buffer information of the current user.
 図12は、本開示の実施形態に係るデコード制御部222が決定するデコード視点情報について説明するための図である。 FIG. 12 is a diagram for explaining the decoding viewpoint information determined by the decoding control unit 222 according to the embodiment of the present disclosure.
 図12では、多視点生成部122で生成し得る複数の視点変換画像をマトリクス状に並べて示している。また、図12では、実際に視点バッファに含まれる視点変換画像の視点情報、換言すると視点バッファ情報に含まれる変換視点情報を十字で示している。また、図12における「×」は、現在のユーザの視点情報VP0を示している。 FIG. 12 shows a plurality of viewpoint conversion images that can be generated by the multi-viewpoint generation unit 122 side by side in a matrix. Further, in FIG. 12, the viewpoint information of the viewpoint converted image actually included in the viewpoint buffer, in other words, the converted viewpoint information included in the viewpoint buffer information is shown by a cross. Further, "x" in FIG. 12 indicates the viewpoint information VP0 of the current user.
 デコード制御部222は、現在のユーザ視点情報VP0に基づき、変換視点情報VP1~VP12の中からM個の変換視点情報をデコード視点情報に決定する。例えば、デコード制御部222は、現在のユーザ視点情報VP0から距離が近い順に上位M個の変換視点情報をデコード視点情報として選択する。 The decoding control unit 222 determines M conversion viewpoint information from the conversion viewpoint information VP1 to VP12 as the decoding viewpoint information based on the current user viewpoint information VP0. For example, the decoding control unit 222 selects the upper M conversion viewpoint information as the decoding viewpoint information in order of increasing distance from the current user viewpoint information VP0.
 図12の例では、デコード制御部222は、デコード視点情報として、ユーザ視点情報VP0から距離が近い順にM=4個の変換視点情報VP4、VP7、VP8、VP11を選択する。なお、図12では、デコード制御部222が選択した変換視点情報VP4、VP7、VP8、VP11を丸で囲った十字で示している。 In the example of FIG. 12, the decoding control unit 222 selects M = 4 conversion viewpoint information VP4, VP7, VP8, VP11 as the decoding viewpoint information in order of distance from the user viewpoint information VP0. In FIG. 12, the conversion viewpoint information VP4, VP7, VP8, and VP11 selected by the decoding control unit 222 are indicated by a circled cross.
 なお、デコード制御部222によるデコード視点情報の選択方法及び個数は図12の例に限定されない。後段の画像生成部224が、現在のユーザの視点情報VP0に対応する表示画像を補間、生成し得る変換視点情報が選択されればよい。選択方法及び個数は、画像生成部224による処理に応じて適宜変更され得る。 The method and number of decoding viewpoint information selected by the decoding control unit 222 is not limited to the example shown in FIG. The conversion viewpoint information that can be generated by interpolating the display image corresponding to the current user's viewpoint information VP0 by the image generation unit 224 in the subsequent stage may be selected. The selection method and the number may be appropriately changed depending on the processing by the image generation unit 224.
 図11に戻る。デコード制御部222は、選択したデコード視点情報を視点バッファデコード部223に出力する。また、デコード制御部222は、現在のユーザの視点情報VP0を画像生成部224に出力する。 Return to Fig. 11. The decoding control unit 222 outputs the selected decoding viewpoint information to the viewpoint buffer decoding unit 223. Further, the decoding control unit 222 outputs the current user's viewpoint information VP0 to the image generation unit 224.
 視点バッファデコード部223は、送受信部210から取得した受信データから、デコード視点情報に対応する視点変換画像をデコードする。図12の例では、視点バッファデコード部223は、デコード視点情報として変換視点情報VP4、VP7、VP8、VP11を取得する。この場合、視点バッファデコード部223は、変換視点情報VP4、VP7、VP8、VP11に対応する視点変換画像をデコードする。 The viewpoint buffer decoding unit 223 decodes the viewpoint conversion image corresponding to the decoding viewpoint information from the received data acquired from the transmission / reception unit 210. In the example of FIG. 12, the viewpoint buffer decoding unit 223 acquires the conversion viewpoint information VP4, VP7, VP8, and VP11 as the decoding viewpoint information. In this case, the viewpoint buffer decoding unit 223 decodes the viewpoint conversion image corresponding to the conversion viewpoint information VP4, VP7, VP8, and VP11.
 視点バッファデコード部223は、デコードしたM個の視点変換画像、及び、デコード視点情報を画像生成部224に出力する。 The viewpoint buffer decoding unit 223 outputs the decoded M viewpoint conversion images and the decoded viewpoint information to the image generation unit 224.
 画像生成部224は、視点バッファデコード部223から取得したM個の視点変換画像、及び、デコード視点情報に基づき、デコード制御部222から取得した現在のユーザの視点情報に対応するゲーム画像を生成する。画像生成部224は、例えば、M個の視点変換画像からゲーム画像を補間、生成する。画像生成部224は、種々の手法を用いてゲーム画像の補間、生成を行い得る。画像生成部224は、例えばモデルベースの手法、あるいは、DNNベースの手法を用いてゲーム画像の補間、生成を行い得る。 The image generation unit 224 generates game images corresponding to the current user's viewpoint information acquired from the decoding control unit 222 based on the M viewpoint conversion images acquired from the viewpoint buffer decoding unit 223 and the decoding viewpoint information. .. The image generation unit 224 interpolates and generates a game image from, for example, M viewpoint conversion images. The image generation unit 224 can interpolate and generate a game image by using various methods. The image generation unit 224 can interpolate and generate a game image using, for example, a model-based method or a DNN-based method.
 画像生成部224は、生成したゲーム画像を周辺機器300の表示部(図示省略)に出力することで、ユーザにゲーム画像を提示する。 The image generation unit 224 presents the game image to the user by outputting the generated game image to the display unit (not shown) of the peripheral device 300.
 このように、デコード部220は、現在のユーザの視点情報VP0に基づき、ゲーム画像の生成に使用する個数の視点変換画像をデコードする。これにより、デコード部220は、視点バッファに含まれる全ての視点変換画像をデコードする場合と比較して、処理負荷を低減させることができる。 In this way, the decoding unit 220 decodes the number of viewpoint conversion images used for generating the game image based on the viewpoint information VP0 of the current user. As a result, the decoding unit 220 can reduce the processing load as compared with the case of decoding all the viewpoint conversion images included in the viewpoint buffer.
 [入力情報取得部230]
 図2に戻る。入力情報取得部230は、ユーザによる周辺機器300(例えば、コントローラ)の操作や、周辺機器300(例えば、HMD)によるユーザの動作を検出する。入力情報取得部230は、検出結果をサーバ装置100の例えばレンダリング部110が解釈し得るコマンド情報に変換し、現在のユーザの視点情報を生成する。
[Input information acquisition unit 230]
Return to FIG. The input information acquisition unit 230 detects the operation of the peripheral device 300 (for example, the controller) by the user and the operation of the user by the peripheral device 300 (for example, HMD). The input information acquisition unit 230 converts the detection result into command information that can be interpreted by, for example, the rendering unit 110 of the server device 100, and generates the viewpoint information of the current user.
 入力情報取得部230は、現在のユーザの視点情報をデコード部220に出力する。また、入力情報取得部230は、現在のユーザの視点情報をコマンド情報に含めて送受信部210を介してサーバ装置100に送信する。 The input information acquisition unit 230 outputs the viewpoint information of the current user to the decoding unit 220. Further, the input information acquisition unit 230 includes the viewpoint information of the current user in the command information and transmits it to the server device 100 via the transmission / reception unit 210.
 <<3.情報処理>>
 <3.1.多視点変換処理>
 図13は、本開示の実施形態に係るサーバ装置100が実行する多視点変換処理の流れを示すフローチャートである。サーバ装置100は、クライアント装置200にサービスを提供している間、例えば所定の周期で図13に示す多視点変換処理を繰り返し実行する。
<< 3. Information processing >>
<3.1. Multi-viewpoint conversion process>
FIG. 13 is a flowchart showing the flow of the multi-viewpoint conversion process executed by the server device 100 according to the embodiment of the present disclosure. While providing the service to the client device 200, the server device 100 repeatedly executes the multi-viewpoint conversion process shown in FIG. 13, for example, at a predetermined cycle.
 図13に示すように、サーバ装置100は、第1の数(例えば1個)の視点位置及び視線方向から見たレンダリング画像を取得する(ステップS101)。また、サーバ装置100は、レンダリング画像の生成に使用したレンダリング付帯情報を取得する(ステップS102)。 As shown in FIG. 13, the server device 100 acquires a rendered image viewed from the viewpoint position and the line-of-sight direction of the first number (for example, one) (step S101). Further, the server device 100 acquires the rendering incidental information used for generating the rendered image (step S102).
 サーバ装置100は、第1の数より多い第2の数(例えば12個)の視点位置及び視線方向を含む視点バッファ情報を生成する(ステップS103)。サーバ装置100は、視点バッファ情報に含まれる視点位置及び視線方向から見た視点変換画像を含む視点バッファを生成する(ステップS104)。サーバ装置100は、視点バッファをレンダリング画像及びレンダリング付帯情報に基づいて生成する。 The server device 100 generates viewpoint buffer information including a second number (for example, 12) of viewpoint positions and line-of-sight directions, which is larger than the first number (step S103). The server device 100 generates a viewpoint buffer including the viewpoint position and the viewpoint conversion image seen from the line-of-sight direction included in the viewpoint buffer information (step S104). The server device 100 generates a viewpoint buffer based on the rendered image and the rendering incidental information.
 サーバ装置100は、視点バッファ情報及び視点バッファをエンコードし(ステップS105)、得られた送信データをクライアント装置200に送信する(ステップS106)。 The server device 100 encodes the viewpoint buffer information and the viewpoint buffer (step S105), and transmits the obtained transmission data to the client device 200 (step S106).
 <3.2.表示画像生成処理>
 図14は、本開示の実施形態に係るクライアント装置200が実行する表示画像生成処理の流れを示すフローチャートである。クライアント装置200は、サーバ装置100から信号を受信すると図14に示す表示画像生成処理を実行する。
<3.2. Display image generation process>
FIG. 14 is a flowchart showing a flow of display image generation processing executed by the client device 200 according to the embodiment of the present disclosure. When the client device 200 receives a signal from the server device 100, the client device 200 executes the display image generation process shown in FIG.
 図14に示すように、クライアント装置200は、受信信号を受信する(ステップS201)。クライアント装置200は、受信信号からまず視点バッファ情報をデコードする(ステップS202)。 As shown in FIG. 14, the client device 200 receives the received signal (step S201). The client device 200 first decodes the viewpoint buffer information from the received signal (step S202).
 クライアント装置200は、視点バッファ情報に含まれる変換視点情報から、デコードする視点変換画像に対応する変換視点情報をデコード視点情報として選択する(ステップS203)。 The client device 200 selects the converted viewpoint information corresponding to the viewpoint converted image to be decoded from the converted viewpoint information included in the viewpoint buffer information as the decoded viewpoint information (step S203).
 クライアント装置200は、視点バッファをデコードし、デコード視点情報に対応する視点変換画像を取得する(ステップS204)。 The client device 200 decodes the viewpoint buffer and acquires a viewpoint conversion image corresponding to the decoded viewpoint information (step S204).
 クライアント装置200は、デコードした視点変換画像を用いて現在のユーザの視点から見た表示画像(ゲーム画像)を補間し生成する(ステップS205)。 The client device 200 interpolates and generates a display image (game image) viewed from the current user's viewpoint using the decoded viewpoint conversion image (step S205).
 <<4.その他の実施形態>>
 上述の実施形態は一例を示したものであり、種々の変更及び応用が可能である。
<< 4. Other embodiments >>
The above embodiment shows an example, and various modifications and applications are possible.
 本実施形態のサーバ装置100又はクライアント装置200を制御する制御装置は、専用のコンピュータシステムで実現してもよいし、汎用のコンピュータシステムで実現してもよい。 The control device for controlling the server device 100 or the client device 200 of the present embodiment may be realized by a dedicated computer system or a general-purpose computer system.
 例えば、上述の動作を実行するためのプログラムを、光ディスク、半導体メモリ、磁気テープ、フレキシブルディスク等のコンピュータ読み取り可能な記録媒体に格納して配布する。そして、例えば、該プログラムをコンピュータにインストールし、上述の処理を実行することによって制御装置を構成する。このとき、制御装置は、サーバ装置100又はクライアント装置200の外部の装置(例えば、パーソナルコンピュータ)であってもよい。また、制御装置は、サーバ装置100又はクライアント装置200の内部の装置であってもよい。 For example, a program for executing the above operation is stored and distributed in a computer-readable recording medium such as an optical disk, a semiconductor memory, a magnetic tape, or a flexible disk. Then, for example, the control device is configured by installing the program in a computer and executing the above-mentioned processing. At this time, the control device may be an external device (for example, a personal computer) of the server device 100 or the client device 200. Further, the control device may be an internal device of the server device 100 or the client device 200.
 また、上記通信プログラムをインターネット等のネットワーク上のサーバ装置が備えるディスク装置に格納しておき、コンピュータにダウンロード等できるようにしてもよい。また、上述の機能を、OS(Operating System)とアプリケーションソフトとの協働により実現してもよい。この場合には、OS以外の部分を媒体に格納して配布してもよいし、OS以外の部分をサーバ装置に格納しておき、コンピュータにダウンロード等できるようにしてもよい。 Further, the above communication program may be stored in a disk device provided in a server device on a network such as the Internet so that it can be downloaded to a computer or the like. Further, the above-mentioned functions may be realized by cooperation between the OS (Operating System) and the application software. In this case, the part other than the OS may be stored in a medium and distributed, or the part other than the OS may be stored in the server device so that it can be downloaded to a computer or the like.
 また、上記実施形態において説明した各処理のうち、自動的に行われるものとして説明した処理の全部又は一部を手動的に行うこともでき、あるいは、手動的に行われるものとして説明した処理の全部又は一部を公知の方法で自動的に行うこともできる。この他、上記文書中や図面中で示した処理手順、具体的名称、各種のデータやパラメータを含む情報については、特記する場合を除いて任意に変更することができる。例えば、各図に示した各種情報は、図示した情報に限られない。 Further, among the processes described in the above-described embodiment, all or a part of the processes described as being automatically performed can be manually performed, or the processes described as being manually performed can be performed. All or part of it can be done automatically by a known method. In addition, information including processing procedures, specific names, various data and parameters shown in the above documents and drawings can be arbitrarily changed unless otherwise specified. For example, the various information shown in each figure is not limited to the information shown in the figure.
 また、図示した各装置の各構成要素は機能概念的なものであり、必ずしも物理的に図示の如く構成されていることを要しない。すなわち、各装置の分散・統合の具体的形態は図示のものに限られず、その全部又は一部を、各種の負荷や使用状況などに応じて、任意の単位で機能的又は物理的に分散・統合して構成することができる。 Further, each component of each device shown in the figure is a functional concept, and does not necessarily have to be physically configured as shown in the figure. That is, the specific form of distribution / integration of each device is not limited to the one shown in the figure, and all or part of them may be functionally or physically distributed / physically in arbitrary units according to various loads and usage conditions. Can be integrated and configured.
 また、上記してきた実施形態は、処理内容を矛盾させない領域で適宜組み合わせることが可能である。また、本実施形態のシーケンス図或いはフローチャートに示された各ステップは、適宜順序を変更することが可能である。 Further, the above-described embodiments can be appropriately combined in an area where the processing contents do not contradict each other. Further, the order of each step shown in the sequence diagram or the flowchart of the present embodiment can be changed as appropriate.
 また、例えば、本実施形態は、装置またはシステムを構成するあらゆる構成、例えば、システムLSI(Large Scale Integration)等としてのプロセッサ、複数のプロセッサ等を用いるモジュール、複数のモジュール等を用いるユニット、ユニットにさらにその他の機能を付加したセット等(すなわち、装置の一部の構成)として実施することもできる。 Further, for example, the present embodiment includes a device or any configuration constituting the system, for example, a processor as a system LSI (Large Scale Integration), a module using a plurality of processors, a unit using a plurality of modules, and a unit. It can also be implemented as a set or the like (that is, a configuration of a part of the device) to which other functions are added.
 なお、本実施形態において、システムとは、複数の構成要素(装置、モジュール(部品)等)の集合を意味し、全ての構成要素が同一筐体中にあるか否かは問わない。したがって、別個の筐体に収納され、ネットワークを介して接続されている複数の装置、及び、1つの筐体の中に複数のモジュールが収納されている1つの装置は、いずれも、システムである。 In the present embodiment, the system means a set of a plurality of components (devices, modules (parts), etc.), and it does not matter whether all the components are in the same housing. Therefore, a plurality of devices housed in separate housings and connected via a network, and a device in which a plurality of modules are housed in one housing are both systems. ..
 また、例えば、本実施形態は、1つの機能を、ネットワークを介して複数の装置で分担、共同して処理するクラウドコンピューティングの構成をとることができる。 Further, for example, the present embodiment can have a cloud computing configuration in which one function is shared by a plurality of devices via a network and jointly processed.
 <<5.まとめ>>
 上述してきたように、本開示の実施形態に係るサーバ装置100(情報処理装置の一例)は、多視点変換部120(生成部の一例)と、送受信部140(送信部の一例)と、を備える。多視点変換部120は、レンダリング視点情報に含まれる視点位置及び視線方向(第1の数の視点位置及び視線方向の一例)から見たレンダリング画像(第1のレンダリング画像の一例)、及び、レンダリング付帯情報(レンダリング画像のレンダリングに使用したレンダリング情報の一例)を取得する。多視点変換部120は、視点バッファ情報(第1の数より多い第2の数の視点位置及び視線方向に関する視線情報の一例)を生成する。多視点変換部120は、レンダリング画像及びレンダリング付帯情報に基づき、デコード視点情報に含まれる視点位置及び視線方向から見た視点変換画像(第2のレンダリング画像の一例)を生成する。多視点変換部120は、視点バッファ情報及び視点変換画像を含む視点バッファをクライアント装置200(端末装置の一例)に送信する。
<< 5. Summary >>
As described above, the server device 100 (an example of an information processing device) according to the embodiment of the present disclosure includes a multi-viewpoint conversion unit 120 (an example of a generation unit) and a transmission / reception unit 140 (an example of a transmission unit). Be prepared. The multi-viewpoint conversion unit 120 includes a rendered image (an example of the first rendered image) and rendering as seen from the viewpoint position and the line-of-sight direction (an example of the first number of viewpoint positions and the line-of-sight direction) included in the rendered viewpoint information. Acquires incidental information (an example of rendering information used for rendering a rendered image). The multi-viewpoint conversion unit 120 generates viewpoint buffer information (an example of line-of-sight information regarding a second number of viewpoint positions and line-of-sight directions that is larger than the first number). The multi-viewpoint conversion unit 120 generates a viewpoint conversion image (an example of the second rendered image) seen from the viewpoint position and the line-of-sight direction included in the decoded viewpoint information based on the rendered image and the rendering incidental information. The multi-viewpoint conversion unit 120 transmits the viewpoint buffer including the viewpoint buffer information and the viewpoint conversion image to the client device 200 (an example of the terminal device).
 これにより、サーバ装置100は、遅延をより低減した高画質な画像をユーザに提示することができ、ユーザに提供するサービスの品質劣化をより抑制することができる。 Thereby, the server device 100 can present the user with a high-quality image with further reduced delay, and can further suppress the deterioration of the quality of the service provided to the user.
 また、本開示の実施形態に係るサーバ装置100の多視点変換部120は、クライアント装置200との通信で発生する遅延状況に応じて、視点バッファ情報に含める視点位置を決定する。 Further, the multi-viewpoint conversion unit 120 of the server device 100 according to the embodiment of the present disclosure determines the viewpoint position to be included in the viewpoint buffer information according to the delay situation generated in the communication with the client device 200.
 これにより、サーバ装置100は、遅延状況に応じた視点変換画像を生成することができ、遅延をより低減した高画質な画像をユーザに提示することができる。 As a result, the server device 100 can generate a viewpoint conversion image according to the delay situation, and can present the user with a high-quality image with further reduced delay.
 また、本開示の実施形態に係るサーバ装置100の多視点変換部120は、レンダリング視点情報の履歴(過去に取得したレンダリング画像の視点位置の一例)に基づき、視点バッファに含める視点位置を決定する。 Further, the multi-viewpoint conversion unit 120 of the server device 100 according to the embodiment of the present disclosure determines the viewpoint position to be included in the viewpoint buffer based on the history of the rendering viewpoint information (an example of the viewpoint position of the rendered image acquired in the past). ..
 これにより、サーバ装置100は、ユーザの操作又は動きに応じた視点変換画像を生成することができ、遅延をより低減した高画質な画像をユーザに提示することができる。 Thereby, the server device 100 can generate a viewpoint conversion image according to the user's operation or movement, and can present the user with a high-quality image with further reduced delay.
 また、本開示の実施形態に係るサーバ装置100の多視点変換部120は、ゲームの種別やタイトル(レンダリング画像を含むコンテンツの一例)に応じて、視点バッファ情報に含める視点位置を決定する。 Further, the multi-viewpoint conversion unit 120 of the server device 100 according to the embodiment of the present disclosure determines the viewpoint position to be included in the viewpoint buffer information according to the type and title of the game (an example of the content including the rendered image).
 これにより、サーバ装置100は、コンテンツ特有のユーザの操作や動きに応じた視点変換画像を生成することができ、遅延をより低減した高画質な画像をユーザに提示することができる。 As a result, the server device 100 can generate a viewpoint conversion image according to the user's operation or movement peculiar to the content, and can present the user with a high-quality image with further reduced delay.
 また、本開示の実施形態に係るサーバ装置100の多視点変換部120は、ゲームのプレイ状況(レンダリング画像のシーンの一例)に応じて、視点バッファ情報に含める視点位置を決定する。 Further, the multi-viewpoint conversion unit 120 of the server device 100 according to the embodiment of the present disclosure determines the viewpoint position to be included in the viewpoint buffer information according to the game play situation (an example of the scene of the rendered image).
 これにより、サーバ装置100は、コンテンツのシーン(例えば、プレイ状況)に応じたユーザの操作や動きに対応しうる視点変換画像を生成することができ、遅延をより低減した高画質な画像をユーザに提示することができる。 As a result, the server device 100 can generate a viewpoint conversion image that can correspond to the user's operation and movement according to the scene of the content (for example, the play situation), and the user can obtain a high-quality image with further reduced delay. Can be presented to.
 また、本開示の実施形態に係るレンダリング情報は、実行時メタデータ(レンダリング画像をレンダリングする時に生成される情報の一例)である。 Further, the rendering information according to the embodiment of the present disclosure is run-time metadata (an example of information generated when rendering a rendered image).
 これにより、サーバ装置100は、実行時メタデータを使用して視点変換画像を生成することで、より高画質な視点変換画像を生成することができる。 As a result, the server device 100 can generate a viewpoint conversion image with higher image quality by generating the viewpoint conversion image using the runtime metadata.
 また、本開示の実施形態に係るレンダリング情報は、制作時メタデータ(レンダリング画像をレンダリングするためにあらかじめ設定されている情報の一例)である。 Further, the rendering information according to the embodiment of the present disclosure is production-time metadata (an example of information preset for rendering a rendered image).
 これにより、サーバ装置100は、制作時メタデータを使用して視点変換画像を生成することで、より高画質な視点変換画像を生成することができる。 As a result, the server device 100 can generate a viewpoint-converted image with higher image quality by generating the viewpoint-converted image using the metadata at the time of production.
 また、本開示の実施形態に係るサーバ装置100の多視点変換部120は、視点密度関数(視点位置の分布を表す視点関数の一例)、及び、視線関数を算出し、算出した視点密度関数及び視線関数を用いて視点バッファ情報を生成する。サーバ装置100の送受信部140は、視点密度関数及び視線関数に関する情報を視点バッファ情報として送信する。 Further, the multi-viewpoint conversion unit 120 of the server device 100 according to the embodiment of the present disclosure calculates the viewpoint density function (an example of the viewpoint function representing the distribution of the viewpoint position), the line-of-sight function, and the calculated viewpoint density function. The viewpoint buffer information is generated using the line-of-sight function. The transmission / reception unit 140 of the server device 100 transmits information regarding the viewpoint density function and the line-of-sight function as viewpoint buffer information.
 これにより、サーバ装置100は、視点バッファ情報の情報量を削減することができる。 Thereby, the server device 100 can reduce the amount of information of the viewpoint buffer information.
 また、本開示の実施形態に係るクライアント装置200(情報処理装置の一例)は、送受信部210(受信部の一例)と、デコード部220(生成部の一例)と、を備える。受信部210は、レンダリング視点情報に含まれる視点位置及び視線方向から見たレンダリング画像、及び、レンダリング付帯情報に基づいて生成される視点変換画像であって、変換視点情報に含まれる視点位置及び視線方向から見た視点変換画像を含む視点バッファと、変換視点情報を含む視点バッファ情報と、を受信する。デコード部220は、視点変換画像及び視点バッファ情報に基づき、表示画像を生成する。 Further, the client device 200 (an example of an information processing device) according to the embodiment of the present disclosure includes a transmission / reception unit 210 (an example of a reception unit) and a decoding unit 220 (an example of a generation unit). The receiving unit 210 is a viewpoint converted image generated based on the rendered image seen from the viewpoint position and the line-of-sight direction included in the rendered viewpoint information and the rendering incidental information, and the viewpoint position and the line of sight included in the converted viewpoint information. The viewpoint buffer including the viewpoint conversion image viewed from the direction and the viewpoint buffer information including the conversion viewpoint information are received. The decoding unit 220 generates a display image based on the viewpoint conversion image and the viewpoint buffer information.
 これにより、クライアント装置200は、遅延をより低減した高画質な画像をユーザに提示することができ、ユーザに提供するサービスの品質劣化をより抑制することができる。 Thereby, the client device 200 can present the user with a high-quality image with further reduced delay, and can further suppress the deterioration of the quality of the service provided to the user.
 また、本開示の実施形態に係るクライアント装置200のデコード部220は、ユーザの現在の視点位置及び視線方向に基づき、視点バッファ情報に含まれる少なくとも1つの視点位置及び視線方向を選択し、選択した視点位置及び視線方向から見た視点変換画像に基づいて前記表示画像を生成する。 Further, the decoding unit 220 of the client device 200 according to the embodiment of the present disclosure selects and selects at least one viewpoint position and line-of-sight direction included in the viewpoint buffer information based on the user's current viewpoint position and line-of-sight direction. The display image is generated based on the viewpoint conversion image viewed from the viewpoint position and the line-of-sight direction.
 これにより、クライアント装置200は、デコードする視点変換画像の数を低減しつつ、高画質な画像をユーザに提示することができる。 As a result, the client device 200 can present a high-quality image to the user while reducing the number of viewpoint-converted images to be decoded.
 また、本開示の実施形態に係る情報処理システム10は、サーバ装置100と、クライアント装置200と、を備える。サーバ装置100は、多視点変換部120と、送受信部140と、を備える。多視点変換部120は、レンダリング視点情報に含まれる視点位置及び視線方向から見たレンダリング画像、及び、レンダリング画像付帯情報を取得する。多視点変換部120は、視点バッファ情報を生成する。多視点変換部120は、レンダリング画像及びレンダリング付帯情報に基づき、デコード視点情報に含まれる視点位置及び視線方向から見た視点変換画像を生成する。送受信部140は、視点バッファ情報及び視点変換画像を含む視点バッファをクライアント装置200に送信する。クライアント装置200は、送受信部210と、デコード部220と、を備える。送受信部210は、視点変換画像と、視点バッファ情報と、を受信する。デコード部220は、視点変換画像及び視点バッファ情報に基づき、表示画像を生成する。 Further, the information processing system 10 according to the embodiment of the present disclosure includes a server device 100 and a client device 200. The server device 100 includes a multi-viewpoint conversion unit 120 and a transmission / reception unit 140. The multi-viewpoint conversion unit 120 acquires the rendered image as seen from the viewpoint position and the line-of-sight direction included in the rendered viewpoint information, and the rendered image incidental information. The multi-viewpoint conversion unit 120 generates viewpoint buffer information. The multi-viewpoint conversion unit 120 generates a viewpoint conversion image seen from the viewpoint position and the line-of-sight direction included in the decoded viewpoint information based on the rendered image and the rendering incidental information. The transmission / reception unit 140 transmits the viewpoint buffer including the viewpoint buffer information and the viewpoint conversion image to the client device 200. The client device 200 includes a transmission / reception unit 210 and a decoding unit 220. The transmission / reception unit 210 receives the viewpoint conversion image and the viewpoint buffer information. The decoding unit 220 generates a display image based on the viewpoint conversion image and the viewpoint buffer information.
 これにより、情報処理システム10は、遅延をより低減した高画質な画像をユーザに提示することができ、ユーザに提供するサービスの品質劣化をより抑制することができる。 Thereby, the information processing system 10 can present the user with a high-quality image with further reduced delay, and can further suppress the deterioration of the quality of the service provided to the user.
 また、本明細書に記載された効果は、あくまで説明的または例示的なものであって限定的ではない。つまり、本開示に係る技術は、上記の効果とともに、または上記の効果に代えて、本明細書の記載から当業者には明らかな他の効果を奏しうる。 Further, the effects described in the present specification are merely explanatory or exemplary and are not limited. That is, the techniques according to the present disclosure may have other effects apparent to those skilled in the art from the description herein, in addition to or in place of the above effects.
 なお、本技術は以下のような構成も取ることができる。
(1)
 第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得し、
 前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成し、
 前記第1のレンダリング画像及び前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成する生成部と、
 前記視線情報及び前記第2のレンダリング画像を端末装置に送信する送信部と、
 を備える情報処理装置。
(2)
 前記生成部は、前記端末装置との通信で発生する遅延状況に応じて、前記視線情報に含める前記視点位置を決定する、(1)に記載の情報処理装置。
(3)
 前記生成部は、過去に取得した前記第1のレンダリング画像の視点位置に基づき、前記視線情報に含める前記視点位置を決定する、(1)又は(2)に記載の情報処理装置。
(4)
 前記生成部は、前記第1のレンダリング画像を含むコンテンツに応じて、前記視線情報に含める前記視点位置を決定する、(1)~(3)のいずれか1つに記載の情報処理装置。
(5)
 前記生成部は、前記第1のレンダリング画像のシーンに応じて、前記視線情報に含める前記視点位置を決定する、(1)~(4)のいずれか1つに記載の情報処理装置。
(6)
 前記レンダリング情報は、前記第1のレンダリング画像をレンダリングする時に生成される情報である、(1)~(5)のいずれか1つに記載の情報処理装置。
(7)
 前記レンダリング情報は、前記第1のレンダリング画像をレンダリングするためにあらかじめ設定されている情報である、(1)~(5)のいずれか1つに記載の情報処理装置。
(8)
 前記生成部は、前記視点位置の分布を表す視点関数、及び、前記視点位置に応じた前記視線方向を表す視線関数を算出し、算出した前記視点関数及び前記視線関数を用いて前記第2の数の前記視点位置及び前記視線方向を生成し、
 前記送信部は、前記視点関数及び前記視線関数に関する情報を前記視線情報として送信する、
 (1)~(7)のいずれか1つに記載の情報処理装置。
(9)
 第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報に基づいて生成される第2のレンダリング画像であって、前記第1の数より多い第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像と、前記第2の数の前記視点位置及び前記視線方向に関する視線情報と、を受信する受信部と、
 前記第2のレンダリング画像及び前記視線情報に基づき、表示画像を生成する生成部と、
 を備える情報処理装置。
(10)
 前記生成部は、ユーザの現在の前記視点位置及び前記視線方向に基づき、前記視線情報に含まれる少なくとも1つの前記視点位置及び前記視線方向を選択し、選択した前記視点位置及び前記視線方向から見た前記第2のレンダリング画像に基づいて前記表示画像を生成する、(9)に記載の情報処理装置。
(11)
 情報処理装置と、端末装置と、を備える情報処理システムであって、
 前記情報処理装置は、
 第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得し、
 前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成し、
 前記第1のレンダリング画像及び前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成する生成部と、
 前記視線情報及び前記第2のレンダリング画像を送信する送信部と、
 を備え、
 前記端末装置は、
 前記第2のレンダリング画像と、前記視線情報と、を受信する受信部と、
 前記第2のレンダリング画像及び前記視線情報に基づき、表示画像を生成する生成部と、
 を備える情報処理システム。
(12)
 第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得し、
 前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成し、
 前記第1のレンダリング画像及び前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成し、
 前記視線情報及び前記第2のレンダリング画像を、端末装置に送信する、
 情報処理方法。
The present technology can also have the following configurations.
(1)
The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
The line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
A generation unit that generates a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
A transmission unit that transmits the line-of-sight information and the second rendered image to the terminal device, and
Information processing device equipped with.
(2)
The information processing device according to (1), wherein the generation unit determines the viewpoint position to be included in the line-of-sight information according to a delay situation generated in communication with the terminal device.
(3)
The information processing apparatus according to (1) or (2), wherein the generation unit determines the viewpoint position to be included in the line-of-sight information based on the viewpoint position of the first rendered image acquired in the past.
(4)
The information processing apparatus according to any one of (1) to (3), wherein the generation unit determines the viewpoint position to be included in the line-of-sight information according to the content including the first rendered image.
(5)
The information processing apparatus according to any one of (1) to (4), wherein the generation unit determines the viewpoint position to be included in the line-of-sight information according to the scene of the first rendered image.
(6)
The information processing apparatus according to any one of (1) to (5), wherein the rendering information is information generated when the first rendered image is rendered.
(7)
The information processing apparatus according to any one of (1) to (5), wherein the rendering information is information preset for rendering the first rendered image.
(8)
The generation unit calculates a viewpoint function representing the distribution of the viewpoint position and a line-of-sight function representing the line-of-sight direction according to the viewpoint position, and uses the calculated viewpoint function and the line-of-sight function to obtain the second line-of-sight function. Generate the viewpoint position and the line-of-sight direction of the number,
The transmitting unit transmits information about the viewpoint function and the line-of-sight function as the line-of-sight information.
The information processing apparatus according to any one of (1) to (7).
(9)
The first rendered image viewed from the viewpoint position and the line-of-sight direction of the first number, and the second rendered image generated based on the rendering information used for rendering the first rendered image. Reception of receiving a second rendered image viewed from the viewpoint position and the line-of-sight direction of a second number larger than the first number, and line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number. Department and
A generation unit that generates a display image based on the second rendered image and the line-of-sight information,
Information processing device equipped with.
(10)
The generation unit selects at least one of the viewpoint positions and the line-of-sight directions included in the line-of-sight information based on the user's current viewpoint position and the line-of-sight direction, and views from the selected viewpoint position and the line-of-sight direction. The information processing apparatus according to (9), which generates the display image based on the second rendered image.
(11)
An information processing system including an information processing device and a terminal device.
The information processing device is
The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
The line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
A generation unit that generates a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
A transmission unit that transmits the line-of-sight information and the second rendered image, and
Equipped with
The terminal device is
A receiving unit that receives the second rendered image and the line-of-sight information.
A generation unit that generates a display image based on the second rendered image and the line-of-sight information,
Information processing system equipped with.
(12)
The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
The line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
Based on the first rendered image and the rendered information, a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number is generated.
The line-of-sight information and the second rendered image are transmitted to the terminal device.
Information processing method.
 10 情報処理システム
 100 サーバ装置
 110 レンダリング部
 120 多視点変換部
 130 エンコード部
 140、210 送受信部
 150 コマンド情報取得部
 200 クライアント装置
 220 デコード部
 230 入力情報取得部
 300 周辺機器
10 Information processing system 100 Server equipment 110 Rendering unit 120 Multi-viewpoint conversion unit 130 Encoding unit 140, 210 Transmission / reception unit 150 Command information acquisition unit 200 Client equipment 220 Decoding unit 230 Input information acquisition unit 300 Peripheral devices

Claims (12)

  1.  第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得し、
     前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成し、
     前記第1のレンダリング画像および前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成する生成部と、
     前記視線情報及び前記第2のレンダリング画像を端末装置に送信する送信部と、
     を備える情報処理装置。
    The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
    The line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
    A generation unit that generates a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
    A transmission unit that transmits the line-of-sight information and the second rendered image to the terminal device, and
    Information processing device equipped with.
  2.  前記生成部は、前記端末装置との通信で発生する遅延状況に応じて、前記視線情報に含める前記視点位置を決定する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the generation unit determines the viewpoint position to be included in the line-of-sight information according to a delay situation generated in communication with the terminal device.
  3.  前記生成部は、過去に取得した前記第1のレンダリング画像の視点位置に基づき、前記視線情報に含める前記視点位置を決定する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the generation unit determines the viewpoint position to be included in the line-of-sight information based on the viewpoint position of the first rendered image acquired in the past.
  4.  前記生成部は、前記第1のレンダリング画像を含むコンテンツに応じて、前記視線情報に含める前記視点位置を決定する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the generation unit determines the viewpoint position to be included in the line-of-sight information according to the content including the first rendered image.
  5.  前記生成部は、前記第1のレンダリング画像のシーンに応じて、前記視線情報に含める前記視点位置を決定する、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the generation unit determines the viewpoint position to be included in the line-of-sight information according to the scene of the first rendered image.
  6.  前記レンダリング情報は、前記第1のレンダリング画像をレンダリングする時に生成される情報である、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the rendering information is information generated when the first rendered image is rendered.
  7.  前記レンダリング情報は、前記第1のレンダリング画像をレンダリングするためにあらかじめ設定されている情報である、請求項1に記載の情報処理装置。 The information processing device according to claim 1, wherein the rendering information is information preset for rendering the first rendered image.
  8.  前記生成部は、前記視点位置の分布を表す視点関数、及び、前記視点位置に応じた前記視線方向を表す視線関数を算出し、算出した前記視点関数及び前記視線関数を用いて前記第2の数の前記視点位置及び前記視線方向を生成し、
     前記送信部は、前記視点関数及び前記視線関数に関する情報を前記視線情報として送信する、
     請求項1に記載の情報処理装置。
    The generation unit calculates a viewpoint function representing the distribution of the viewpoint position and a line-of-sight function representing the line-of-sight direction according to the viewpoint position, and uses the calculated viewpoint function and the line-of-sight function to obtain the second line-of-sight function. Generate the viewpoint position and the line-of-sight direction of the number,
    The transmitting unit transmits information about the viewpoint function and the line-of-sight function as the line-of-sight information.
    The information processing apparatus according to claim 1.
  9.  第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報に基づいて生成される第2のレンダリング画像であって、前記第1の数より多い第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像と、前記第2の数の前記視点位置及び前記視線方向に関する視線情報と、を受信する受信部と、
     前記第2のレンダリング画像及び前記視線情報に基づき、表示画像を生成する生成部と、
     を備える情報処理装置。
    The first rendered image viewed from the viewpoint position and the line-of-sight direction of the first number, and the second rendered image generated based on the rendering information used for rendering the first rendered image. Reception of receiving a second rendered image viewed from the viewpoint position and the line-of-sight direction of a second number larger than the first number, and line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number. Department and
    A generation unit that generates a display image based on the second rendered image and the line-of-sight information,
    Information processing device equipped with.
  10.  前記生成部は、ユーザの現在の前記視点位置及び前記視線方向に基づき、前記視線情報に含まれる少なくとも1つの前記視点位置及び前記視線方向を選択し、選択した前記視点位置及び前記視線方向から見たレンダリング画像に基づいて前記表示画像を生成する、請求項9に記載の情報処理装置。 The generation unit selects at least one of the viewpoint positions and the line-of-sight directions included in the line-of-sight information based on the user's current viewpoint position and the line-of-sight direction, and looks from the selected viewpoint position and the line-of-sight direction. The information processing apparatus according to claim 9, which generates the display image based on the rendered image.
  11.  情報処理装置と、端末装置と、を備える情報処理システムであって、
     前記情報処理装置は、
     第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得し、
     前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成し、
     前記第1のレンダリング画像および前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成する生成部と、
     前記視線情報及び前記第2のレンダリング画像を送信する送信部と、
     を備え、
     前記端末装置は、
     前記第2のレンダリング画像と、前記視線情報と、を受信する受信部と、
     前記第2のレンダリング画像及び前記視線情報に基づき、表示画像を生成するする生成部と、
     を備える情報処理システム。
    An information processing system including an information processing device and a terminal device.
    The information processing device is
    The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
    The line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
    A generation unit that generates a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number based on the first rendered image and the rendered information.
    A transmission unit that transmits the line-of-sight information and the second rendered image, and
    Equipped with
    The terminal device is
    A receiving unit that receives the second rendered image and the line-of-sight information.
    A generation unit that generates a display image based on the second rendered image and the line-of-sight information.
    Information processing system equipped with.
  12.  第1の数の視点位置及び視線方向から見た第1のレンダリング画像、及び、前記第1のレンダリング画像のレンダリングに使用したレンダリング情報を取得し、
     前記第1の数より多い第2の数の前記視点位置及び前記視線方向に関する視線情報を生成し、
     前記第1のレンダリング画像および前記レンダリング情報に基づき、前記第2の数の前記視点位置及び前記視線方向から見た第2のレンダリング画像を生成し、
     前記視線情報及び前記第2のレンダリング画像を、端末装置に送信する、
     情報処理方法。
    The first rendered image viewed from the first number of viewpoint positions and the line-of-sight direction, and the rendering information used for rendering the first rendered image are acquired, and the rendering information is acquired.
    The line-of-sight information regarding the viewpoint position and the line-of-sight direction of the second number, which is larger than the first number, is generated.
    Based on the first rendered image and the rendered information, a second rendered image viewed from the viewpoint position and the line-of-sight direction of the second number is generated.
    The line-of-sight information and the second rendered image are transmitted to the terminal device.
    Information processing method.
PCT/JP2021/042524 2020-12-07 2021-11-19 Information processing device, information processing system, and information processing method WO2022124047A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2020202607 2020-12-07
JP2020-202607 2020-12-07

Publications (1)

Publication Number Publication Date
WO2022124047A1 true WO2022124047A1 (en) 2022-06-16

Family

ID=81974406

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/042524 WO2022124047A1 (en) 2020-12-07 2021-11-19 Information processing device, information processing system, and information processing method

Country Status (1)

Country Link
WO (1) WO2022124047A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020010301A (en) * 2018-07-12 2020-01-16 キヤノン株式会社 Information processing apparatus, control method of the same, and program
JP2020095741A (en) * 2020-03-03 2020-06-18 キヤノン株式会社 Image processing apparatus, image processing system, image processing method, and program
JP2021114216A (en) * 2020-01-21 2021-08-05 Kddi株式会社 Information system, terminal, server and program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020010301A (en) * 2018-07-12 2020-01-16 キヤノン株式会社 Information processing apparatus, control method of the same, and program
JP2021114216A (en) * 2020-01-21 2021-08-05 Kddi株式会社 Information system, terminal, server and program
JP2020095741A (en) * 2020-03-03 2020-06-18 キヤノン株式会社 Image processing apparatus, image processing system, image processing method, and program

Similar Documents

Publication Publication Date Title
US11909984B2 (en) Video encoding and decoding for cloud gaming
US10354430B2 (en) Image update method, system, and apparatus
US11196973B2 (en) Providing apparatus, providing method and computer readable storage medium for performing processing relating to a virtual viewpoint image
KR102562877B1 (en) Methods and apparatus for distribution of application computations
KR102441514B1 (en) Hybrid streaming
TW201921921A (en) Processing of 3D image information based on texture maps and meshes
EP4037789A1 (en) Game application providing scene change hint for encoding at a cloud gaming server
US8860720B1 (en) System and method for delivering graphics over network
Joachimczak et al. Real-time mixed-reality telepresence via 3D reconstruction with HoloLens and commodity depth sensors
JP6306089B2 (en) System and method for transmitting media over a network
JP2020522801A (en) Method and system for creating a virtual projection of a customized view of a real world scene for inclusion in virtual reality media content
CN111464828A (en) Virtual special effect display method, device, terminal and storage medium
WO2022124047A1 (en) Information processing device, information processing system, and information processing method
US20220084282A1 (en) Image processing device, image generation method, and image processing method
JPWO2014171066A1 (en) 3D image display system, 3D image display system server, and 3D image display method
US20220321861A1 (en) Method and apparatus for providing conversational services in mobile communication system
WO2023079987A1 (en) Distribution device, distribution method, and program
EP3564905A1 (en) Conversion of a volumetric object in a 3d scene into a simpler representation model
US8897821B2 (en) Method for providing visual effect messages and associated communication system and transmitting end
US20230316656A1 (en) Information processing device, information processing method, and program
WO2022224964A1 (en) Information processing device and information processing method
US20230412724A1 (en) Controlling an Augmented Call Based on User Gaze
CN118160315A (en) Distribution device, distribution method, and program
Aumüller D5. 3.4–Remote hybrid rendering: revision of system and protocol definition for exascale systems

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21903151

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21903151

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP