WO2024087883A1 - 视频画面渲染方法、装置、设备和介质 - Google Patents

视频画面渲染方法、装置、设备和介质 Download PDF

Info

Publication number
WO2024087883A1
WO2024087883A1 PCT/CN2023/116670 CN2023116670W WO2024087883A1 WO 2024087883 A1 WO2024087883 A1 WO 2024087883A1 CN 2023116670 W CN2023116670 W CN 2023116670W WO 2024087883 A1 WO2024087883 A1 WO 2024087883A1
Authority
WO
WIPO (PCT)
Prior art keywords
rendering
sub
video
picture
pictures
Prior art date
Application number
PCT/CN2023/116670
Other languages
English (en)
French (fr)
Inventor
李想
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Publication of WO2024087883A1 publication Critical patent/WO2024087883A1/zh

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • A63F13/525Changing parameters of virtual cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44012Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving rendering scenes according to scene graphs, e.g. MPEG-4 scene graphs

Definitions

  • the present application relates to video processing technology, and in particular to a video picture rendering method, device, equipment and medium.
  • a video picture rendering method, apparatus, device and medium are provided.
  • the present application provides a video picture rendering method, which is executed by a computer device, comprising:
  • the video frame is split to obtain a plurality of video sub-pictures
  • the plurality of video sub-pictures are synchronously rendered by using a plurality of pre-set rendering machines to obtain a rendering picture corresponding to the video frame picture.
  • the present application provides a video picture rendering device, the device comprising:
  • a determination module is used to determine a simulated camera, where the simulated camera is obtained by simulating a virtual camera of a rendering engine; determine a mesh patch, where the mesh patch matches the size of a virtual sensor in the virtual camera; determine a positional relationship between a viewpoint in the simulated camera and each vertex in the mesh patch, and determine a rendering area according to the positional relationship; obtain a video frame to be rendered in the rendering area; obtain a video frame to be rendered in the rendering area;
  • a splitting module used for splitting the video frame according to the grid surface to obtain multiple video sub-pictures
  • the rendering module is used to synchronously render the multiple video sub-pictures through a plurality of pre-set rendering machines to obtain a rendering picture surface corresponding to the video frame picture.
  • the present application further provides a computer device, including a memory and a processor, wherein the memory stores computer-readable instructions, and the processor executes the steps of the method embodiments of the present application when executing the computer-readable instructions.
  • the present application further provides a computer-readable storage medium storing computer-readable instructions, which, when executed by a processor, perform the steps of the method embodiments of the present application.
  • the present application also provides a computer program product, comprising computer-readable instructions, which, when executed by a processor, perform the steps of the method embodiments of the present application.
  • FIG1 is a diagram of an application environment of a video image rendering method according to an embodiment
  • FIG2 is a schematic diagram of a flow chart of a method for rendering a video image in one embodiment
  • FIG3 is a schematic diagram of the positional relationship between viewpoints and mesh patches in one embodiment
  • FIG4 is a schematic diagram showing the principle of determining the distance between a viewpoint and a mesh patch based on the focal length of a virtual camera in one embodiment
  • FIG5 is a schematic diagram showing the principle of determining the position of a simulated camera based on the position of a virtual camera in one embodiment
  • FIG6 is a schematic diagram showing the principle of segmentation modeling and distributed rendering based on the size of a virtual sensor in one embodiment
  • FIG7 is a schematic diagram of a mapping relationship between a rendering engine, a mesh sub-face, and a screen viewport in one embodiment
  • FIG8 is a schematic diagram of a hardware environment constructed for video image rendering in one embodiment
  • FIG9 is a schematic diagram of a flow chart of a video image rendering method in another embodiment
  • FIG10 is a schematic diagram of an application scenario of a video image rendering method according to an embodiment
  • FIG11 is a schematic diagram of a flow chart of a video picture rendering method in yet another embodiment
  • FIG12 is a structural block diagram of a video picture rendering device in one embodiment
  • FIG. 13 is a diagram showing the internal structure of a computer device in one embodiment.
  • the video rendering method provided in this application can be applied to the application environment shown in Figure 1.
  • the terminal 102 communicates with the server 104 through the network.
  • the data storage system can store the data that the server 104 needs to process.
  • the data storage system can be integrated on the server 104, or it can be placed on the cloud or other servers.
  • the terminal 102 can be, but is not limited to, various desktop computers, laptops, smart phones, tablet computers, Internet of Things devices and portable wearable devices.
  • the Internet of Things devices can be smart speakers, smart TVs, smart air conditioners, smart car-mounted devices, etc.
  • Portable wearable devices can be smart watches, smart bracelets, head-mounted devices, etc.
  • the server 104 can be an independent physical server, or a server cluster or distributed system composed of multiple physical servers, or a cloud server that provides basic cloud computing services such as cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, CDN, and big data and artificial intelligence platforms.
  • the terminal 102 and the server 104 can be directly or indirectly connected by wired or wireless communication, and this application is not limited here.
  • the server 104 may determine a simulated camera, which is obtained by simulating a virtual camera of a rendering engine.
  • the server 104 may determine a mesh patch, which matches the size of a virtual sensor in a virtual camera.
  • the server 104 may determine the positional relationship between the viewpoint in the simulated camera and each vertex in the mesh patch, and determine a rendering area based on the positional relationship.
  • the server 104 may obtain a video frame to be rendered in the rendering area.
  • the server 104 may split the video frame according to the mesh patch to obtain a plurality of video sub-pictures.
  • the server 104 may synchronously render a plurality of video sub-pictures through a plurality of pre-set rendering machines to obtain a rendering picture corresponding to the video frame.
  • the server 104 can send the synthesized rendering picture to the terminal 102 for display. This embodiment does not limit this. It is understood that the application scenario in FIG. 1 is only for illustrative purposes and is not limited thereto.
  • a video image rendering method is provided.
  • the method can be applied to a computer device, which can be a terminal or a server.
  • the method can be executed by the terminal or the server itself, or can be implemented through interaction between the terminal and the server.
  • This embodiment is described by taking the method applied to a computer device as an example, and includes the following steps:
  • Step 202 determine the simulated camera, which is obtained by simulating the virtual camera of the rendering engine. Determine the mesh patch, which matches the size of the virtual sensor in the virtual camera. Determine the viewpoint in the simulated camera and The positional relationship between each vertex in the mesh patch determines the rendering area based on the positional relationship.
  • the rendering engine is a three-dimensional real-time rendering engine for rendering images, and may be a game engine, such as Unreal Engine.
  • a virtual camera is a virtual camera in a rendering engine.
  • a virtual camera can be used to implement the functions of a physical camera.
  • a simulated camera is a camera simulated for a virtual camera.
  • a simulated camera may have the functions of a simulated virtual camera.
  • the simulated camera has a viewpoint and a mesh patch.
  • the viewpoint can be called an observation point. By observing the same object from different viewpoints, you can get pictures of the same object from different perspectives.
  • the mesh patch is a three-dimensional patch constructed based on the size of the virtual sensor in the virtual camera, and the three-dimensional patch is located in the three-dimensional virtual scene.
  • the position of the viewpoint of the simulated camera can be determined according to the position of the simulated camera. For example, the position of the simulated camera can be directly used as the position of the viewpoint of the simulated camera, or the position of the simulated camera can be offset according to a preset offset relationship to obtain the position of the viewpoint.
  • the virtual sensor is a virtual sensor in the virtual camera.
  • the virtual sensor has the function of a physical sensor.
  • the size of the virtual sensor can be expressed in physical size, for example, the size of the virtual sensor can be expressed in physical size as 23.76mm*13.365mm, where mm represents millimeters.
  • the mesh patch matches the size of the virtual sensor in the virtual camera, and the mesh patch can be the same size as the virtual sensor in the virtual camera, or keep the same ratio.
  • the rendering area is the area inside the cone when the viewpoint is connected to each vertex of the mesh surface to form a cone.
  • the rendering area is used to determine the area of the video frame to be rendered.
  • the rendering area can be the area inside the cone formed from the viewpoint to the edge of the mesh patch.
  • the rendering area can be the space occupied in the three-dimensional virtual scene, which is used to render the two-dimensional video screen.
  • the vertex is the point used to determine the edge of the mesh patch, and the same vertex connects different edges of the mesh patch.
  • the mesh patch is a rectangle
  • the four right-angled vertices of the rectangle are the vertices of the mesh patch.
  • the rendering area is the area inside the vertebral body.
  • the rendering area can be used to accurately determine the content to be rendered in the three-dimensional virtual scene, and then accurately generate the video frame to be rendered, thereby preparing to ensure the quality of the final rendered image.
  • the viewpoint and the vertices of the mesh patch can determine the rendering area. Therefore, the computer device can determine the positional relationship between the viewpoint and each vertex of the mesh patch to determine the rendering area.
  • the positional relationship can be the relative position between the viewpoint and the vertex, such as the offset of the vertex relative to the viewpoint on the plane where the mesh patch is located.
  • the positional relationship can also be represented by the position of the viewpoint and the vertex in the three-dimensional virtual scene.
  • the computer device can determine the position of the viewpoint and the position of each vertex of the mesh surface, thereby determining the positional relationship between the viewpoint and the vertices of the mesh surface, and then determine the rendering area using the positional relationship between the viewpoint and the vertices of the mesh surface.
  • the computer device can determine a cone based on the viewpoint and each vertex in the mesh surface, and the area inside the cone is the rendering area.
  • the viewpoint connects the four vertices of the mesh surface to form a cone, and the inside of the cone is the rendering area.
  • the computer device can connect the viewpoint O in the simulated camera to each vertex (ie, A, B, C, D) in the mesh patch ABCD to obtain a tetrahedron O-ABCD, and the area inside the cone is the rendering area.
  • Step 204 Obtain the video frame to be rendered in the rendering area.
  • the video frame picture is an image of a video frame in a video or video stream.
  • the video frame picture can be a real-time picture, and specifically can be a picture of a video frame corresponding to the current moment in the real-time video stream. It can be understood that the video stream includes multiple video frames, and the multiple video frames correspond to multiple video frame pictures.
  • the position of the simulated camera may be the position of the simulated camera in the three-dimensional virtual scene.
  • the position of the simulated camera corresponds to the position of the viewpoint of the simulated camera. Therefore, the computer device may determine the position of the viewpoint according to the position of the camera, thereby determining the spatial position of the rendering area in the three-dimensional virtual scene, and thus determining the video frame based on the content in the rendering area.
  • the computer device may project the three-dimensional content in the rendering area onto a two-dimensional mesh patch to obtain a video frame to be rendered.
  • obtaining the video frame to be rendered in the rendering area includes: obtaining the position of the simulated camera, and determining the video frame to be rendered in the rendering area according to the position of the virtual camera.
  • the computer device can obtain the current position of the simulated camera, and thus determine the real-time video frame to be rendered in the rendering area according to the current position of the virtual camera.
  • the computer device can obtain the current position of the simulated camera, and then determine the current position of the simulated camera's viewpoint based on the current position of the simulated camera, thereby determining the content of the rendering area in the three-dimensional virtual scene at the current moment, and then determine the real-time video frame to be rendered in the rendering area based on the content.
  • the position of the simulated camera can be used to accurately determine the video frame to be rendered in the rendering area, thereby preparing for ensuring the quality of the final rendered image.
  • the computer device may obtain the position of the virtual camera and determine the position of the simulated camera according to the position of the virtual camera. It is understood that the computer device may use the position of the virtual camera as the position of the simulated camera. The computer device may also adjust the position of the virtual camera, such as offsetting it according to a preset offset relationship, and use the offset position as the position of the simulated camera.
  • Step 206 split the video frame according to the grid surface patches to obtain a plurality of video sub-pictures.
  • the video sub-picture is a video picture obtained by splitting a video frame into regions. It can be understood that the multiple video sub-pictures obtained by splitting are independent of each other and can be merged into a video frame.
  • the computer device can determine the size of the video frame according to the size of the grid surface, determine the splitting method according to the size of the grid surface, and split the video frame according to the splitting method to obtain multiple video sub-pictures.
  • the video frame can be split according to a preset splitting method or a splitting method determined in real time.
  • the splitting method can be equal or unequal. If it is unequal, some video sub-pictures can have the same size, some video sub-pictures can have different sizes, or all video sub-pictures can have different sizes.
  • the computer device can spatially split the rendering area corresponding to each real-time video frame according to the grid surface to obtain multiple sub-areas. Furthermore, the computer device can respectively determine the real-time picture content to be rendered in each sub-area, and obtain multiple video sub-pictures based on the real-time picture content to be rendered in each sub-area. Splitting the video frame in space means splitting the video frame from the picture dimension, and the split video sub-pictures can be rendered separately.
  • Step 208 using a plurality of pre-set rendering machines, synchronously render a plurality of video sub-pictures to obtain a rendering picture corresponding to the video frame picture.
  • the rendering machine may be a computer program set in a computer device for performing a rendering task.
  • the rendering machine may also be a computer device for performing a rendering task.
  • the synchronous rendering of multiple video sub-pictures by multiple pre-set rendering machines is that multiple rendering machines perform rendering tasks in parallel to render multiple video sub-pictures in parallel.
  • the number of rendering machines may be the same as the number of video sub-pictures, or may be different from the number of video sub-pictures.
  • the rendered picture is a picture obtained after rendering the video frame picture.
  • the rendered picture may be displayed by a display unit of a computer device or an independent display device.
  • multiple rendering machines can synchronously render multiple video sub-pictures to obtain multiple rendered sub-pictures corresponding to the multiple video sub-pictures one by one, and the multiple rendered sub-pictures constitute a rendered picture corresponding to the video frame picture.
  • Each rendered sub-picture is a picture obtained by rendering the corresponding video sub-picture.
  • Each rendering machine can be used to render at least one video sub-picture.
  • the rendering machine has a network address, and the rendering machine can render a video sub-picture having a preset mapping relationship with its network address.
  • multiple rendering machines can synchronously render multiple video sub-pictures to obtain multiple rendering sub-pictures corresponding to the multiple video sub-pictures one by one, and splice the multiple rendering sub-pictures according to the positional relationship of the multiple video sub-pictures to obtain a rendering picture corresponding to the video frame picture.
  • the computer device executing steps 202 to 206 may be one of the multiple rendering machines, which may be referred to as the master rendering machine.
  • the master rendering machine may coordinate the slave rendering machines among the multiple rendering machines, so that the master rendering machine and the slave rendering machines synchronously render multiple video sub-pictures.
  • the computer device executing steps 202-206 may be a device independent of any rendering machine. At this time, the computer device can instruct multiple rendering machines to synchronously render multiple video sub-pictures.
  • the position of the video sub-picture rendered by each rendering machine relative to the video frame can be fixed. For example, there are 4 rendering machines, numbered 1, 2, 3, and 4, and each video frame is split into 4 video sub-pictures, numbered I, II, III, and IV according to the position, then rendering machine No. 1 can be fixed to render the video sub-picture at position I, rendering machine No. 2 can be fixed to render the video sub-picture at position II, rendering machine No. 3 can be fixed to render the video sub-picture at position III, and rendering machine No. 4 can be fixed to render the video sub-picture at position IV.
  • At least one rendering machine among the multiple rendering machines can render at least two video sub-pictures at the same time.
  • the rendering machine that renders at least two video sub-pictures at the same time can be a fixed rendering machine among the multiple rendering machines, such as a rendering machine with stronger computing performance than the rendering machine that renders one video sub-picture at the same time.
  • the rendering machine that renders at least two video sub-pictures at the same time can be a rendering machine that renders one video sub-picture at the same time at the next adjacent moment.
  • the rendering machine that renders at least two video sub-pictures at the same time can be selected from multiple rendering machines in turn, or can be selected according to the load situation.
  • each rendering machine when the number of rendering machines is greater than the number of video sub-pictures, each rendering machine can only render one video sub-picture at a time.
  • the rendering machines that do not render video sub-pictures are idle rendering machines, and the idle rendering machines at different times can be the same or different.
  • the above-mentioned video screen rendering method splits the video frame screen through a grid patch that matches the size of the virtual sensor in the virtual camera, and synchronously renders the split sub-screens through a plurality of pre-set rendering machines, so as to efficiently obtain a high-quality rendering screen, and meet the rendering requirements in scenes with high picture quality and rendering efficiency.
  • the larger the screen size the more prominent the effect of the above-mentioned rendering screen rendering method is compared with the traditional method.
  • the above-mentioned video screen rendering method can better meet the high real-time requirements of video screen rendering.
  • the above-mentioned video screen rendering method has a particularly prominent effect.
  • the position of the virtual camera in the three-dimensional virtual scene is determined according to the position of the virtual camera in the three-dimensional virtual scene.
  • Determining the video frame to be rendered in the rendering area according to the position of the virtual camera includes: determining the real-time scene content in the rendering area in the three-dimensional virtual scene according to the position of the simulated camera, and obtaining the real-time video frame to be rendered.
  • the position of the simulated camera is determined according to the position of the virtual camera in the three-dimensional virtual scene; according to the position of the simulated camera, the real-time video frame to be rendered in the rendering area is determined, including: according to the position of the simulated camera, the real-time scene content located in the rendering area in the three-dimensional virtual scene is determined to obtain the real-time video frame to be rendered.
  • the computer device can determine the position of the simulation camera according to the position of the virtual camera in the three-dimensional virtual scene, and determine the real-time scene content in the rendering area of the three-dimensional virtual scene according to the position of the simulation camera, and obtain the real-time video frame to be rendered. It can be understood that the real-time scene content in the rendering area of the three-dimensional virtual scene can be rendered, that is, it is visible. The real-time scene content outside the rendering area of the three-dimensional virtual scene will not be rendered, that is, it is invisible.
  • the real-time scene content in the rendering area of the three-dimensional virtual scene is determined, and the real-time video frame to be rendered is obtained, which can improve the rendering speed of the video frame.
  • a mesh patch includes multiple mesh sub-patches, and the common vertices of two adjacent mesh sub-patches in the multiple mesh sub-patches have the same coordinates.
  • a video frame is split to obtain multiple video sub-patterns, including: for each video frame, a rendering area corresponding to the video frame is spatially split according to the common vertices in the multiple mesh sub-patches to obtain multiple sub-areas; and real-time scene content in each sub-area in a three-dimensional virtual scene is determined, and the real-time scene content in each sub-area is respectively determined as a video sub-area.
  • the shared vertices are the mesh vertices shared by two mesh sub-faces, and the sub-regions are the regions obtained by spatially splitting the rendering region.
  • the computer device can spatially split the rendering area corresponding to the video frame according to the common vertices in the multiple mesh sub-faces to obtain multiple sub-areas. Furthermore, the computer device can determine the real-time scene content in each sub-area in the three-dimensional virtual scene, and determine the real-time scene content in each sub-area after the split as the video sub-picture.
  • the rendering machine has a network address, the network address of the rendering machine corresponds to the mesh sub-face, and the mesh sub-face corresponds to the video sub-picture.
  • the rendering machine can render the video sub-picture corresponding to the mesh sub-face that has a preset mapping relationship with its network address.
  • the mesh patch includes two mesh sub-patterns, namely, mesh sub-pattern AEFD (i.e., mesh sub-pattern 1) and mesh sub-pattern EBCF (i.e., mesh sub-pattern 2).
  • mesh sub-pattern AEFD and mesh sub-pattern EBCF are two adjacent mesh sub-patterns, which have two common vertices, namely, vertex E and vertex F.
  • the computer device can spatially split the rendering area corresponding to the video frame according to the common vertices E and F in the mesh sub-pattern AEFD and the mesh sub-pattern EBCF, and obtain two sub-areas, namely, tetrahedron O-AEFD and tetrahedron O-EBCF. It can be understood that the real-time scene content in the tetrahedron O-AEFD and the tetrahedron O-EBCF is the video sub-picture.
  • the rendering area corresponding to the video frame is spatially split by using the common vertices in multiple mesh sub-faces, which can ensure that subsequent seamless splicing can eventually synthesize a complete picture. Furthermore, by determining the real-time scene content in each of the split sub-areas as the video sub-picture, the acquisition accuracy of the video sub-picture can be improved.
  • the video image rendering method further includes: determining the focal length of the virtual camera; determining the distance between the viewpoint in the simulated camera and the mesh patch based on the focal length; and determining the positional relationship between the viewpoint and each vertex in the mesh patch based on the distance.
  • the computer device can determine the distance between the viewpoint and the mesh surface according to the focal length of the virtual camera in the rendering engine; and obtain the relative position relationship between the viewpoint and each vertex in the mesh surface according to the distance.
  • the computer device can determine the focal length of the virtual camera, and determine the distance between the viewpoint and the mesh surface according to the focal length of the virtual camera in the rendering engine. Furthermore, the computer device can obtain the relative position relationship between the viewpoint and each vertex in the mesh surface according to the distance.
  • the computer device may calculate the focal length of the virtual camera in the rendering engine, and use the calculated distance as the distance between the viewpoint and the mesh patch.
  • the computer device may directly use the focal length of the virtual camera in the rendering engine as the distance between the viewpoint and the mesh patch.
  • the simulated camera is obtained by simulating the virtual camera based on the display component, and the distance between the viewpoint and the mesh patch in the simulated camera is determined based on the focal length of the virtual camera.
  • the display component is a component used in the rendering engine to render the picture.
  • the computer device can obtain the current focal length of the virtual camera, and use the obtained current focal length of the virtual camera as the distance between the viewpoint and the mesh surface in the simulated camera. It can be understood that once the focal length of the virtual camera changes, the distance between the viewpoint and the mesh surface in the simulated camera will also change accordingly.
  • the distance between the viewpoint and the mesh patch is determined by the focal length of the virtual camera in the rendering engine, and the relative position between the viewpoint and each vertex in the mesh patch is obtained based on the distance. This can improve the accuracy of the relative position between the viewpoint and each vertex in the mesh patch, thereby determining a more accurate rendering area.
  • the simulated camera is simulated by a display component
  • the video image rendering method further includes: determining the position of the virtual camera; assigning the position of the virtual camera to the display component to obtain the position of the simulated camera.
  • the display component is a component used in the rendering engine to render the image.
  • the simulated camera is simulated by a display component; the display component is the parent of the viewpoint and the mesh patch; the video screen rendering method further includes: determining the position of the virtual camera; assigning the position of the virtual camera to the display component to obtain the position of the simulated camera.
  • the display component is the parent of the viewpoint and the mesh patch, so the display component can have the attributes of the viewpoint and the mesh patch, and the position of the virtual camera is assigned to the display component, and the position of the simulated camera, the viewpoint and the mesh patch can all be determined.
  • the computer device can simulate the virtual camera through the display component to obtain the simulated camera.
  • the computer device can determine the position of the virtual camera and assign the position of the virtual camera to the display component to obtain the position of the simulated camera.
  • the position of the simulated camera in the three-dimensional virtual scene is determined based on the position of the virtual camera in the three-dimensional virtual scene.
  • the computer device can obtain the position of the virtual camera and use the obtained position of the virtual camera as the position of the simulated camera. It can be understood that once the position of the virtual camera in the three-dimensional virtual scene changes, the position of the simulated camera in the three-dimensional virtual scene will also change accordingly.
  • the position of the virtual camera by assigning the position of the virtual camera to the display component, the position of the simulated camera can be obtained, and then the simulated camera can simulate the movement of the virtual camera, so that a more accurate video frame image can be determined.
  • multiple video sub-screens are synchronously rendered through multiple pre-set rendering machines to obtain rendering screens corresponding to video frame screens, including: synchronously rendering multiple video sub-screens through multiple pre-set rendering machines to obtain rendering sub-screens corresponding to the multiple video sub-screens; synthesizing the rendering sub-screens corresponding to the multiple video sub-screens to obtain the rendering screen corresponding to the video frame screen.
  • the computer device may be pre-set with multiple rendering machines, and through the pre-set multiple rendering machines, the computer device may synchronously render multiple video sub-pictures to obtain rendered sub-pictures corresponding to the multiple video sub-pictures. It can be understood that each rendering machine may be responsible for rendering at least one video sub-picture.
  • multiple rendering machines can respectively receive synchronous rendering signals and rendering instructions, and control the multiple rendering machines to synchronously render multiple video sub-pictures through the synchronous rendering signals and rendering instructions to obtain rendering sub-pictures corresponding to the multiple video sub-pictures.
  • the picture synthesizer can be a computer program or computer device that synthesizes the rendering sub-pictures corresponding to the multiple video sub-pictures to obtain the rendering picture corresponding to the video frame picture.
  • the picture synthesizer can be one of the multiple rendering machines, such as the main rendering machine.
  • the picture synthesizer can also be a computer device independent of any rendering machine.
  • a picture synthesizer may be pre-deployed in the computer device.
  • the computer device may receive rendering sub-pictures corresponding to multiple video sub-pictures output by each rendering machine through the picture synthesizer, and synthesize the rendering sub-pictures corresponding to the multiple video sub-pictures to obtain a rendering picture corresponding to the video frame picture.
  • the size of the virtual sensor is represented by a physical size
  • the video image rendering method further includes: modeling according to the size of the virtual sensor and the number of rendering machines to obtain a plurality of grid sub-faces.
  • the number of mesh sub-patches is consistent with the number of rendering machines; and the sum of the areas of the multiple mesh sub-patches is consistent with the area of the virtual sensor.
  • the size of the grid sub-pattern is obtained by dividing the physical size of the virtual sensor.
  • the size of the virtual sensor is expressed in physical size as: 23.76 mm*13.365 mm, and the number of rendering machines is 2, so the computer device can model two grid sub-surfaces with physical sizes of 11.88 mm*13.365 mm.
  • the computer device can synchronously render the video sub-pictures corresponding to the two grid sub-pictures through rendering machine 1 and rendering machine 2, respectively, to obtain rendering sub-picture 1 and rendering sub-picture 2.
  • the computer device may pre-set the mapping relationship between the network address of the rendering machine, the grid sub-face, and the screen viewport in the rendering machine. Specifically, the computer device may bind the network address 192.168.0.6 of the rendering machine 1, the grid sub-face 1, and the screen viewport 1 in the rendering machine 1. The computer device may bind the network address 192.168.0.7 of the rendering machine 2, the grid sub-face 2, and the screen viewport 2 in the rendering machine 2, so as to render the video sub-picture corresponding to the grid sub-face 1 by the rendering machine 1, and render the video sub-picture corresponding to the grid sub-face 2 by the rendering machine 2.
  • the screen viewport corresponds to the rendering machine, and the rendering machine corresponds to the split video sub-picture. The screen viewport is used to display the rendered sub-picture obtained by rendering the corresponding video sub-picture.
  • multiple grid sub-planes are obtained by modeling the physical size of the virtual sensor and the number of rendering machines, so that each rendering machine is responsible for rendering a video sub-screen corresponding to a corresponding grid sub-plane, thereby improving the rendering efficiency of the video sub-screen.
  • the multiple rendering machines include a master rendering machine and at least one slave rendering machine, and a synchronization card is deployed in the master rendering machine; multiple video sub-pictures are synchronously rendered through the pre-set multiple rendering machines to obtain rendering sub-pictures corresponding to the multiple video sub-pictures, including: receiving a synchronization rendering signal generated by a synchronization signal generator based on a preset frame rate through the synchronization card in the master rendering machine; synchronizing the synchronization rendering signal to the slave rendering machine through the synchronization card; controlling the master rendering machine and the slave rendering machine to synchronously render the multiple video sub-pictures respectively through the synchronization rendering signals respectively received by the master rendering machine and the slave rendering machine, to obtain rendering sub-pictures respectively corresponding to the multiple video sub-pictures.
  • the synchronous rendering signal is a synchronization signal used to instruct multiple rendering machines to synchronously render multiple video sub-pictures.
  • the synchronization signal generator is a phase synchronization signal generator. Multiple rendering machines can achieve phase-level time alignment based on the synchronous rendering signal.
  • the preset frame rate is a pre-set frame rate that can be set as needed. Within the frame rate range that can be perceived by the naked eye, the higher the preset frame rate, the higher the quality and the higher the hardware performance requirements.
  • the multiple rendering machines include a master rendering machine and at least one slave rendering machine, wherein a synchronization card is deployed in the master rendering machine.
  • a synchronization signal generator is also deployed in the computer device, and the synchronization signal generator can generate a synchronization rendering signal based on a preset frame rate.
  • the computer device can receive the synchronous rendering signal generated by the synchronous signal generator based on the preset frame rate through the synchronization card in the master rendering machine.
  • the master rendering machine can synchronize the synchronous rendering signal to each slave rendering machine through the synchronization card.
  • the computer device can control the master rendering machine and the slave rendering machine to synchronously render multiple video sub-pictures respectively through the synchronous rendering signals received by the master rendering machine and the slave rendering machine, and the rendering instructions sent by the master rendering machine, to obtain rendering sub-pictures corresponding to the multiple video sub-pictures.
  • the synchronization card in the master rendering machine receives the synchronization rendering signal generated by the synchronization signal generator based on the preset frame rate, and the synchronization rendering signal is synchronized to the slave rendering machine through the synchronization card, so that each rendering machine can receive the same synchronization rendering signal. Furthermore, through the synchronization rendering signals received by the master rendering machine and the slave rendering machine respectively, and the rendering instructions sent by the master rendering machine, the master rendering machine and the slave rendering machine are controlled to synchronously render multiple video sub-pictures respectively, and the rendering sub-pictures corresponding to the multiple video sub-pictures are obtained, which can avoid the problem of video picture tearing, thereby further improving the quality of the final rendered video picture.
  • rendering sub-pictures corresponding to a plurality of video sub-pictures are synthesized to obtain a rendering picture corresponding to a video frame picture, including: when a video signal acquisition card receives a synchronization acquisition signal generated by a synchronization signal generator, synchronously acquiring rendering sub-picture signals corresponding to a plurality of video sub-pictures through the video signal acquisition card; and synthesizing the synchronously acquired rendering sub-picture signals to obtain a rendering picture corresponding to the video frame picture.
  • the rendered sub-picture corresponds to a rendered sub-picture signal.
  • the synchronous acquisition signal is a synchronous signal for instructing multiple video signal acquisition cards to synchronously acquire multiple rendered sub-picture signals.
  • the computer device is also equipped with a video signal acquisition card.
  • the synchronization signal generator may generate a synchronization acquisition signal based on a preset frame rate, and the video signal acquisition card may receive the synchronization acquisition signal generated by the synchronization signal generator.
  • the computer device may synchronously acquire rendering sub-picture signals corresponding to a plurality of video sub-pictures through the video signal acquisition card.
  • the computer device may synthesize the rendering sub-picture signals corresponding to the plurality of video sub-pictures acquired synchronously, and obtain a rendering picture corresponding to the video frame picture.
  • the rendering sub-screen signals corresponding to the multiple video sub-screens are synchronously acquired through the video signal acquisition card, and the rendering sub-screen signals corresponding to the multiple synchronously acquired video sub-screens are synthesized to obtain the rendering screen corresponding to the video frame screen.
  • the problem of video screen tearing can be further avoided, thereby further improving the quality of the final rendered video screen.
  • the video screen rendering method further includes: when the signal format of the rendered sub-picture signal obtained by synchronous rendering is inconsistent with the signal format specified by the video signal acquisition card, after receiving the synchronous conversion signal through the format converter, formatting the rendered sub-picture signals corresponding to the plurality of video sub-pictures respectively through the format converter.
  • the rendered sub-picture signal is converted synchronously to obtain the signal format consistent with that specified by the video signal acquisition card for synchronous acquisition by the video signal acquisition card.
  • the synchronous conversion signal is a synchronous signal used to instruct multiple format converters to synchronously convert the formats of multiple rendered sub-picture signals.
  • the computer device is also equipped with a plurality of format converters, and the number of format converters can be consistent with the number of rendering machines.
  • the synchronization signal generator generates a synchronization conversion signal based on a preset frame rate, and the format converter can receive the synchronization conversion signal generated by the synchronization signal generator.
  • the computer device can perform format synchronization conversion on the rendered sub-picture signals corresponding to the plurality of video sub-pictures through the format converter after receiving the synchronization conversion signal through the format converter, and obtain the rendered sub-picture signal consistent with the signal format specified by the video signal acquisition card, so as to provide the video signal acquisition card with synchronous acquisition.
  • the number of format converters is consistent with the number of rendering machines, and each format converter is responsible for the conversion processing of the rendering sub-picture signal output by the corresponding rendering machine.
  • the computer device can synchronously convert the format of the rendering sub-picture signals corresponding to the multiple video sub-pictures through each format converter to obtain the rendering sub-picture signal consistent with the signal format specified by the video signal acquisition card, so as to provide the video signal acquisition card with synchronous acquisition.
  • the signal format of the rendering sub-picture signal output by each rendering machine is HDMI (High Definition Multimedia Interface) format
  • the signal format specified by the video signal acquisition card is SDI (Serial Digital Interface) format.
  • the computer device can perform format synchronous conversion on the rendering sub-picture signals corresponding to multiple HDMI format video sub-pictures through the format converter to obtain rendering sub-picture signals consistent with the SDI format specified by the video signal acquisition card for synchronous acquisition by the video signal acquisition card.
  • the signal format of the rendering sub-picture signal output by each rendering machine is in DP format
  • the signal format specified by the video signal acquisition card is in SDI format.
  • the computer device can perform format synchronous conversion on the rendering sub-picture signals corresponding to multiple DP (Display Port) format video sub-pictures through the format converter to obtain rendering sub-picture signals consistent with the SDI format specified by the video signal acquisition card for synchronous acquisition by the video signal acquisition card.
  • the format converter when the signal format of the rendered sub-picture signal obtained by synchronous rendering of the master and slave rendering machines is inconsistent with the signal format specified by the video signal acquisition card, after receiving the synchronous conversion signal through the format converter, the format converter performs synchronous conversion on the rendered sub-picture signals corresponding to the multiple video sub-pictures, and obtains the rendered sub-picture signal consistent with the signal format specified by the video signal acquisition card, so as to provide the video signal acquisition card with synchronous acquisition.
  • a video signal acquisition card is deployed on a screen synthesizer; the screen synthesizer provides a synthetic video canvas; the synchronously acquired rendering sub-picture signals are synthesized to obtain a rendering picture corresponding to the video frame picture, including: setting the frame rate of the synthetic video canvas to a preset frame rate; through the synthetic video canvas that meets the preset frame rate, the synchronously acquired rendering sub-picture signals are synchronously synthesized to obtain a rendering picture corresponding to the video frame picture.
  • a picture synthesizer is also deployed in the computer device, and a video signal acquisition card, synthesis software, and a synthetic video canvas created based on the synthesis software are deployed on the picture synthesizer.
  • the computer device can set the frame rate of the synthetic video canvas to the same preset frame rate as the synchronization signal generator. Furthermore, the computer device can synchronously synthesize the rendering sub-picture signals corresponding to the multiple video sub-pictures synchronously acquired by the video signal acquisition card through the synthetic video canvas that meets the preset frame rate, and obtain the rendering picture corresponding to the video frame picture.
  • the frame rate of the synthetic video canvas is set to be the same as the preset frame rate of the synchronization signal generator, and the rendering sub-picture signals corresponding to the multiple video sub-pictures synchronously collected by the video signal acquisition card are synchronously synthesized through the synthetic video canvas that meets the preset frame rate to obtain the rendering picture corresponding to the video frame picture, which can further avoid The problem of video tearing occurs, thereby further improving the quality of the final rendered video image.
  • the synchronization signal generator can generate a synchronization signal. It can be understood that when the synchronization signal generated by the synchronization signal generator is received by a rendering machine, the synchronization signal is a synchronization rendering signal. When the synchronization signal generated by the synchronization signal generator is received by a format converter, the synchronization signal is a synchronization conversion signal. When the synchronization signal generated by the synchronization signal generator is received by a video signal acquisition card, the synchronization signal is a synchronization acquisition signal. It should be noted that the synchronization signal generator is a phase synchronization signal generator. Phase-level time alignment can be achieved between multiple systems based on the received synchronization signal.
  • a computer device is internally integrated with a synchronization signal generator, two rendering machines (rendering machine 1 and rendering machine 2), two format converters (format converter 1 and format converter 2) and a picture synthesizer equipped with a video signal acquisition card.
  • a synchronization card is integrated in the rendering machine 1.
  • the synchronization signal generator can generate a synchronization rendering signal, a synchronization conversion signal and a synchronization acquisition signal.
  • the multiple video sub-pictures are specifically two video sub-pictures.
  • the rendering machine 1 can receive the synchronization rendering signal generated by the synchronization signal generator through the synchronization card, and synchronize the synchronization rendering signal to the rendering machine 2.
  • the rendering machine 1 and the rendering machine 2 are controlled to perform synchronous rendering of the two video sub-pictures respectively, and the rendering sub-picture signals corresponding to the two video sub-pictures are obtained.
  • the signal format of the rendering sub-picture signal obtained by the synchronous rendering of the rendering machine 1 and the rendering machine 2 is inconsistent with the signal format specified by the video signal acquisition card
  • the format conversion of the rendering sub-picture signal output by the rendering machine 1 is performed through the format converter 1
  • the format conversion of the rendering sub-picture signal output by the rendering machine 2 is performed through the format converter 2
  • the rendering sub-picture signal consistent with the signal format specified by the video signal acquisition card is obtained.
  • the video signal acquisition card When the video signal acquisition card receives the synchronous acquisition signal generated by the synchronous signal generator, the video signal acquisition card synchronously acquires the rendering sub-picture signals corresponding to the two video sub-pictures, and synthesizes the rendering sub-picture signals corresponding to the two synchronously acquired video sub-pictures to obtain the rendering picture corresponding to the video frame. It can be understood that the output rendering picture can be applied to multiple business scenarios.
  • the computer device may determine the rendering area according to the relative position between the viewpoint in the simulated camera and each vertex in the mesh patch, and determine the real-time video frame to be rendered in the rendering area according to the position of the simulated camera.
  • the computer device may spatially split the video frame according to the mesh patch to obtain N video sub-pictures.
  • the computer device may synchronously render the N video sub-pictures through N pre-set rendering machines to obtain rendering sub-pictures corresponding to the N video sub-pictures.
  • the computer device may synthesize the rendering sub-pictures corresponding to the N video sub-pictures through a picture synthesizer to obtain a rendering picture corresponding to the video frame.
  • N is a positive integer greater than 2 and is a constant.
  • the video frame includes a virtual object real-time picture
  • the video sub-picture includes a real-time sub-picture.
  • Obtaining the video frame to be rendered in the rendering area includes: obtaining the position of the simulation camera, and determining the real-time virtual object real-time picture to be rendered in the rendering area according to the position of the simulation camera.
  • Synthesizing the rendering sub-pictures corresponding to the multiple video sub-pictures to obtain the rendering picture corresponding to the video frame includes: synthesizing the rendering sub-pictures corresponding to the multiple real-time sub-pictures to obtain the rendering picture corresponding to the virtual object real-time picture.
  • determining the video frame to be rendered in the rendering area according to the position of the virtual camera includes: determining the real-time virtual object real-time picture to be rendered in the rendering area according to the position of the simulation camera.
  • the real-time picture of the virtual object is a video picture determined in the real-time rendering scene of the virtual object.
  • the real-time sub-picture is a video picture obtained by spatially splitting the real-time picture of the virtual object.
  • the virtual object is a virtual entity object, which may include at least one of a virtual person, a virtual animal and a virtual object.
  • the computer device can determine the rendering area according to the relative position between the viewpoint in the simulated camera and each vertex in the mesh patch.
  • the computer device can determine the real-time virtual object real-time picture to be rendered in the rendering area according to the position of the simulated camera.
  • the computer device can divide the real-time picture of the virtual object in space according to the mesh patch to obtain multiple real-time sub-pictures, and synchronously render the multiple real-time sub-pictures through multiple pre-set rendering machines to obtain the corresponding real-time sub-pictures of the multiple real-time sub-pictures. Rendering sub-picture.
  • the computer device may synthesize the rendering sub-pictures corresponding to the multiple real-time sub-pictures to obtain a rendering picture corresponding to the real-time picture of the virtual object.
  • the real-time virtual object real-time picture to be rendered in the rendering area is determined, and the acquisition accuracy of the virtual object real-time picture can be improved.
  • the rendering sub-pictures corresponding to the multiple real-time sub-pictures respectively the rendering picture corresponding to the virtual object real-time picture is obtained, and the rendering picture instruction in the virtual object real-time rendering scene can be improved.
  • the rendering area is determined according to the relative position between the viewpoint in the simulated camera and each vertex in the mesh patch, and the real-time video frame image to be rendered in the rendering area is determined according to the position of the simulated camera. Since the simulated camera is obtained by simulating the virtual camera in the rendering engine, and the mesh patch is a patch that can be used for image splitting processing based on the physical size of the virtual sensor in the virtual camera.
  • the video frame image is spatially split according to the mesh patch to obtain multiple video sub-images.
  • multiple video sub-images are synchronously rendered to obtain rendering sub-images corresponding to the plurality of video sub-images, and the rendering sub-images corresponding to the plurality of video sub-images are synthesized to obtain the rendering image corresponding to the video frame image.
  • this application spatially splits the video frame image through a grid patch constructed based on the physical size of the virtual sensor in the virtual camera, and synchronously renders the split sub-images through multiple rendering machines, distributing the rendering pressure to multiple rendering machines, thereby obtaining high-quality real-time video images and meeting the needs of real-time video rendering scenarios.
  • a video image rendering method is provided.
  • the method can be applied to a computer device, which can be a terminal or a server.
  • the method can be executed by the terminal or the server itself, or can be implemented through interaction between the terminal and the server.
  • This embodiment is described by taking the method applied to a computer device as an example. The method specifically includes the following steps:
  • Step 1102 modeling is performed according to the size of the virtual sensor represented by the physical size and the number of rendering machines to obtain multiple grid sub-faces; wherein the number of grid sub-faces is consistent with the number of rendering machines; and the sum of the areas of the multiple grid sub-faces is consistent with the area of the virtual sensor.
  • Step 1104 determining the distance between the viewpoint and the plurality of mesh sub-faces according to the focal length of the virtual camera in the rendering engine.
  • Step 1106 obtaining the relative position between the viewpoint and each vertex in the plurality of mesh sub-faces according to the distance.
  • Step 1108 determining the rendering area according to the relative position between the viewpoint in the simulated camera and each vertex in each mesh sub-face; the simulated camera is obtained by simulating the virtual camera in the rendering engine through the display component.
  • Step 1110 determine the position of the virtual camera, assign the position of the virtual camera to the display component, and obtain the position of the simulated camera.
  • Step 1112 according to the position of the simulated camera, determine the real-time scene content located in the rendering area in the three-dimensional virtual scene, and obtain the real-time video frame to be rendered.
  • Step 1114 for each video frame, the rendering area corresponding to the video frame is spatially split according to the common vertices in multiple mesh sub-faces, and the real-time scene content in each of the split sub-areas is determined as the video sub-picture.
  • Step 1116 receiving a synchronous rendering signal generated by a synchronous signal generator based on a preset frame rate through a synchronous card in the main rendering machine.
  • Step 1118 synchronize the synchronous rendering signal to each slave rendering machine through the synchronization card.
  • Step 1120 through the synchronous rendering signals respectively received by the master rendering machine and each slave rendering machine, the master rendering machine and the slave rendering machine are controlled to synchronously render the multiple video sub-pictures respectively, to obtain rendering sub-pictures corresponding to the multiple video sub-pictures respectively.
  • Step 1122 using a picture synthesizer, synthesizes the rendering sub-pictures corresponding to the multiple video sub-pictures to obtain a rendering picture corresponding to the video frame picture.
  • the present application also provides an application scenario, which applies the above-mentioned video screen rendering method.
  • the video screen rendering method can be applied to the video screen rendering scenario for live broadcast of virtual objects.
  • the computer device can model according to the size of the virtual sensor represented by the physical size and the number of rendering machines to obtain multiple grid sub-faces; wherein the number of grid sub-faces is consistent with the number of rendering machines; the sum of the areas of the multiple grid sub-faces is consistent with the area of the virtual sensor.
  • the focal length of the virtual camera in the rendering engine the distance between the viewpoint and the multiple grid sub-faces is determined. According to the distance, the relative position between the viewpoint and each vertex in the multiple grid sub-faces is obtained.
  • the rendering area is determined; the simulated camera is obtained by simulating the virtual camera in the rendering engine through the display component. Determine the position of the virtual camera, assign the position of the virtual camera to the display component, and obtain the position of the simulated camera. According to the position of the simulated camera, determine the real-time scene content located in the rendering area in the three-dimensional virtual scene, and obtain the real-time virtual object live screen to be rendered.
  • the computer device can spatially split the rendering area corresponding to the live broadcast screen of the virtual object according to the common vertices in multiple mesh sub-facets, and determine the real-time scene content in each sub-area after the split as the live broadcast sub-screen.
  • the synchronization rendering signal generated by the synchronization signal generator based on the preset preset frame rate is received.
  • the synchronization rendering signal is synchronized to each slave rendering machine through the synchronization card.
  • the master rendering machine and the slave rendering machine are controlled to synchronously render multiple live sub-screens respectively, and the rendering sub-screens corresponding to the multiple live sub-screens are obtained.
  • the rendering sub-screens corresponding to the multiple live sub-screens are synthesized to obtain the rendering screen corresponding to the virtual object live screen.
  • the present application spatially splits the virtual object live screen by the mesh facets constructed based on the physical size of the virtual sensor in the virtual camera, and synchronously renders the split live sub-screens through multiple rendering machines, so as to obtain high-quality real-time virtual object live screens, meeting the needs of real-time virtual object live screen rendering scenes.
  • the present application also provides an application scenario, which applies the above-mentioned video screen rendering method.
  • the video screen rendering method can be applied to the video screen rendering scene for XR (Extended Reality) live broadcast.
  • extended reality refers to the combination of reality and virtuality through computers to create a virtual environment for human-computer interaction, which can bring an immersive sense of seamless transition between the virtual world and the real world to the experiencer.
  • the extended reality live screen is spatially split into regions, and the split live sub-screens are synchronously rendered by multiple rendering machines, so that high-quality real-time extended reality live screens can be obtained, which meets the needs of real-time extended reality live screen rendering scenarios.
  • each step in the flow chart of the above-mentioned embodiments is shown in order, these steps are not necessarily performed in order. Unless there is a clear explanation in this article, the execution of these steps does not have strict order restrictions, and these steps can be performed in other orders. Moreover, at least a portion of the steps in the above-mentioned embodiments may include a plurality of sub-steps or a plurality of stages, and these sub-steps or stages are not necessarily performed at the same time, but can be performed at different times, and the execution order of these sub-steps or stages is not necessarily performed in order, but can be performed in turn or alternately with at least a portion of other steps or sub-steps or stages of other steps.
  • a video image rendering device 1200 is provided.
  • the device may adopt a software module or a hardware module, or a combination of the two to form a part of a computer device.
  • the device specifically includes:
  • Determine module 1202 which is used to determine a simulated camera, which is obtained by simulating the virtual camera of the rendering engine; determine a mesh patch, which matches the size of the virtual sensor in the virtual camera; determine the positional relationship between the viewpoint in the simulated camera and each vertex in the mesh patch, and determine the rendering area according to the positional relationship; and obtain the video frame to be rendered in the rendering area.
  • the splitting module 1204 is used to split the video frame according to the grid surface to obtain multiple video sub-pictures.
  • the rendering module 1206 is used to synchronously render the multiple video sub-pictures through multiple pre-set rendering machines to obtain To the rendering picture corresponding to the video frame picture.
  • the rendering area is the area inside the cone when the viewpoint is connected to each vertex of the mesh surface to form a cone.
  • the determination module 1202 is further used to obtain the position of the simulated camera, and determine the video frame to be rendered in the rendering area according to the position of the virtual camera.
  • the position of the virtual camera in the three-dimensional virtual scene and the position of the simulated camera are determined according to the position of the virtual camera in the three-dimensional virtual scene.
  • the determination module 1202 is also used to determine the real-time scene content located in the rendering area in the three-dimensional virtual scene according to the position of the simulated camera, and obtain the real-time video frame to be rendered.
  • the mesh patch includes multiple mesh sub-patches, and the common vertices of two adjacent mesh sub-patches in the multiple mesh sub-patches have the same coordinates.
  • the splitting module 1204 is also used to spatially split the rendering area corresponding to the video frame picture according to the common vertices in the multiple mesh sub-patches to obtain multiple sub-areas for each of the video frame pictures; determine the real-time scene content in each of the sub-areas in the three-dimensional virtual scene, and determine the real-time scene content in each of the sub-areas as video sub-pictures.
  • the determination module 1202 is also used to determine the focal length of the virtual camera; based on the focal length, determine the distance between the viewpoint in the simulated camera and the mesh patch; based on the distance, determine the positional relationship between the viewpoint and each vertex in the mesh patch.
  • the simulated camera is obtained by simulating a display component, and the determination module 1202 is further used to determine the position of the virtual camera; the position of the virtual camera is assigned to the display component to obtain the position of the simulated camera.
  • the rendering module 1206 is further configured to synchronously render the plurality of video sub-pictures through a plurality of pre-set rendering machines to obtain rendering sub-pictures corresponding to the plurality of video sub-pictures.
  • the video picture rendering device 1200 further includes a synthesis module, configured to synthesize the rendering sub-pictures corresponding to the plurality of video sub-pictures to obtain a rendering picture corresponding to the video frame picture.
  • the size of the virtual sensor is represented by a physical size.
  • the video image rendering device 1200 also includes a construction module, which is used to perform modeling according to the size of the virtual sensor and the number of the rendering machines to obtain the multiple grid sub-faces.
  • the number of the grid sub-patterns is consistent with the number of the rendering machines; and the sum of the areas of the plurality of grid sub-patterns is consistent with the area of the virtual sensor.
  • the multiple rendering machines include a master rendering machine and at least one slave rendering machine, and a synchronization card is deployed in the master rendering machine; the rendering module 1206 is also used to receive a synchronization rendering signal generated by a synchronization signal generator based on a preset frame rate through the synchronization card in the master rendering machine; synchronize the synchronization rendering signal to the slave rendering machine through the synchronization card; and control the master rendering machine and the slave rendering machine to synchronously render the multiple video sub-pictures respectively through the synchronization rendering signals respectively received by the master rendering machine and the slave rendering machine, so as to obtain rendering sub-pictures corresponding to the multiple video sub-pictures respectively.
  • the rendering module 1206 is also used to synchronously acquire the rendering sub-picture signals corresponding to the multiple video sub-pictures through the video signal acquisition card when the video signal acquisition card receives the synchronous acquisition signal generated by the synchronous signal generator; and synthesize the synchronously acquired rendering sub-picture signals to obtain the rendering picture corresponding to the video frame picture.
  • the video screen rendering device 1200 also includes a conversion module, which is used to, when the signal format of the rendered sub-screen signal obtained by synchronous rendering is inconsistent with the signal format specified by the video signal acquisition card, after receiving the synchronous conversion signal through the format converter, perform format synchronous conversion on the rendered sub-screen signals corresponding to the multiple video sub-screens respectively through the format converter to obtain the rendered sub-screen signal consistent with the signal format specified by the video signal acquisition card for synchronous acquisition by the video signal acquisition card.
  • a conversion module which is used to, when the signal format of the rendered sub-screen signal obtained by synchronous rendering is inconsistent with the signal format specified by the video signal acquisition card, after receiving the synchronous conversion signal through the format converter, perform format synchronous conversion on the rendered sub-screen signals corresponding to the multiple video sub-screens respectively through the format converter to obtain the rendered sub-screen signal consistent with the signal format specified by the video signal acquisition card for synchronous acquisition by the video signal acquisition card.
  • the video signal acquisition card is deployed on a picture synthesizer; the picture synthesizer provides a synthesized video canvas; the synthesis module is further used to set the frame rate of the synthesized video canvas to the preset frame rate; The synthetic video canvas that meets the preset frame rate synchronously synthesizes the synchronously collected rendering sub-picture signals to obtain the rendering picture corresponding to the video frame picture.
  • the video frame includes a virtual object real-time picture
  • the video sub-picture includes a real-time sub-picture.
  • the determination module 1202 is also used to determine the real-time virtual object real-time picture to be rendered in the rendering area according to the position of the simulated camera; the synthesis module is also used to synthesize the rendering sub-pictures corresponding to the multiple real-time sub-pictures to obtain the rendering picture corresponding to the virtual object real-time picture.
  • the above-mentioned video picture rendering device splits the video frame picture through a grid patch that matches the size of the virtual sensor in the virtual camera, and synchronously renders the split sub-pictures through a plurality of pre-set rendering machines, so as to efficiently obtain a high-quality rendering picture, and meet the rendering requirements in scenes with high picture quality and rendering efficiency.
  • the larger the picture size the more prominent the effect of the above-mentioned rendering picture rendering method is compared with the traditional method.
  • the above-mentioned video picture rendering method can better meet the high real-time requirements of video picture rendering.
  • the above-mentioned video picture rendering method has a particularly prominent effect.
  • Each module in the above video rendering device can be implemented in whole or in part by software, hardware, or a combination thereof.
  • Each module can be embedded in or independent of a processor in a computer device in the form of hardware, or can be stored in a memory in a computer device in the form of software, so that the processor can call and execute operations corresponding to each module.
  • a computer device which can be a server or a terminal, and its internal structure diagram can be shown in Figure 13.
  • the computer device includes a processor, a memory, an input/output interface (Input/Output, referred to as I/O) and a communication interface.
  • the processor, the memory and the input/output interface are connected through a system bus, and the communication interface is connected to the system bus through the input/output interface.
  • the processor of the computer device is used to provide computing and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system and computer-readable instructions.
  • the internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium.
  • the input/output interface of the computer device is used to exchange information between the processor and an external device.
  • the communication interface of the computer device is used to communicate with an external terminal through a network connection.
  • FIG. 13 is merely a block diagram of a partial structure related to the scheme of the present application, and does not constitute a limitation on the computer device to which the scheme of the present application is applied.
  • the specific computer device may include more or fewer components than shown in the figure, or combine certain components, or have a different arrangement of components.
  • a computer device including a memory and a processor, wherein the memory stores computer-readable instructions, and the processor implements the steps in the above-mentioned method embodiments when executing the computer-readable instructions.
  • a computer-readable storage medium which stores computer-readable instructions.
  • the steps in the above-mentioned method embodiments are implemented.
  • a computer program product including computer-readable instructions, which implement the steps in the above-mentioned method embodiments when executed by a processor.
  • user information including but not limited to user device information, user personal information, etc.
  • data including but not limited to data used for analysis, stored data, displayed data, etc.
  • Non-volatile memory can include read-only memory (ROM), magnetic tape, floppy disk, flash memory or optical storage, etc.
  • Volatile memory can include
  • the RAM may be a random access memory (RAM) or an external cache memory.
  • the RAM may be in various forms, such as a static random access memory (SRAM) or a dynamic random access memory (DRAM).

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Geometry (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

一种视频画面渲染方法,包括:确定模拟摄像机,所述模拟摄像机是模拟渲染引擎的虚拟摄像机得到的;确定网格面片,所述网格面片与所述虚拟摄像机中虚拟传感器的尺寸匹配;确定所述模拟摄像机中的视点与所述网格面片中各个顶点之间的位置关系,根据所述位置关系确定渲染区域(202);获取所述渲染区域内待渲染的视频帧画面(204);根据所述网格面片,将所述视频帧画面进行拆分,得到多个视频子画面(206);通过预先设置的多个渲染机,同步渲染所述多个视频子画面,得到所述视频帧画面对应的渲染画面(208)。

Description

视频画面渲染方法、装置、设备和介质
相关申请:
本申请要求于2022年10月28日提交中国专利局,申请号为2022113357187、发明名称为“视频画面渲染方法、装置、设备和介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请涉及视频处理技术,特别是涉及一种视频画面渲染方法、装置、设备和介质。
背景技术
随着计算机技术的发展,人们对视频画面的质量要求越来越高。比如,在游戏领域中,玩家对游戏画面的质量也有较高的要求。高质量的视频画面需要同时具备高分辨率和高帧率,通过渲染机对视频画面进行渲染,渲染机的渲染压力较重,超负荷的渲染机无法满足视频画面的高质量要求。
发明内容
根据本申请提供的各种实施例,提供一种视频画面渲染方法、装置、设备和介质。
第一方面,本申请提供了一种视频画面渲染方法,由计算机设备执行,包括:
确定模拟摄像机,所述模拟摄像机是模拟渲染引擎的虚拟摄像机得到的;
确定网格面片,所述网格面片与所述虚拟摄像机中虚拟传感器的尺寸匹配;
确定所述模拟摄像机中的视点与所述网格面片中各个顶点之间的位置关系,根据所述位置关系确定渲染区域;
获取所述渲染区域内待渲染的视频帧画面;
根据所述网格面片,将所述视频帧画面进行拆分,得到多个视频子画面;
通过预先设置的多个渲染机,同步渲染所述多个视频子画面,得到所述视频帧画面对应的渲染画面。
第二方面,本申请提供了一种视频画面渲染装置,所述装置包括:
确定模块,用于确定模拟摄像机,所述模拟摄像机是模拟渲染引擎的虚拟摄像机得到的;确定网格面片,所述网格面片与所述虚拟摄像机中虚拟传感器的尺寸匹配;确定所述模拟摄像机中的视点与所述网格面片中各个顶点之间的位置关系,根据所述位置关系确定渲染区域;获取所述渲染区域内待渲染的视频帧画面;获取所述渲染区域内待渲染的视频帧画面;
拆分模块,用于根据所述网格面片,将所述视频帧画面拆分,得到多个视频子画面;
渲染模块,用于通过预先设置的多个渲染机,同步渲染所述多个视频子画面,得到所述视频帧画面对应的渲染画面面。
第三方面,本申请还提供了一种计算机设备,包括存储器和处理器,所述存储器存储有计算机可读指令,所述处理器执行所述计算机可读指令时执行本申请各方法实施例的步骤。
第四方面,本申请还提供了一种计算机可读存储介质,存储有计算机可读指令,所述计算机可读指令被处理器执行时执行本申请各方法实施例的步骤。
第五方面,本申请还提供了一种计算机程序产品,包括计算机可读指令,所述计算机可读指令被处理器执行时执行本申请各方法实施例的步骤。
本申请的一个或多个实施例的细节在下面的附图和描述中提出。本申请的其它特征、目的和优点将从说明书、附图以及权利要求书变得明显。
附图说明
为了更清楚地说明本申请实施例或传统技术中的技术方案,下面将对实施例或传统技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据公 开的附图获得其他的附图。
图1为一个实施例中视频画面渲染方法的应用环境图;
图2为一个实施例中视频画面渲染方法的流程示意图;
图3为一个实施例中视点与网格面片之间的位置关系示意图;
图4为一个实施例中基于虚拟摄像机的焦距确定视点与网格面片之间的距离的原理示意图;
图5为一个实施例中基于虚拟摄像机的位置确定模拟摄像机的位置的原理示意图;
图6为一个实施例中基于虚拟传感器的尺寸进行切分建模,并进行分布式渲染的原理示意图;
图7为一个实施例中渲染机、网格子面片和画面视口之间的映射关系示意图;
图8为一个实施例中针对视频画面渲染构建的硬件环境示意图;
图9为另一个实施例中视频画面渲染方法的流程示意图;
图10为一个实施例中视频画面渲染方法的应用场景示意图;
图11为又一个实施例中视频画面渲染方法的流程示意图;
图12为一个实施例中视频画面渲染装置的结构框图;
图13为一个实施例中计算机设备的内部结构图。
具体实施方式
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
本申请提供的视频画面渲染方法,可以应用于如图1所示的应用环境中。其中,终端102通过网络与服务器104进行通信。数据存储系统可以存储服务器104需要处理的数据。数据存储系统可以集成在服务器104上,也可以放在云上或其他服务器上。其中,终端102可以但不限于是各种台式计算机、笔记本电脑、智能手机、平板电脑、物联网设备和便携式可穿戴设备,物联网设备可为智能音箱、智能电视、智能空调、智能车载设备等。便携式可穿戴设备可为智能手表、智能手环、头戴设备等。服务器104可以是独立的物理服务器,也可以是多个物理服务器构成的服务器集群或者分布式系统,还可以是提供云服务、云数据库、云计算、云函数、云存储、网络服务、云通信、中间件服务、域名服务、安全服务、CDN、以及大数据和人工智能平台等基础云计算服务的云服务器。终端102以及服务器104可以通过有线或无线通信方式进行直接或间接地连接,本申请在此不做限制。
服务器104可确定模拟摄像机,模拟摄像机是模拟渲染引擎的虚拟摄像机得到的。服务器104可确定网格面片,网格面片与虚拟摄像机中虚拟传感器的尺寸匹配。服务器104可确定模拟摄像机中的视点与网格面片中各个顶点之间的位置关系,根据位置关系确定渲染区域。服务器104可获取渲染区域内待渲染的视频帧画面。服务器104可根据网格面片,将视频帧画面进行拆分,得到多个视频子画面。服务器104可通过预先设置的多个渲染机,同步渲染多个视频子画面,得到视频帧画面对应的渲染画面。
可以理解,服务器104可将合成得到的渲染画面发送至终端102进行显示。本实施例对此不做限定,可以理解,图1中的应用场景仅为示意说明,并不限定于此。
在一个实施例中,如图2所示,提供了一种视频画面渲染方法,该方法可应用于计算机设备,计算机设备可以是终端或服务器,由终端或服务器自身单独执行,也可以通过终端和服务器之间的交互来实现。本实施例以该方法应用于计算机设备为例进行说明,包括以下步骤:
步骤202,确定模拟摄像机,模拟摄像机是模拟渲染引擎的虚拟摄像机得到的。确定网格面片,网格面片与虚拟摄像机中虚拟传感器的尺寸匹配。确定模拟摄像机中的视点与 网格面片中各个顶点之间的位置关系,根据位置关系确定渲染区域。
其中,渲染引擎是用于渲染画面的三维实时渲染引擎,可以是游戏引擎,如虚幻引擎。虚拟摄像机是渲染引擎中虚拟的摄像机。虚拟摄像机可用于实现物理摄像机的功能。模拟摄像机是针对虚拟摄像机进行模拟得到的摄像机。模拟摄像机可以具有所模拟虚拟摄像机的功能。
模拟摄像机具有视点和网格面片。视点可以称之为观察点,从不同视点观察同一对象,可以得到同一对象在不同视角下的画面。网格面片,是基于虚拟摄像机中虚拟传感器的尺寸构建的三维面片,该三维面片位于三维虚拟场景中。模拟摄像机的视点的位置,可以根据模拟摄像机的位置确定,如可以直接将模拟摄像机的位置作为模拟摄像机的视点的位置,也可以将模拟摄像机的位置按照预设偏移关系偏移后得到视点的位置。
虚拟传感器是虚拟摄像机中虚拟的传感器。虚拟传感器具有物理传感器的功能。虚拟传感器的尺寸可以是采用物理尺寸表示的,比如虚拟传感器的尺寸可以采用物理尺寸表示为23.76mm*13.365mm,mm表示毫米。网格面片与虚拟摄像机中虚拟传感器的尺寸匹配,可以是网格面片与虚拟摄像机中虚拟传感器的尺寸相同,或者保持相同的比例。
在一个实施例中,渲染区域,是当视点分别与网格面片的每个顶点相连构成锥体时,椎体内部的区域。
渲染区域是用于确定需渲染的视频帧画面的区域。渲染区域可以是从视点到包围网格面片的边缘形成的椎体内部的区域。渲染区域可以是在三维虚拟场景中占用的空间,该空间用来渲染成二维的视频画面。顶点是用于确定网格面片的边缘的点,同一顶点连接网格面片的不同边缘。当网格面片是矩形,矩形的四个直角顶点即为网格面片的顶点。
本实施例中,渲染区域是椎体内部的区域,利用渲染区域可以准确地在三维虚拟场景中确定待渲染的内容,进而可以准确生成待渲染的视频帧画面,进而为保障最终渲染画面的质量做好准备。
视点和网格面片的顶点可以确定渲染区域。于是,计算机设备可确定视点与网格面片各个顶点之间的位置关系,确定渲染区域。位置关系可以是视点与顶点之间的相对位置,如可以是顶点在网络面片所在平面上相对于视点的偏移量。位置关系也可以用视点与顶点各自在三维虚拟场景中的位置表示。
计算机设备可确定视点的位置以及网格面片各个的顶点的位置,从而确定视点和网格面片的顶点之间的位置关系,从而利用视点和网格面片各个顶点之间的位置关系,确定渲染区域。
可以理解,计算机设备可基于视点与网格面片中各个顶点确定得到一个锥体,这个锥体内部的区域即渲染区域。当网格面片是矩形,视点连接网格面片四个顶点构成椎体,椎体内部为渲染区域。
举例说明,如图3所示,计算机设备可将模拟摄像机中的视点O分别与网格面片ABCD中各个顶点(即A、B、C、D)进行相连,得到一个四棱锥体O-ABCD,这个锥体内部的区域即渲染区域。
步骤204,获取渲染区域内待渲染的视频帧画面。
其中,视频帧画面是视频或视频流中一帧视频帧的图像。视频帧画面可以是实时的画面,具体可以是实时的视频流中当前时刻对应的一帧视频帧的画面。可以理解,视频流包括多帧视频帧,多帧视频帧对应多张视频帧画面。
模拟摄像机的位置,可以是模拟摄像机在三维虚拟场景中的位置。模拟摄像机的位置对应了模拟摄像机的视点的位置。于是,计算机设备可根据摄像机的位置,确定视点的位置,从而确定渲染区域在三维虚拟场景中的空间位置,从而可以基于渲染区域内的内容确定视频帧画面。计算机设备可将渲染区域内三维的内容投射到二维的网格面片上,得到待渲染的视频帧画面。
在一个实施例中,获取渲染区域内待渲染的视频帧画面包括:获取模拟摄像机的位置,根据虚拟摄像机的位置,确定渲染区域内待渲染的视频帧画面。
计算机设备可获取模拟摄像机当前时刻的位置,从而根据虚拟摄像机当前时刻的位置,确定渲染区域内待渲染的实时的视频帧画面。
进一步地,计算机设备可获取模拟摄像机当前时刻的位置,进而根据模拟摄像机当前时刻的位置,确定模拟摄像机的视点当前时刻的位置,从而确定三维虚拟场景中当前时刻在渲染区域的内容,进而根据该内容确定渲染区域内待渲染的实时的视频帧画面。
本实施例中,利用模拟摄像机的位置,可以准确地确定渲染区域内待渲染的视频帧画面,进而为保障最终渲染画面的质量做好准备。
在一个实施例中,计算机设备可获取虚拟摄像机的位置,并根据虚拟摄像机的位置确定模拟摄像机的位置。可以理解,计算机设备可将虚拟摄像机的位置作为模拟摄像机的位置。计算机设备还可对虚拟摄像机的位置进行调整,如按照预设偏移关系进行偏移,并将偏移后的位置作为模拟摄像机的位置。
步骤206,根据网格面片,将视频帧画面进行拆分,得到多个视频子画面。
其中,视频子画面是将视频帧画面在进行区域拆分得到的视频画面,可以理解,拆分得到的多个视频子画面相互独立,且能够合并成视频帧画面。计算机设备可根据网格面片的尺寸,确定视频帧画面的尺寸,按照网格面片的尺寸确定拆分方式,按照该拆分方式将视频帧画面进行拆分,得到多个视频子画面。
将视频帧画面进行拆分,可以是按照预设的拆分方式进行拆分,也可以是按照即时确定的拆分方式进行拆分。拆分方式可以是等分,也可以是不等分。如果是不等分,可以是部分视频子画面尺寸相同,部分视频子画面尺寸不相同,也可以是每个视频子画面尺寸都不相同。
在一个实施例中,计算机设备可针对每张实时的视频帧画面,根据网格面片对视频帧画面对应的渲染区域在空间上进行拆分,得到多个子区域。进而,计算机设备可分别确定各个子区域内的待渲染的实时画面内容,并基于各个子区域内的待渲染的实时画面内容,得到多个视频子画面。将视频帧画面在空间上拆分,是指将视频帧画面从画面维度进行拆分,拆分后的视频子画面能够分别渲染。
步骤208,通过预先设置的多个渲染机,同步渲染多个视频子画面,得到视频帧画面对应的渲染画面。
渲染机可以是设置于计算机设备中用于执行渲染任务的计算机程序。渲染机还可以是用于执行渲染任务的计算机设备。通过预先设置的多个渲染机同步渲染多个视频子画面,是多个渲染机并行地执行渲染任务,以并行地渲染多个视频子画面。渲染机的数量可以与视频子画面的数量相同,也可以与视频子画面的数量不同。渲染画面是针对视频帧画面进行渲染后得到的画面。渲染画面可通过计算机设备的显示单元或独立的显示设备进行显示。
在一个实施例中,多个渲染机可同步渲染多个视频子画面,得到与多个视频子画面一一对应的多个渲染子画面,多个渲染子画面构成视频帧画面对应的渲染画面。其中,每个渲染子画面是对对应的视频子画面进行渲染得到的画面。每个渲染机可用于渲染至少一个视频子画面。渲染机具有网络地址,渲染机可渲染与其网络地址具有预设的映射关系的视频子画面。
在一个实施例中,多个渲染机可同步渲染多个视频子画面,得到与多个视频子画面一一对应的多个渲染子画面,将多个渲染子画面按照多个视频子画面的位置关系拼接,得到视频帧画面对应的渲染画面。
执行步骤202至步骤206的计算机设备,可以是多个渲染机中的一个,可以称之为主渲染机,此时主渲染机可以协调多个渲染机中的从渲染机,从而主渲染机与从渲染机同步渲染多个视频子画面。执行步骤202-206的计算机设备,可以是独立于任一渲染机的设备, 此时计算机设备可指示多个渲染机同步渲染多个视频子画面。
在一个实施例中,当渲染机的数量与视频子画面的数量相同,每个渲染机所渲染视频子画面相对于视频帧画面的位置可以是固定的。比如,有4个渲染机,编号为1、2、3、4,每个视频帧画面拆分出4个视频子画面,按照位置编号为I、II、III、IV,则渲染机1号可固定渲染位置I处的视频子画面,渲染机2号可固定渲染位置II处的视频子画面,渲染机3号可固定渲染位置III处的视频子画面,渲染机4号可固定渲染位置IV处的视频子画面。
在一个实施例中,当渲染机的数量小于视频子画面的数量,多个渲染机中的至少一个渲染机,可在同一时刻渲染至少两个视频子画面。同一时刻渲染至少两个视频子画面的渲染机,可以是多个渲染机中固定的渲染机,如计算性能比同一时刻渲染一个视频子画面的渲染机强的渲染机。同一时刻渲染至少两个视频子画面的渲染机,在相邻的下一时刻可以是同一时刻渲染一个视频子画面的渲染机。同一时刻渲染至少两个视频子画面的渲染机,可以从多个渲染机中轮流选择,也可以根据负载情况选择。
在一个实施例中,当渲染机的数量大于视频子画面的数量,每个渲染机同一时刻可仅渲染一个视频子画面。不渲染视频子画面的渲染机为空闲的渲染机,不同时刻空闲的渲染机可以相同也可以不同。
上述视频画面渲染方法,通过与虚拟摄像机中虚拟传感器的尺寸匹配的网格面片,对视频帧画面进行拆分,并通过预先设置的多个渲染机,对拆分的子画面进行同步渲染,可以高效地得到高质量的渲染画面,满足了对画面质量和渲染效率高的场景下的渲染需求。比如,画面尺寸越大,采用上述渲染画面渲染方法,相比传统方法效果则更加突出。还比如,对实时性要求较高的场景,如渲染实时的视频画面的场景,采用上述视频画面渲染方法能够更好地满足视频画面渲染的高实时性要求。特别低,对于建筑场所内部或外部设置的屏幕,商用巨型屏幕,采用上述视频画面渲染方法效果特别突出。
在一个实施例中,虚拟摄像机在三维虚拟场景中,模拟摄像机的位置,是根据虚拟摄像机在三维虚拟场景中的位置确定的。根据虚拟摄像机的位置,确定渲染区域内待渲染的视频帧画面包括:根据模拟摄像机的位置,确定三维虚拟场景中位于渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面。
在一个实施例中,模拟摄像机的位置是根据虚拟摄像机在三维虚拟场景中的位置确定的;根据模拟摄像机的位置,确定渲染区域内待渲染的实时的视频帧画面,包括:根据模拟摄像机的位置,确定三维虚拟场景中位于渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面。
计算机设备可根据虚拟摄像机在三维虚拟场景中的位置,确定出模拟摄像机的位置,并根据模拟摄像机的位置,确定三维虚拟场景中位于渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面。可以理解,三维虚拟场景中位于渲染区域之内的实时场景内容是可被渲染出来的,即是可见的。三维虚拟场景中位于渲染区域之外的实时场景内容是不会被渲染出来的,即是不可见的。
上述实施例中,通过模拟摄像机的位置,确定三维虚拟场景中位于渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面,可以提升视频帧画面的渲染速度。
在一个实施例中,网格面片包括多个网格子面片,多个网格子面片中相邻两个网格子面片的共有顶点具有相同坐标,根据网格面片,将视频帧画面拆分,得到多个视频子画面,包括:针对每张视频帧画面,根据多个网格子面片中的共有顶点,将视频帧画面对应的渲染区域,在空间上进行拆分,获得多个子区域;确定三维虚拟场景中在每个子区域内的实时场景内容,将每个子区域内的实时场景内容分别确定为视频子画面。
其中,共有顶点是两个网格子面片之间共有的网格顶点。子区域是对渲染区域在空间上进行拆分后得到的区域。
针对每张视频帧画面,计算机设备可根据多个网格子面片中的共有顶点,将视频帧画面对应的渲染区域在空间上进行拆分,得到多个子区域。进而,计算机设备可确定三维虚拟场景中分别位于各个子区域内的实时场景内容,并确定拆分后的各个子区域内的实时场景内容为视频子画面。渲染机具有网络地址,渲染机的网络地址与网格子面片对应,网格子面片与视频子画面对应,则渲染机可渲染与其网络地址具有预设的映射关系网格子面片所对应的视频子画面。
在一个实施例中,继续参考图3,网格面片包括两个网格子面片,即网格子面片AEFD(即网格子面片1)和网格子面片EBCF(即网格子面片2)。网格子面片AEFD和网格子面片EBCF是相邻的两个网格子面片,它们有两个共有顶点,即顶点E和顶点F。计算机设备可根据网格子面片AEFD和网格子面片EBCF中的共有顶点E和F,将视频帧画面对应的渲染区域在空间上进行拆分,得到两个子区域,即四棱锥体O-AEFD和四棱锥体O-EBCF。可以理解,四棱锥体O-AEFD和四棱锥体O-EBCF内的实时场景内容为视频子画面。
上述实施例中,通过多个网格子面片中的共有顶点,将视频帧画面对应的渲染区域在空间上进行拆分,可以确保后续无缝拼接能最终合成一个完整的画面。进而通过确定拆分后的各个子区域内的实时场景内容为视频子画面,可以提升视频子画面的获取准确率。
在一个实施例中,视频画面渲染方法还包括:确定虚拟摄像机的焦距;根据焦距,确定模拟摄像机中的视点与网格面片之间的距离;根据距离,确定视点与网格面片中各个顶点之间的位置关系。
计算机设备可根据渲染引擎中虚拟摄像机的焦距,确定视点与网格面片之间的距离;根据距离,得到视点与网格面片中各个顶点之间的相对位置关系。
进一步地,计算机设备可确定虚拟摄像机的焦距,并根据渲染引擎中虚拟摄像机的焦距,确定视点与网格面片之间的距离。进而,计算机设备可根据距离,得到视点与网格面片中各个顶点之间的相对位置关系。
在一个实施例中,计算机设备可将渲染引擎中虚拟摄像机的焦距进行运算,并将运算后的距离作为视点与网格面片之间的距离。
在一个实施例中,计算机设备可将渲染引擎中虚拟摄像机的焦距,直接作为视点与网格面片之间的距离。
在一个实施例中,模拟摄像机是基于显示组件针对虚拟摄像机进行模拟得到的,模拟摄像机中视点与网格面片之间的距离,是基于虚拟摄像机的焦距确定得到的。显示组件是渲染引擎中用于渲染画面的组件。
如图4所示,计算机设备可获取虚拟摄像机的当前焦距,并将获取到的虚拟摄像机的当前焦距作为模拟摄像机中视点与网格面片之间的距离。可以理解,一旦虚拟摄像机的焦距发生了改变,模拟摄像机中视点与网格面片之间的距离也会随之改变。
上述实施例中,通过渲染引擎中虚拟摄像机的焦距,确定视点与网格面片之间的距离,并根据距离得到视点与网格面片中各个顶点之间的相对位置,可以提升视点与网格面片中各个顶点之间的相对位置的准确率,从而可以确定得到更准确的渲染区域。
在一个实施例中,模拟摄像机是通过显示组件模拟得到的,视频画面渲染方法还包括:确定虚拟摄像机的位置;将虚拟摄像机的位置赋予显示组件,得到模拟摄像机的位置。其中,显示组件是渲染引擎中用于渲染画面的组件。
在一个实施例中,模拟摄像机是通过显示组件模拟得到的;显示组件是视点和网格面片的父级;视频画面渲染方法还包括:确定虚拟摄像机的位置;将虚拟摄像机的位置赋予显示组件,得到模拟摄像机的位置。显示组件是视点和网格面片的父级,则显示组件可具有视点和网格面片的属性,将虚拟摄像机的位置赋予显示组件,模拟摄像机的位置、视点和网格面片就都可以确定。
在一个实施例中,计算机设备可通过显示组件针对虚拟摄像机进行模拟,得到模拟摄像机。计算机设备可确定虚拟摄像机的位置,并将虚拟摄像机的位置赋予显示组件,得到模拟摄像机的位置。
在一个实施例中,由于模拟摄像机是基于显示组件针对虚拟摄像机进行模拟得到的,因此,模拟摄像机位于三维虚拟场景的位置,是基于虚拟摄像机位于三维虚拟场景的位置确定得到的。如图5所示,计算机设备可获取虚拟摄像机的位置,并将获取到的虚拟摄像机的位置作为模拟摄像机的位置。可以理解,一旦虚拟摄像机位于三维虚拟场景的位置发生了改变,模拟摄像机位于三维虚拟场景的位置也会随之改变。
上述实施例中,通过将虚拟摄像机的位置赋予显示组件,可以得到模拟摄像机的位置,进而使得模拟摄像机可模拟虚拟摄影机的运动,从而可以确定得到更准确的视频帧画面。
在一个实施例中,通过预先设置的多个渲染机,同步渲染多个视频子画面,以得到视频帧画面对应的渲染画面,包括:通过预先设置的多个渲染机,将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面;将多个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面。
计算机设备中可预先设置有多个渲染机,通过预先设置的多个渲染机,计算机设备可将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面。可以理解,每个渲染机可负责渲染至少一个视频子画面。
在一个实施例中,多个渲染机可以分别接收同步渲染信号及渲染指令,通过同步渲染信号和渲染指令,控制多个渲染机分别将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面。
画面合成机可以是将多个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面的计算机程序或计算机设备。画面合成机可以是多个渲染机中的一个,如主渲染机。画面合成机也可以是独立于任一渲染机的计算机设备。
计算机设备中可预先部署有画面合成机,计算机设备可通过画面合成机接收各个渲染机输出的多个视频子画面分别对应的渲染子画面,并将多个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面。
在一个实施例中,虚拟传感器的尺寸采用物理尺寸表示,视频画面渲染方法还包括:根据虚拟传感器的尺寸和渲染机的数量进行建模,得到多个网格子面片。
在一个实施例中,网格子面片的数量与渲染机的数量一致;多个网格子面片的面积之和与虚拟传感器的面积一致。
可以理解,网格子面片的大小是基于虚拟传感器的物理尺寸均分得到的。
在一个实施例中,如图6,虚拟传感器的尺寸以物理尺寸表示为:23.76mm*13.365mm,渲染机的数量为2个,则计算机设备可建模得到物理尺寸分别为11.88mm*13.365mm的两个网格子面片。计算机设备可通过渲染机1和渲染机2,分别将这两个网格子面片分别对应的视频子画面进行同步渲染,得到渲染子画面1和渲染子画面2。
在一个实施例中,如图7所示,计算机设备可预先设置渲染机的网络地址、网格子面片和渲染机中的画面视口之间的映射关系。具体地,计算机设备可将渲染机1的网络地址192.168.0.6、网格子面片1和渲染机1中的画面视口1进行绑定。计算机设备可将渲染机2的网络地址192.168.0.7、网格子面片2和渲染机2中的画面视口2进行绑定,以便于后续通过渲染机1对网格子面片1对应的视频子画面进行渲染,以及通过渲染机2对网格子面片2对应的视频子画面进行渲染。画面视口与渲染机对应,渲染机与拆分出的视频子画面对应,该画面视口用于显示将对应视频子画面进行渲染得的的渲染子画面。
上述实施例中,通过虚拟传感器的物理尺寸和渲染机的数量进行建模,得到多个网格子面片,可以使得每一个渲染机负责渲染一个相应网格子面片对应的视频子画面,从而提升视频子画面的渲染效率。
在一个实施例中,多个渲染机包括一个主渲染机和至少一个从渲染机,主渲染机中部署有同步卡;通过预先设置的多个渲染机,将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面,包括:通过主渲染机中的同步卡,接收同步信号发生器基于预设帧率产生的同步渲染信号;通过同步卡将同步渲染信号同步至从渲染机中;通过主渲染机和从渲染机分别接收的同步渲染信号,控制主渲染机和从渲染机分别将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面。
其中,同步渲染信号,是用于指示多个渲染机对多个视频子画面进行同步渲染的同步信号。需要说明的是,同步信号发生器是一种相位同步信号发生器。多台渲染机可基于同步渲染信号实现相位级别的时间对齐。预设帧率是预先设置的帧率,可以根据需要设定,在肉眼可觉察的帧率范围内,预设帧率越高,质量越高,对硬件性能要求越高。
多个渲染机中包括一个主渲染机和至少一个从渲染机,其中,主渲染机中部署有同步卡。计算机设备中还部署有同步信号发生器,同步信号发生器可基于预设的预设帧率产生的同步渲染信号。
计算机设备可通过主渲染机中的同步卡,接收同步信号发生器基于预设的预设帧率产生的同步渲染信号。主渲染机可通过同步卡将同步渲染信号同步至各个从渲染机中。进而,计算机设备可通过主渲染机和从渲染机分别接收的同步渲染信号,及主渲染机发送的渲染指令,控制主渲染机和从渲染机分别将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面。
上述实施例中,通过主渲染机中的同步卡,接收同步信号发生器基于预设帧率产生的同步渲染信号,并通过同步卡将同步渲染信号同步至从渲染机中,使得各个渲染机可以接收到相同的同步渲染信号。进而,通过主渲染机和从渲染机分别接收的同步渲染信号,及主渲染机发送的渲染指令,控制主渲染机和从渲染机分别将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面,可以避免出现视频画面撕裂的问题,从而进一步提升最终渲染视频画面的质量。
在一个实施例中,将多个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面,包括:在视频信号采集卡接收到同步信号发生器产生的同步采集信号的情况下,通过视频信号采集卡对多个视频子画面分别对应的渲染子画面信号进行同步采集;将同步采集到的渲染子画面信号进行合成,得到视频帧画面对应的渲染画面。
其中,渲染子画面对应有渲染子画面信号。其中,同步采集信号,是用于指示多个视频信号采集卡对多路渲染子画面信号进行同步采集的同步信号。计算机设备中还部署有视频信号采集卡。
同步信号发生器可基于预设的预设帧率产生的同步采集信号,视频信号采集卡可接收同步信号发生器产生的同步采集信号。在视频信号采集卡接收到同步信号发生器产生的同步采集信号的情况下,计算机设备可通过视频信号采集卡对多个视频子画面分别对应的渲染子画面信号进行同步采集。进而,计算机设备可将同步采集到的多个视频子画面分别对应的渲染子画面信号进行合成,得到视频帧画面对应的渲染画面。
上述实施例中,在视频信号采集卡接收到同步信号发生器产生的同步采集信号的情况下,通过视频信号采集卡对多个视频子画面分别对应的渲染子画面信号进行同步采集,将同步采集到的多个视频子画面分别对应的渲染子画面信号进行合成,得到视频帧画面对应的渲染画面,这样,通过对多个渲染子画面信号进行同步采集,再将同步采集的渲染子画面信号进行合成,可以进一步避免出现视频画面撕裂的问题,从而进一步提升最终渲染视频画面的质量。
在一个实施例中,视频画面渲染方法还包括:在同步渲染得到的渲染子画面信号的信号格式与视频信号采集卡所规定的信号格式不一致的情况下,则在通过格式转换器接收到同步转换信号后,通过格式转换器,将多个视频子画面分别对应的渲染子画面信号进行格 式同步转换,得到与视频信号采集卡所规定的信号格式一致的渲染子画面信号,以供视频信号采集卡进行同步采集。
其中,同步转换信号,是用于指示多个格式转换器对多路渲染子画面信号进行格式同步转换的同步信号。
计算机设备中还部署有多个格式转换器,格式转换器的数量可以与渲染机的数量一致。同步信号发生器基于预设的预设帧率产生的同步转换信号,格式转换器可接收同步信号发生器所产生的同步转换信号。在主从渲染机同步渲染得到的渲染子画面信号的信号格式与视频信号采集卡所规定的信号格式不一致的情况下,则计算机设备可在通过格式转换器接收到同步转换信号后,通过格式转换器将多个视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与视频信号采集卡所规定的信号格式一致的渲染子画面信号,以供视频信号采集卡进行同步采集。
在一个实施例中,格式转换器的数量与渲染机的数量一致,每个格式转换器负责相应渲染机输出的渲染子画面信号的转换处理。计算机设备可通过各个格式转换器,将多个视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与视频信号采集卡所规定的信号格式一致的渲染子画面信号,以供视频信号采集卡进行同步采集。
在一个实施例中,各个渲染机所输出的渲染子画面信号的信号格式为HDMI(High Definition Multimedia Interface,高清多媒体接口)格式,视频信号采集卡所规定的信号格式为SDI(Serial Digital Interface,高清数字分量串行接口)格式,则计算机设备可在通过格式转换器接收到同步转换信号后,通过格式转换器将多个HDMI格式的视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与视频信号采集卡所规定的SDI格式一致的渲染子画面信号,以供视频信号采集卡进行同步采集。
在一个实施例中,各个渲染机所输出的渲染子画面信号的信号格式为DP格式,视频信号采集卡所规定的信号格式为SDI格式,则计算机设备可在通过格式转换器接收到同步转换信号后,通过格式转换器将多个DP(Display Port,显示端口)格式的视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与视频信号采集卡所规定的SDI格式一致的渲染子画面信号,以供视频信号采集卡进行同步采集。
上述实施例中,在主从渲染机同步渲染得到的渲染子画面信号的信号格式与视频信号采集卡所规定的信号格式不一致的情况下,则在通过格式转换器接收到同步转换信号后,通过格式转换器将多个视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与视频信号采集卡所规定的信号格式一致的渲染子画面信号,以供视频信号采集卡进行同步采集。这样,通过对多个渲染子画面信号进行同步格式转换,再对同步转换后的渲染子画面信号进行同步采集,进而再将同步采集的渲染子画面信号进行合成,可以进一步避免出现视频画面撕裂的问题,从而进一步提升最终渲染视频画面的质量。
在一个实施例中,视频信号采集卡部署在画面合成机上;画面合成机提供合成视频画布;将同步采集到的渲染子画面信号进行合成,得到视频帧画面对应的渲染画面,包括:将合成视频画布的帧率设置为预设帧率;通过满足预设帧率的合成视频画布,将将同步采集到的渲染子画面信号进行同步合成,得到视频帧画面对应的渲染画面。
具体地,计算机设备中还部署有画面合成机,画面合成机上部署有视频信号采集卡、合成软件和基于合成软件创建的合成视频画布。计算机设备可将合成视频画布的帧率设置为与同步信号发生器相同的预设帧率。进而,计算机设备可通过满足预设帧率的合成视频画布,将视频信号采集卡同步采集到的多个视频子画面分别对应的渲染子画面信号进行同步合成,得到视频帧画面对应的渲染画面。
上述实施例中,将合成视频画布的帧率设置为与同步信号发生器相同的预设帧率,并通过满足预设帧率的合成视频画布,将视频信号采集卡同步采集到的多个视频子画面分别对应的渲染子画面信号进行同步合成,得到视频帧画面对应的渲染画面,可以进一步避免 出现视频画面撕裂的问题,从而进一步提升最终渲染视频画面的质量。
在一个实施例中,同步信号发生器可产生同步信号,可以理解,同步信号发生器所产生的同步信号在被渲染机接收的情况下,该同步信号即为同步渲染信号。同步信号发生器所产生的同步信号在被格式转换器接收的情况下,该同步信号即为同步转换信号。同步信号发生器所产生的同步信号在被视频信号采集卡接收的情况下,该同步信号即为同步采集信号。需要说明的是,同步信号发生器是一种相位同步信号发生器。多个系统之间可基于接收到的同步信号实现相位级别的时间对齐。
在一个实施例中,如图8所示,计算机设备内部集成有同步信号发生器,两个渲染机(渲染机1和渲染机2)、两个格式转换器(格式转换器1和格式转换器2)和部署有视频信号采集卡的画面合成机。渲染机1中集成有同步卡。同步信号发生器可产生同步渲染信号、同步转换信号和同步采集信号。多个视频子画面具体为两个视频子画面。渲染机1可通过同步卡接收同步信号发生器所产生同步渲染信号,并将同步渲染信号同步至渲染机2。通过渲染机1和渲染机2分别接收的同步渲染信号,控制渲染机1和渲染机2分别将两个视频子画面进行同步渲染,得到两个视频子画面分别对应的渲染子画面信号。在渲染机1和渲染机2同步渲染得到的渲染子画面信号的信号格式与视频信号采集卡所规定的信号格式不一致的情况下,则通过格式转换器1对渲染机1输出的渲染子画面信号进行格式转换,通过格式转换器2对渲染机2输出的渲染子画面信号进行格式转换,得到与视频信号采集卡所规定的信号格式一致的渲染子画面信号。在视频信号采集卡接收到同步信号发生器产生的同步采集信号的情况下,通过视频信号采集卡对上述两个视频子画面分别对应的渲染子画面信号进行同步采集,并将同步采集到的两个视频子画面分别对应的渲染子画面信号进行合成,得到视频帧画面对应的渲染画面。可以理解,输出的渲染画面可应用于多个业务场景中。
在一个实施例中,如图9所示,计算机设备可根据模拟摄像机中的视点与网格面片中各个顶点之间的相对位置,确定渲染区域,根据模拟摄像机的位置,确定渲染区域内待渲染的实时的视频帧画面。针对每张视频帧画面,计算机设备可根据网格面片将视频帧画面在空间上进行区域拆分,得到N个视频子画面。计算机设备可通过预先设置的N个渲染机,将N个视频子画面进行同步渲染,得到N个视频子画面分别对应的渲染子画面。进而,计算机设备可通过画面合成机将N个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面。其中,N为大于2的正整数,且是常数。
在一个实施例中,在虚拟对象实时渲染场景下,视频帧画面包括虚拟对象实时画面,视频子画面包括实时子画面。获取所述渲染区域内待渲染的视频帧画面,包括:获取所述模拟摄像机的位置,根据所述模拟摄像机的位置,确定所述渲染区域内待渲染的实时的虚拟对象实时画面。所述将所述多个视频子画面分别对应的渲染子画面进行合成,得到所述视频帧画面对应的渲染画面,包括:将多个所述实时子画面分别对应的渲染子画面进行合成,得到所述虚拟对象实时画面对应的渲染画面。
在一个实施例中,根据虚拟摄像机的位置,确定渲染区域内待渲染的视频帧画面包括:根据模拟摄像机的位置,确定渲染区域内待渲染的实时的虚拟对象实时画面。
其中,虚拟对象实时画面,是在虚拟对象实时渲染场景下确定得到的视频画面。实时子画面,是将虚拟对象实时画面在空间上进行区域拆分之后得到的视频画面。虚拟对象,是虚拟的实体对象,具体可包括虚拟人物、虚拟动物和虚拟物体中的至少一种。
计算机设备可根据模拟摄像机中的视点与网格面片中各个顶点之间的相对位置,确定渲染区域。在虚拟对象实时渲染场景下,计算机设备可根据模拟摄像机的位置,确定渲染区域内待渲染的实时的虚拟对象实时画面。针对每张虚拟对象实时画面,计算机设备可根据网格面片将虚拟对象实时画面在空间上进行区域拆分,得到多个实时子画面,并通过预先设置的多个渲染机,将多个实时子画面进行同步渲染,得到多个实时子画面分别对应的 渲染子画面。进而,计算机设备可将多个实时子画面分别对应的渲染子画面进行合成,得到虚拟对象实时画面对应的渲染画面。
上述实施例中,在虚拟对象实时渲染场景下,通过模拟摄像机的位置,确定渲染区域内待渲染的实时的虚拟对象实时画面,可以提升虚拟对象实时画面的获取准确率。通过将多个实时子画面分别对应的渲染子画面进行合成,得到虚拟对象实时画面对应的渲染画面,可以提升虚拟对象实时渲染场景下渲染画面的指令。
在一个实施例中,如图10所示,场景(a)和场景(b)中都存在大量的复杂场景元素,比如,灯光和毛发等。若通过单个渲染机对场景(a)和场景(b)中的画面进行渲染,渲染机的渲染压力较大。因此,通过本申请的视频画面渲染方法,根据模拟摄像机中的视点与网格面片中各个顶点之间的相对位置,确定渲染区域,根据模拟摄像机的位置,确定渲染区域内待渲染的实时的视频帧画面。由于模拟摄像机是针对渲染引擎中虚拟摄像机进行模拟得到的,以及网格面片是基于虚拟摄像机中虚拟传感器的物理尺寸构建得到的、可用于做画面拆分处理的面片。因此,针对每张视频帧画面,根据网格面片将视频帧画面在空间上进行区域拆分,得到多个视频子画面。通过预先设置的多个渲染机,将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面,将多个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面。相较于传统的离线渲染方式,本申请通过基于虚拟摄像机中虚拟传感器的物理尺寸构建得到的网格面片,对视频帧画面在空间上进行区域拆分,并通过多个渲染机对拆分的子画面进行同步渲染,将渲染压力分配至多个渲染机中,可以获取到高质量的实时的视频画面,满足了实时的视频画面渲染场景的需求。
如图11所示,在一个实施例中,提供了一种视频画面渲染方法,该方法可应用于计算机设备,计算机设备可以是终端或服务器,由终端或服务器自身单独执行,也可以通过终端和服务器之间的交互来实现。本实施例以该方法应用于计算机设备为例进行说明,该方法具体包括以下步骤:
步骤1102,根据虚拟传感器以物理尺寸表示的尺寸和渲染机的数量进行建模,得到多个网格子面片;其中,网格子面片的数量与渲染机的数量一致;多个网格子面片的面积之和与虚拟传感器的面积一致。
步骤1104,根据渲染引擎中虚拟摄像机的焦距,确定视点与多个网格子面片之间的距离。
步骤1106,根据距离,得到视点与多个网格子面片中各个顶点之间的相对位置。
步骤1108,根据模拟摄像机中的视点与各个网格子面片中各个顶点之间的相对位置,确定渲染区域;模拟摄像机是通过显示组件针对渲染引擎中虚拟摄像机进行模拟得到的。
步骤1110,确定虚拟摄像机的位置,将虚拟摄像机的位置赋予显示组件,得到模拟摄像机的位置。
步骤1112,根据模拟摄像机的位置,确定三维虚拟场景中位于渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面。
步骤1114,针对每张视频帧画面,根据多个网格子面片中的共有顶点,将视频帧画面对应的渲染区域在空间上进行拆分,并确定拆分后的各个子区域内的实时场景内容为视频子画面。
步骤1116,通过主渲染机中的同步卡,接收同步信号发生器基于预设的预设帧率产生的同步渲染信号。
步骤1118,通过同步卡将同步渲染信号同步至各个从渲染机中。
步骤1120,通过主渲染机和各个从渲染机分别接收的同步渲染信号,控制主渲染机和从渲染机分别将多个视频子画面进行同步渲染,得到多个视频子画面分别对应的渲染子画面。
步骤1122,通过画面合成机,将多个视频子画面分别对应的渲染子画面进行合成,得到视频帧画面对应的渲染画面。
本申请还提供一种应用场景,该应用场景应用上述的视频画面渲染方法。具体地,该视频画面渲染方法可应用于针对虚拟对象直播的视频画面渲染场景。计算机设备可根据虚拟传感器以物理尺寸表示的尺寸和渲染机的数量进行建模,得到多个网格子面片;其中,网格子面片的数量与渲染机的数量一致;多个网格子面片的面积之和与虚拟传感器的面积一致。根据渲染引擎中虚拟摄像机的焦距,确定视点与多个网格子面片之间的距离。根据距离,得到视点与多个网格子面片中各个顶点之间的相对位置。根据模拟摄像机中的视点与各个网格子面片中各个顶点之间的相对位置,确定渲染区域;模拟摄像机是通过显示组件针对渲染引擎中虚拟摄像机进行模拟得到的。确定虚拟摄像机的位置,将虚拟摄像机的位置赋予显示组件,得到模拟摄像机的位置。根据模拟摄像机的位置,确定三维虚拟场景中位于渲染区域内的实时场景内容,得到待渲染的实时的虚拟对象直播画面。
针对每张虚拟对象直播画面,计算机设备可根据多个网格子面片中的共有顶点,将虚拟对象直播画面对应的渲染区域在空间上进行拆分,并确定拆分后的各个子区域内的实时场景内容为直播子画面。通过主渲染机中的同步卡,接收同步信号发生器基于预设的预设帧率产生的同步渲染信号。通过同步卡将同步渲染信号同步至各个从渲染机中。通过主渲染机和各个从渲染机分别接收的同步渲染信号,控制主渲染机和从渲染机分别将多个直播子画面进行同步渲染,得到多个直播子画面分别对应的渲染子画面。将多个直播子画面分别对应的渲染子画面进行合成,得到虚拟对象直播画面对应的渲染画面。本申请通过基于虚拟摄像机中虚拟传感器的物理尺寸构建得到的网格面片,对虚拟对象直播画面在空间上进行区域拆分,并通过多个渲染机对拆分的直播子画面进行同步渲染,可以获取到高质量的实时的虚拟对象直播画面,满足了实时的虚拟对象直播画面渲染场景的需求。
本申请还另外提供一种应用场景,该应用场景应用上述的视频画面渲染方法。具体地,该视频画面渲染方法可应用于针对XR(Extended Reality,扩展现实)直播的视频画面渲染场景。其中,扩展现实(XR),是指通过计算机将真实与虚拟相结合,打造一个可人机交互的虚拟环境,可以为体验者带来虚拟世界与现实世界之间无缝转换的沉浸感。通过本申请的视频画面渲染方法,基于虚拟摄像机中虚拟传感器以物理尺寸表示的尺寸构建得到的网格面片,对扩展现实直播画面在空间上进行区域拆分,并通过多个渲染机对拆分的直播子画面进行同步渲染,可以获取到高质量的实时的扩展现实直播画面,满足了实时的扩展现实直播画面渲染场景的需求。
应该理解的是,虽然上述各实施例的流程图中的各个步骤按照顺序依次显示,但是这些步骤并不是必然按照顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,上述各实施例中的至少一部分步骤可以包括多个子步骤或者多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些子步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤的子步骤或者阶段的至少一部分轮流或者交替地执行。
在一个实施例中,如图12所示,提供了一种视频画面渲染装置1200,该装置可以采用软件模块或硬件模块,或者是二者的结合成为计算机设备的一部分,该装置具体包括:
确定模块1202,用于确定模拟摄像机,所述模拟摄像机是模拟渲染引擎的虚拟摄像机得到的;确定网格面片,所述网格面片与所述虚拟摄像机中虚拟传感器的尺寸匹配;确定所述模拟摄像机中的视点与所述网格面片中各个顶点之间的位置关系,根据所述位置关系确定渲染区域;获取所述渲染区域内待渲染的视频帧画面。
拆分模块1204,用于根据所述网格面片,将所述视频帧画面进行拆分,得到多个视频子画面。
渲染模块1206,用于通过预先设置的多个渲染机,同步渲染所述多个视频子画面,得 到所述视频帧画面对应的渲染画面。
在一个实施例中,所述渲染区域,是当所述视点分别与所述网格面片的每个顶点相连构成锥体时,所述椎体内部的区域。
在一个实施例中,确定模块1202还用于获取所述模拟摄像机的位置,根据所述虚拟摄像机的位置,确定所述渲染区域内待渲染的视频帧画面。
在一个实施例中,虚拟摄像机在三维虚拟场景中,所述模拟摄像机的位置,是根据所述虚拟摄像机在所述三维虚拟场景中的位置确定的,确定模块1202还用于根据所述模拟摄像机的位置,确定所述三维虚拟场景中位于所述渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面。
在一个实施例中,所述网格面片包括多个网格子面片,所述多个网格子面片中相邻两个网格子面片的共有顶点具有相同坐标,拆分模块1204还用于针对每张所述视频帧画面,根据所述多个网格子面片中的共有顶点,将所述视频帧画面对应的所述渲染区域,在空间上进行拆分,获得多个子区域;确定所述三维虚拟场景中在每个所述子区域内的实时场景内容,将每个所述子区域内的实时场景内容分别确定为视频子画面。
在一个实施例中,确定模块1202还用于确定虚拟摄像机的焦距;根据所述焦距,确定所述模拟摄像机中的视点与所述网格面片之间的距离;根据所述距离,确定所述视点与所述网格面片中各个顶点之间的位置关系。
在一个实施例中,所述模拟摄像机是通过显示组件模拟得到的,确定模块1202还用于确定所述虚拟摄像机的位置;将所述虚拟摄像机的位置赋予所述显示组件,得到所述模拟摄像机的位置。
在一个实施例中,渲染模块1206还用于通过预先设置的多个渲染机,将所述多个视频子画面进行同步渲染,得到所述多个视频子画面分别对应的渲染子画面。视频画面渲染装置1200还包括合成模块,用于将所述多个视频子画面分别对应的渲染子画面进行合成,得到所述视频帧画面对应的渲染画面。
所述虚拟传感器的尺寸采用物理尺寸表示,视频画面渲染装置1200还包括构建模块,用于根据所述虚拟传感器的尺寸和所述渲染机的数量进行建模,得到所述多个网格子面片。
在一个实施例中,所述网格子面片的数量与所述渲染机的数量一致;所述多个网格子面片的面积之和与所述虚拟传感器的面积一致。
在一个实施例中,所述多个渲染机包括一个主渲染机和至少一个从渲染机,所述主渲染机中部署有同步卡;渲染模块1206还用于通过所述主渲染机中的同步卡,接收同步信号发生器基于预设帧率产生的同步渲染信号;通过所述同步卡将所述同步渲染信号同步至所述从渲染机中;通过所述主渲染机和所述从渲染机分别接收的同步渲染信号,控制所述主渲染机和所述从渲染机分别将所述多个视频子画面进行同步渲染,得到所述多个视频子画面分别对应的渲染子画面。
在一个实施例中,渲染模块1206还用于在视频信号采集卡接收到所述同步信号发生器产生的同步采集信号的情况下,通过所述视频信号采集卡对所述多个视频子画面分别对应的渲染子画面信号进行同步采集;将同步采集到的渲染子画面信号进行合成,得到所述视频帧画面对应的渲染画面。
在一个实施例中,视频画面渲染装置1200还包括转换模块,用于在同步渲染得到的渲染子画面信号的信号格式与所述视频信号采集卡所规定的信号格式不一致的情况下,则在通过所述格式转换器接收到同步转换信号后,通过所述格式转换器,将所述多个视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与所述视频信号采集卡所规定的信号格式一致的渲染子画面信号,以供所述视频信号采集卡进行同步采集。
在一个实施例中,所述视频信号采集卡部署在画面合成机上;所述画面合成机提供合成视频画布;合成模块,还用于将所述合成视频画布的帧率设置为所述预设帧率;通过满 足所述预设帧率的所述合成视频画布,将将同步采集到的渲染子画面信号进行同步合成,得到所述视频帧画面对应的渲染画面。
在一个实施例中,在虚拟对象实时渲染场景下,所述视频帧画面包括虚拟对象实时画面,所述视频子画面包括实时子画面,确定模块1202还用根据所述模拟摄像机的位置,确定所述渲染区域内待渲染的实时的虚拟对象实时画面;合成模块还用于将多个所述实时子画面分别对应的渲染子画面进行合成,得到所述虚拟对象实时画面对应的渲染画面。
上述视频画面渲染装置,通过与虚拟摄像机中虚拟传感器的尺寸匹配的网格面片,对视频帧画面进行拆分,并通过预先设置的多个渲染机,对拆分的子画面进行同步渲染,可以高效地得到高质量的渲染画面,满足了对画面质量和渲染效率高的场景下的渲染需求。比如,画面尺寸越大,采用上述渲染画面渲染方法,相比传统方法效果则更加突出。还比如,对实时性要求较高的场景,如渲染实时的视频画面的场景,采用上述视频画面渲染方法能够更好地满足视频画面渲染的高实时性要求。特别低,对于建筑场所内部或外部设置的屏幕,商用巨型屏幕,采用上述视频画面渲染方法效果特别突出。
上述视频画面渲染装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。
在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务器,也可以是终端,其内部结构图可以如图13所示。该计算机设备包括处理器、存储器、输入/输出接口(Input/Output,简称I/O)和通信接口。其中,处理器、存储器和输入/输出接口通过系统总线连接,通信接口通过输入/输出接口连接到系统总线。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质和内存储器。该非易失性存储介质存储有操作系统、计算机可读指令。该内存储器为非易失性存储介质中的操作系统和计算机可读指令的运行提供环境。该计算机设备的输入/输出接口用于处理器与外部设备之间交换信息。该计算机设备的通信接口用于与外部的终端通过网络连接通信。该计算机可读指令被处理器执行时以实现一种视频画面渲染方法。
本领域技术人员可以理解,图13中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。
在一个实施例中,还提供了一种计算机设备,包括存储器和处理器,存储器中存储有计算机可读指令,该处理器执行计算机可读指令时实现上述各方法实施例中的步骤。
在一个实施例中,提供了一种计算机可读存储介质,存储有计算机可读指令,该计算机可读指令被处理器执行时实现上述各方法实施例中的步骤。
在一个实施例中,提供了一种计算机程序产品,包括计算机可读指令,计算机可读指令被处理器执行时实现上述各方法实施例中的步骤。
需要说明的是,本申请所涉及的用户信息(包括但不限于用户设备信息、用户个人信息等)和数据(包括但不限于用于分析的数据、存储的数据、展示的数据等),均为经用户授权或者经过各方充分授权的信息和数据,且相关数据的收集、使用和处理需要遵守相关国家和地区的相关法律法规和标准。
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和易失性存储器中的至少一种。非易失性存储器可包括只读存储器(Read-Only Memory,ROM)、磁带、软盘、闪存或光存储器等。易失性存储器可包 括随机存取存储器(Random Access Memory,RAM)或外部高速缓冲存储器。作为说明而非局限,RAM可以是多种形式,比如静态随机存取存储器(Static Random Access Memory,SRAM)或动态随机存取存储器(Dynamic Random Access Memory,DRAM)等。
以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。
以上所述实施例仅表达了本申请的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本申请构思的前提下,还可以做出若干变形和改进,这些都属于本申请的保护范围。因此,本申请专利的保护范围应以所附权利要求为准。

Claims (19)

  1. 一种视频画面渲染方法,由计算机设备执行,包括:
    确定模拟摄像机,所述模拟摄像机是模拟渲染引擎的虚拟摄像机得到的;
    确定网格面片,所述网格面片与所述虚拟摄像机中虚拟传感器的尺寸匹配;
    确定所述模拟摄像机中的视点与所述网格面片中各个顶点之间的位置关系,根据所述位置关系确定渲染区域;
    获取所述渲染区域内待渲染的视频帧画面;
    根据所述网格面片,将所述视频帧画面进行拆分,得到多个视频子画面;
    通过预先设置的多个渲染机,同步渲染所述多个视频子画面,得到所述视频帧画面对应的渲染画面。
  2. 根据权利要求1所述的方法,所述渲染区域,是当所述视点分别与所述网格面片的每个顶点相连构成锥体时,所述椎体内部的区域。
  3. 根据权利要求1或2所述的方法,所述获取所述渲染区域内待渲染的视频帧画面,包括:
    获取所述模拟摄像机的位置,根据所述虚拟摄像机的位置,确定所述渲染区域内待渲染的视频帧画面。
  4. 根据权利要求3所述的方法,所述虚拟摄像机在三维虚拟场景中,所述模拟摄像机的位置,是根据所述虚拟摄像机在所述三维虚拟场景中的位置确定的,所述根据所述虚拟摄像机的位置,确定所述渲染区域内待渲染的视频帧画面包括:
    根据所述模拟摄像机的位置,确定所述三维虚拟场景中位于所述渲染区域内的实时场景内容,得到待渲染的实时的视频帧画面。
  5. 根据权利要求1至4中任一项所述的方法,所述网格面片包括多个网格子面片,所述多个网格子面片中相邻两个网格子面片的共有顶点具有相同坐标,所述根据所述网格面片,将所述视频帧画面拆分,得到多个视频子画面,包括:
    针对每张所述视频帧画面,根据所述多个网格子面片中的共有顶点,将所述视频帧画面对应的所述渲染区域,在空间上进行拆分,获得多个子区域;
    确定所述三维虚拟场景中在每个所述子区域内的实时场景内容,将每个所述子区域内的实时场景内容分别确定为视频子画面。
  6. 根据权利要求1至5中任一项所述的方法,所述方法还包括:
    确定虚拟摄像机的焦距;
    根据所述焦距,确定所述模拟摄像机中的视点与所述网格面片之间的距离;
    根据所述距离,确定所述视点与所述网格面片中各个顶点之间的位置关系。
  7. 根据权利要求1至6中任一项所述的方法,所述模拟摄像机是通过显示组件模拟得到的,所述方法还包括:
    确定所述虚拟摄像机的位置;
    将所述虚拟摄像机的位置赋予所述显示组件,得到所述模拟摄像机的位置。
  8. 根据权利要求1至7中任一项所述的方法,所述通过预先设置的多个渲染机,同步渲染所述多个视频子画面,以得到所述视频帧画面对应的渲染画面,包括:
    通过预先设置的多个渲染机,将所述多个视频子画面进行同步渲染,得到所述多个视频子画面分别对应的渲染子画面;
    将所述多个视频子画面分别对应的渲染子画面进行合成,得到所述视频帧画面对应的渲染画面。
  9. 根据权利要求1至8中任一项所述的方法,所述虚拟传感器的尺寸采用物理尺寸表示,所述方法还包括:
    根据所述虚拟传感器的尺寸和所述渲染机的数量进行建模,得到所述多个网格子面片。
  10. 根据权利要求9所述的方法,所述网格子面片的数量与所述渲染机的数量一致;所述多个网格子面片的面积之和与所述虚拟传感器的面积一致。
  11. 根据权利要求8至10中任一项所述的方法,所述多个渲染机包括一个主渲染机和至少一个从渲染机,所述主渲染机中部署有同步卡;
    所述通过预先设置的多个渲染机,将所述多个视频子画面进行同步渲染,得到所述多个视频子画面分别对应的渲染子画面,包括:
    通过所述主渲染机中的同步卡,接收同步信号发生器基于预设帧率产生的同步渲染信号;
    通过所述同步卡将所述同步渲染信号同步至所述从渲染机中;
    通过所述主渲染机和所述从渲染机分别接收的同步渲染信号,控制所述主渲染机和所述从渲染机分别将所述多个视频子画面进行同步渲染,得到所述多个视频子画面分别对应的渲染子画面。
  12. 根据权利要求11所述的方法,所述将所述多个视频子画面分别对应的渲染子画面进行合成,得到所述视频帧画面对应的渲染画面,包括:
    在视频信号采集卡接收到所述同步信号发生器产生的同步采集信号的情况下,通过所述视频信号采集卡对所述多个视频子画面分别对应的渲染子画面信号进行同步采集;
    将同步采集到的渲染子画面信号进行合成,得到所述视频帧画面对应的渲染画面。
  13. 根据权利要求12所述的方法,所述方法还包括:
    在同步渲染得到的渲染子画面信号的信号格式与所述视频信号采集卡所规定的信号格式不一致的情况下,则在通过所述格式转换器接收到同步转换信号后,通过所述格式转换器,将所述多个视频子画面分别对应的渲染子画面信号进行格式同步转换,得到与所述视频信号采集卡所规定的信号格式一致的渲染子画面信号,以供所述视频信号采集卡进行同步采集。
  14. 根据权利要求12或13所述的方法,所述视频信号采集卡部署在画面合成机上;所述画面合成机提供合成视频画布;
    所述将同步采集到的渲染子画面信号进行合成,得到所述视频帧画面对应的渲染画面,包括:
    将所述合成视频画布的帧率设置为所述预设帧率;
    通过满足所述预设帧率的所述合成视频画布,将将同步采集到的渲染子画面信号进行同步合成,得到所述视频帧画面对应的渲染画面。
  15. 根据权利要求8至11任一项所述的方法,在虚拟对象实时渲染场景下,所述视频帧画面包括虚拟对象实时画面,所述视频子画面包括实时子画面,所述获取所述渲染区域内待渲染的视频帧画面包括:
    获取所述模拟摄像机的位置,根据所述模拟摄像机的位置,确定所述渲染区域内待渲染的实时的虚拟对象实时画面;
    所述将所述多个视频子画面分别对应的渲染子画面进行合成,得到所述视频帧画面对应的渲染画面,包括:
    将多个所述实时子画面分别对应的渲染子画面进行合成,得到所述虚拟对象实时画面对应的渲染画面。
  16. 一种视频画面渲染装置,所述装置包括:
    确定模块,用于确定模拟摄像机,所述模拟摄像机是模拟渲染引擎的虚拟摄像机得到的;确定网格面片,所述网格面片与所述虚拟摄像机中虚拟传感器的尺寸匹配;确定所述模拟摄像机中的视点与所述网格面片中各个顶点之间的位置关系,根据所述位置关系确定渲染区域;获取所述渲染区域内待渲染的视频帧画面;
    拆分模块,用于根据所述网格面片,将所述视频帧画面拆分,得到多个视频子画面;
    渲染模块,用于同步渲染所述多个视频子画面,得到所述视频帧画面对应的渲染画面面。
  17. 一种计算机设备,包括存储器和处理器,所述存储器存储有计算机可读指令,所述处理器执行所述计算机可读指令时执行权利要求1至15中任一项所述的方法。
  18. 一种计算机可读存储介质,存储有计算机可读指令,所述计算机可读指令被处理器执行时执行权利要求1至15中任一项所述的方法。
  19. 一种计算机程序产品,包括计算机可读指令,所述计算机可读指令被处理器执行时执行权利要求1至15中任一项所述的方法。
PCT/CN2023/116670 2022-10-28 2023-09-04 视频画面渲染方法、装置、设备和介质 WO2024087883A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211335718.7A CN116485966A (zh) 2022-10-28 2022-10-28 视频画面渲染方法、装置、设备和介质
CN202211335718.7 2022-10-28

Publications (1)

Publication Number Publication Date
WO2024087883A1 true WO2024087883A1 (zh) 2024-05-02

Family

ID=87220116

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/116670 WO2024087883A1 (zh) 2022-10-28 2023-09-04 视频画面渲染方法、装置、设备和介质

Country Status (2)

Country Link
CN (1) CN116485966A (zh)
WO (1) WO2024087883A1 (zh)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116485966A (zh) * 2022-10-28 2023-07-25 腾讯科技(深圳)有限公司 视频画面渲染方法、装置、设备和介质
CN116866621B (zh) * 2023-09-05 2023-11-03 湖南马栏山视频先进技术研究院有限公司 一种面向视频实时渲染的云端同步方法及系统

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170243319A1 (en) * 2016-02-18 2017-08-24 Nvidia Corporation Sub-frame scanout for latency reduction in virtual reality applications
CN112001993A (zh) * 2020-07-14 2020-11-27 深圳市规划国土房产信息中心(深圳市空间地理信息中心) 一种面向大场景的多gpu城市仿真系统
CN114255315A (zh) * 2020-09-25 2022-03-29 华为云计算技术有限公司 一种渲染方法、装置及设备
CN114494559A (zh) * 2022-02-15 2022-05-13 成都引力视创科技有限公司 一种基于多gpu协同的三维渲染融合方法、系统、介质
CN114820910A (zh) * 2021-01-21 2022-07-29 华为云计算技术有限公司 一种渲染方法及装置
CN116485966A (zh) * 2022-10-28 2023-07-25 腾讯科技(深圳)有限公司 视频画面渲染方法、装置、设备和介质

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170243319A1 (en) * 2016-02-18 2017-08-24 Nvidia Corporation Sub-frame scanout for latency reduction in virtual reality applications
CN112001993A (zh) * 2020-07-14 2020-11-27 深圳市规划国土房产信息中心(深圳市空间地理信息中心) 一种面向大场景的多gpu城市仿真系统
CN114255315A (zh) * 2020-09-25 2022-03-29 华为云计算技术有限公司 一种渲染方法、装置及设备
CN114820910A (zh) * 2021-01-21 2022-07-29 华为云计算技术有限公司 一种渲染方法及装置
CN114494559A (zh) * 2022-02-15 2022-05-13 成都引力视创科技有限公司 一种基于多gpu协同的三维渲染融合方法、系统、介质
CN116485966A (zh) * 2022-10-28 2023-07-25 腾讯科技(深圳)有限公司 视频画面渲染方法、装置、设备和介质

Also Published As

Publication number Publication date
CN116485966A (zh) 2023-07-25

Similar Documents

Publication Publication Date Title
WO2024087883A1 (zh) 视频画面渲染方法、装置、设备和介质
US11019259B2 (en) Real-time generation method for 360-degree VR panoramic graphic image and video
CN102834849B (zh) 进行立体视图像的描绘的图像描绘装置、图像描绘方法、图像描绘程序
US20180241986A1 (en) Image processing method and device
WO2017113488A1 (zh) 一种虚拟现实设备中显示2d应用界面的方法和装置
WO2022048097A1 (zh) 一种基于多显卡的单帧画面实时渲染方法
WO2017113681A1 (zh) 一种基于虚拟现实技术的视频图像处理方法及装置
CN113436343B (zh) 用于虚拟演播厅的画面生成方法及装置、介质及电子设备
WO2021135320A1 (zh) 一种视频的生成方法、装置及计算机系统
WO2006114898A1 (en) 3d image generation and display system
WO2021171982A1 (ja) 画像処理装置、3dモデルの生成方法、学習方法およびプログラム
JP2006094458A (ja) 映像信号処理装置および仮想現実感生成装置ならびに記録媒体
JP2019191989A (ja) 仮想視点画像を生成するシステム、方法及びプログラム
CN213461894U (zh) 一种xr-扩展现实系统
KR20160003614A (ko) 3차원 리얼타임 가상입체 스튜디오 장치에서의 가상입체 스튜디오 영상 생성 방법
WO2022191070A1 (ja) 3dオブジェクトのストリーミング方法、装置、及びプログラム
Schreer et al. Advanced volumetric capture and processing
CN110213640B (zh) 虚拟物品的生成方法、装置及设备
WO2022024780A1 (ja) 情報処理装置、情報処理方法、映像配信方法及び情報処理システム
US7009606B2 (en) Method and apparatus for generating pseudo-three-dimensional images
JP2012221092A (ja) 画像処理装置、画像処理方法およびプログラム
KR102561903B1 (ko) 클라우드 서버를 이용한 ai 기반의 xr 콘텐츠 서비스 방법
JP3631117B2 (ja) 映像処理装置及びその方法並びに記憶媒体
CN113947671A (zh) 全景360度图像分割合成方法、系统及介质
KR20170059310A (ko) 텔레 프레젠스 영상 송신 장치, 텔레 프레젠스 영상 수신 장치 및 텔레 프레젠스 영상 제공 시스템

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23881464

Country of ref document: EP

Kind code of ref document: A1