IL313019A - Adaptive energy-saving configurations for outsourcing systems - Google Patents

Adaptive energy-saving configurations for outsourcing systems

Info

Publication number
IL313019A
IL313019A IL313019A IL31301924A IL313019A IL 313019 A IL313019 A IL 313019A IL 313019 A IL313019 A IL 313019A IL 31301924 A IL31301924 A IL 31301924A IL 313019 A IL313019 A IL 313019A
Authority
IL
Israel
Prior art keywords
indicator
layers
render
reprojection
engine
Prior art date
Application number
IL313019A
Other languages
Hebrew (he)
Inventor
Wesley James Holland
Cullum James Baldwin
Dam Backer
David Berrebi
Girish Bhat
Krishnaprasad Vinay Melkote
Walker Curtis
Gal Tzvi Graif
Roee Hardoon
Original Assignee
Qualcomm Inc
Wesley James Holland
Cullum James Baldwin
Dam Backer
David Berrebi
Girish Bhat
Krishnaprasad Vinay Melkote
Walker Curtis
Gal Tzvi Graif
Roee Hardoon
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qualcomm Inc, Wesley James Holland, Cullum James Baldwin, Dam Backer, David Berrebi, Girish Bhat, Krishnaprasad Vinay Melkote, Walker Curtis, Gal Tzvi Graif, Roee Hardoon filed Critical Qualcomm Inc
Priority to IL313019A priority Critical patent/IL313019A/en
Priority to PCT/US2025/027792 priority patent/WO2025244825A1/en
Publication of IL313019A publication Critical patent/IL313019A/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4122Peripherals receiving signals from specially adapted client devices additional display device, e.g. video projector
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • G06T1/20Processor architectures; Processor configuration, e.g. pipelining
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • G09G5/363Graphics controllers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • H04N21/41407Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance embedded in a portable device, e.g. video client on a mobile phone, PDA, laptop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/426Internal components of the client ; Characteristics thereof
    • H04N21/42653Internal components of the client ; Characteristics thereof for processing graphics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • H04N21/4355Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream involving reformatting operations of additional data, e.g. HTML pages on a television screen
    • H04N21/4356Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream involving reformatting operations of additional data, e.g. HTML pages on a television screen by altering the spatial resolution, e.g. to reformat additional data on a handheld device, attached to the STB
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/436Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
    • H04N21/4363Adapting the video stream to a specific local network, e.g. a Bluetooth® network
    • H04N21/43637Adapting the video stream to a specific local network, e.g. a Bluetooth® network involving a wireless protocol, e.g. Bluetooth, RF or wireless LAN [IEEE 802.11]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2350/00Solving problems of bandwidth in display systems
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2360/00Aspects of the architecture of display systems
    • G09G2360/08Power processing, i.e. workload management for processors involved in display operations, such as CPUs or GPUs

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • Image Generation (AREA)

Description

ADAPTIVE POWER-SAVING CONFIGURATIONS FOR REPROJECTION SYSTEMS TECHNICAL FIELD
[0001] The present disclosure relates generally to processing systems, and more particularly, to one or more techniques for graphics processing. INTRODUCTION [0002] Computing devices often perform graphics and/or display processing (e.g., utilizing a graphics processing unit (GPU), a central processing unit (CPU), a display processor, etc.) to render and display visual content. Such computing devices may include, for example, computer workstations, mobile phones such as smartphones, embedded systems, personal computers, tablet computers, and video game consoles. GPUs are configured to execute a graphics processing pipeline that includes one or more processing stages, which operate together to execute graphics processing commands and output a frame. A central processing unit (CPU) may control the operation of the GPU by issuing one or more graphics processing commands to the GPU. Modern day CPUs are typically capable of executing multiple applications concurrently, each of which may need to utilize the GPU during execution. A display processor may be configured to convert digital information received from a CPU to analog values and may issue commands to a display panel for displaying the visual content. A device that provides content for visual presentation on a display may utilize a CPU, a GPU, and/or a display processor. [0003] Current techniques may not address hardware limitations of reprojection hardware, rendering units (e.g., GPUs), or wireless devices. It may be beneficial to provide improved rendering configuration techniques to optimize rendered layers prior to reprojection and/or composition of the rendered layers.
BRIEF SUMMARY
[0004] The following presents a simplified summary of one or more aspects in order to provide a basic understanding of such aspects. This summary is not an extensive overview of all contemplated aspects, and is intended to neither identify key or critical elements of all aspects nor delineate the scope of any or all aspects. Its sole purpose is to present some concepts of one or more aspects in a simplified form as a prelude to the more detailed description that is presented later. [0005] In an aspect of the disclosure, a method, a computer-readable medium, and an apparatus are provided. The apparatus may include a memory; and at least one processor coupled to the memory. Based at least in part on information stored in the memory, the at least one processor may be configured to obtain a first indicator of a plurality of layers of graphics content. The at least one processor may be configured to obtain a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device. The at least one processor may be configured to configure at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device. The at least one processor may be configured to output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. [0006] In some aspects, the techniques described herein relate to a method of graphics processing, including: obtaining a first indicator of a plurality of layers of graphics content; obtaining a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device; configuring at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device; and outputting a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. [0007] In some aspects, the techniques described herein relate to a method, where configuring at least one of the render rate or the render resolution includes configuring a remote compression setting, further including: outputting a third indication of the configured remote compression setting. [0008] In some aspects, the techniques described herein relate to a method, where outputting the third indication of the configured remote compression setting includes: transmitting, to a second device remote from the device, the third indication of the configured remote compression setting.
[0009] In some aspects, the techniques described herein relate to a method, where outputting the third indication of the configured remote compression setting includes: outputting, to a compression unit of the device, the third indication of the configured remote compression setting. [0010] In some aspects, the techniques described herein relate to a method, where the first set of characteristics includes at least one of: a third indicator of a downscale capability for the reprojection engine; a fourth indicator of an upscale capability for the reprojection engine; a fifth indicator of a rotation capability for the reprojection engine; or a sixth indicator of a warp capability for the reprojection engine. [0011] In some aspects, the techniques described herein relate to a method, where the second set of content statistics includes at least one of: a third indicator of a timestamp associated with a layer rendered by a rendering unit; or a fourth indicator of a head pose associated with the layer rendered by the rendering unit. [0012] In some aspects, the techniques described herein relate to a method, where the third set of wireless characteristics includes at least one of: a third indicator of an amount of data wirelessly transmitted to the device; a fourth indicator of a fourth set of transmission windows used to transmit wirelessly transmit the amount of data to the device; a fifth indicator of a wireless transmission capacity associated with transmission of rendered layers to the device; or a sixth indicator of a fifth set of available data transmission windows associated with the transmission of rendered layers to the device. [0013] In some aspects, the techniques described herein relate to a method, where configuring at least one of the render rate or the render resolution of the plurality of layers includes: configuring at least one of the render rate or the render resolution for a rendering unit at the device. [0014] In some aspects, the techniques described herein relate to a method, where outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content includes: outputting, to the rendering unit at the device, the second indicator of at least one of the configured render rate or the configured resolution. [0015] In some aspects, the techniques described herein relate to a method, where obtaining the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device includes: obtaining, from the rendering unit of the device, a third indicator of the second set of content statistics. [0016] In some aspects, the techniques described herein relate to a method, where outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content includes: transmitting, to the device, the second indicator of at least one of the configured render rate or the configured resolution. [0017] In some aspects, the techniques described herein relate to a method, where obtaining the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device includes: receiving, from the device, a third indicator of the second set of content statistics. [0018] In some aspects, the techniques described herein relate to a method, where configuring at least one of the render rate or the render resolution of the plurality of layers includes: configuring at least one of the render rate or the render resolution for a rendering unit at a second device remote from the device. [0019] In some aspects, the techniques described herein relate to a method, where outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content includes: outputting, to the rendering unit at the second device, the second indicator of at least one of the configured render rate or the configured resolution. [0020] In some aspects, the techniques described herein relate to a method, where obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device includes: obtaining, from the rendering unit at the second device, a third indicator of the second set of content statistics. [0021] In some aspects, the techniques described herein relate to a method, where outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content includes: transmitting, to the second device, the second indicator of at least one of the configured render rate or the configured resolution. [0022] In some aspects, the techniques described herein relate to a method, where obtaining the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device includes: receiving, from the second device, a third indicator of the second set of content statistics. [0023] In some aspects, the techniques described herein relate to a method, where obtaining the first indicator of the plurality of layers of graphics content includes: obtaining, from a set of sensors at the device, the first indicator of the plurality of layers of graphics content. [0024] In some aspects, the techniques described herein relate to a method, where obtaining the first indicator of the plurality of layers of graphics content includes: receiving, from the device, the first indicator of the plurality of layers of graphics content. [0025] In some aspects, the techniques described herein relate to a method, where obtaining the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device includes: obtaining, from the reprojection engine of the device, a third indicator of the first set of characteristics. [0026] In some aspects, the techniques described herein relate to a method, where obtaining the first set of characteristics of the reprojection engine of a device and at least one of the second set of content statistics or the third set of wireless characteristics of the device includes: receiving, from the device, a third indicator of the first set of characteristics. [0027] In some aspects, the techniques described herein relate to a method, where obtaining the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device includes: receiving, from a transceiver, a third indicator of the third set of wireless characteristics, where the transceiver is configured to transmit at least a subset of the plurality of layers of graphics content. [0028] To the accomplishment of the foregoing and related ends, the one or more aspects include the features hereinafter fully described and particularly pointed out in the claims. The following description and the annexed drawings set forth in detail certain illustrative features of the one or more aspects. These features are indicative, however, of but a few of the various ways in which the principles of various aspects may be employed, and this description is intended to include all such aspects and their equivalents.
BRIEF DESCRIPTION OF THE DRAWINGS
[0029] FIG. 1 is a block diagram that illustrates an example content generation system in accordance with one or more techniques of this disclosure. [0030] FIG. 2 illustrates an example GPU in accordance with one or more techniques of this disclosure. [0031] FIG. 3 illustrates an example image or surface in accordance with one or more techniques of this disclosure. [0032] FIG. 4 illustrates an example of a reprojection engine configured to process rendered content from a local rendering unit and a remote rendering unit in accordance with one or more techniques of this disclosure. [0033] FIG. 5 illustrates another example of a reprojection engine configured to process rendered content from a local rendering unit and a remote rendering unit in accordance with one or more techniques of this disclosure. [0034] FIG. 6 is a call flow diagram illustrating example communications between a processor, a local rendering unit, and a remote rendering unit in accordance with one or more techniques of this disclosure. [0035] FIG. 7 is a flowchart of an example method of configuring a render rate for a rendered layer by a rendering unit. [0036] FIG. 8 is a flowchart of an example method of configuring a render resolution for a rendered layer by a rendering unit. [0037] FIG. 9 is a flowchart of an example method of configuring a remote compression setting for a transmission of a rendered layer. [0038] FIG. 10 is a flowchart of an example method of graphics processing in accordance with one or more techniques of this disclosure.
DETAILED DESCRIPTION
[0039] Various aspects of systems, apparatuses, computer program products, and methods are described more fully hereinafter with reference to the accompanying drawings. This disclosure may, however, be embodied in many different forms and should not be construed as limited to any specific structure or function presented throughout this disclosure. Rather, these aspects are provided so that this disclosure will be thorough and complete, and will fully convey the scope of this disclosure to those skilled in the art. Based on the teachings herein one skilled in the art should appreciate that the scope of this disclosure is intended to cover any aspect of the systems, apparatuses, computer program products, and methods disclosed herein, whether implemented independently of, or combined with, other aspects of the disclosure. For example, an apparatus may be implemented or a method may be practiced using any number of the aspects set forth herein. In addition, the scope of the disclosure is intended to cover such an apparatus or method which is practiced using other structure, functionality, or structure and functionality in addition to or other than the various aspects of the disclosure set forth herein. Any aspect disclosed herein may be embodied by one or more elements of a claim. [0040] Although various aspects are described herein, many variations and permutations of these aspects fall within the scope of this disclosure. Although some potential benefits and advantages of aspects of this disclosure are mentioned, the scope of this disclosure is not intended to be limited to particular benefits, uses, or objectives. Rather, aspects of this disclosure are intended to be broadly applicable to different wireless technologies, system configurations, processing systems, networks, and transmission protocols, some of which are illustrated by way of example in the figures and in the following description. The detailed description and drawings are merely illustrative of this disclosure rather than limiting, the scope of this disclosure being defined by the appended claims and equivalents thereof. [0041] Several aspects are presented with reference to various apparatus and methods. These apparatus and methods are described in the following detailed description and illustrated in the accompanying drawings by various blocks, components, circuits, processes, algorithms, and the like (collectively referred to as “elements”). These elements may be implemented using electronic hardware, computer software, or any combination thereof. Whether such elements are implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. [0042] By way of example, an element, or any portion of an element, or any combination of elements may be implemented as a “processing system” that includes one or more processors (which may also be referred to as processing units). Examples of processors include microprocessors, microcontrollers, graphics processing units (GPUs), general purpose GPUs (GPGPUs), central processing units (CPUs), application processors, digital signal processors (DSPs), reduced instruction set computing (RISC) processors, systems-on-chip (SOCs), baseband processors, application specific integrated circuits (ASICs), field programmable gate arrays (FPGAs), programmable logic devices (PLDs), state machines, gated logic, discrete hardware circuits, and other suitable hardware configured to perform the various functionality described throughout this disclosure. One or more processors in the processing system may execute software. Software can be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software components, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise. [0043] The term application may refer to software. As described herein, one or more techniques may refer to an application (e.g., software) being configured to perform one or more functions. In such examples, the application may be stored in a memory (e.g., on-chip memory of a processor, system memory, or any other memory). Hardware described herein, such as a processor may be configured to execute the application. For example, the application may be described as including code that, when executed by the hardware, causes the hardware to perform one or more techniques described herein. As an example, the hardware may access the code from a memory and execute the code accessed from the memory to perform one or more techniques described herein. In some examples, components are identified in this disclosure. In such examples, the components may be hardware, software, or a combination thereof. The components may be separate components or sub-components of a single component. [0044] In one or more examples described herein, the functions described may be implemented in hardware, software, or any combination thereof. If implemented in software, the functions may be stored on or encoded as one or more instructions or code on a computer-readable medium. Computer-readable media includes computer storage media. Storage media may be any available media that can be accessed by a computer. By way of example, and not limitation, such computer-readable media can include a random access memory (RAM), a read-only memory (ROM), an electrically erasable programmable ROM (EEPROM), optical disk storage, magnetic disk storage, other magnetic storage devices, combinations of the aforementioned types of computer-readable media, or any other medium that can be used to store computer executable code in the form of instructions or data structures that can be accessed by a computer. [0045] As used herein, instances of the term “content” may refer to “graphical content,” an “image,” etc., regardless of whether the terms are used as an adjective, noun, or other parts of speech. In some examples, the term “graphical content,” as used herein, may refer to a content produced by one or more processes of a graphics processing pipeline. In further examples, the term “graphical content,” as used herein, may refer to a content produced by a processing unit configured to perform graphics processing. In still further examples, as used herein, the term “graphical content” may refer to a content produced by a graphics processing unit. [0046] The following description is directed to examples for the purposes of describing innovative aspects of this disclosure. However, a person having ordinary skill in the art may recognize that the teachings herein may be applied in a multitude of ways. Some or all of the described examples may be implemented in any device or system that is capable of processing graphics commands. Various aspects relate generally to configuration for reprojection systems. Some aspects more specifically relate to configuring attributes of rendered layers by local and remote rendering units. A rendering unit may be a processor configured to render a set of layers, for example a graphics processing unit (GPU). A reprojection engine may be configured to process layers rendered by a set of rendering units. A set of local rendering units may be local to the reprojection engine and a set of remote rendering units may be remote from the reprojection engine. Local rendering units may include rendering units that are co-located with the reprojection engine on a common device (i.e., the set of local rendering units and the reprojection engine are co-located on the same device). Remote rendering units may include rendering units that are not co-located with the reprojection engine on a common device (e.g., the set of remote rendering units are located on a second device remote from the first device having the reprojection engine). A set of remote rendering units may render a set of layers and may wirelessly transmit the set of layers to the device having the reprojection engine for processing. The device having a reprojection engine may provide a first set of layers rendered by the set of remote rendering units and a second set of layers rendered by the set of local rendering units to the same reprojection engine for processing. Various data and information may be used to optimize configuration of the rendered layers by rendering units. For example, a set of characteristics of a reprojection engine may be used to determine the capabilities of the reprojection engine, such as a downscale capability of the reprojection engine (e.g., a maximum downscale ratio), an upscale capability of the reprojection engine (e.g., a maximum upscale ratio), a rotation capability of the reprojection engine (e.g., a maximum rotation), or a warp capability of the reprojection engine (e.g., whether the reprojection is configured to perform a matrix warp or a gridded warp). In another example, a set of content statistics of a rendering unit may be used to determine the capabilities of a rendering unit, such as a timestamp of when a previous layer was rendered by the rendering unit, or a head pose associated with a rendering of a previous layer rendered by the rendering unit. In another example, a set of wireless characteristics association with a transmission of a set of rendered layers may be used to configure a compression setting for a future transmission of a set of rendered layers, such as an amount of data transmitted for the transmission of the set of rendered layers, a set of transmission windows used to transmit the set of rendered layers, a wireless transmission capacity available to transmit a set of rendered layers in the future, or a set of available data transmission windows that may be used to transmit the set of rendered layers in the future. [0047] In some examples, a graphics processor may obtain a first indicator of a plurality of layers of graphics content. The graphics processor may obtain a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device. The graphics processor may configure at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device. The graphics processor may output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. [0048] To reduce power and minimize visual artifacts, a graphics processing system may configure a render rate (i.e., the rate at which a rendering unit renders layers), a render resolution (i.e., the resolution of rendered layers that a rendering unit renders), and/or a remote compression setting for transmission of layers rendered by a set of remote rendering units based on content, known characteristics of the reprojection engine hardware, rendered content statistics, and/or wireless statistics. For example, certain reprojection engines may have different performances when performing downscale, rotation, matrix warp, gridded warp, etc.
[0049] Particular aspects of the subject matter described in this disclosure can be implemented to realize one or more of the following potential advantages. In some examples, by configuring a render rate or a render resolution of a set of layers based on content, known characteristics of the reprojection engine hardware, rendered content statistics, and/or wireless statistics, the described techniques can be used to optimize rendered layers provided to a reprojection engine for processing, thereby reducing the amount of power used and/or the number of visual artifacts produced by the reprojection engine. [0050] The examples describe herein may refer to a use and functionality of a graphics processing unit (GPU). As used herein, a GPU can be any type of graphics processor, and a graphics processor can be any type of processor that is designed or configured to process graphics content. For example, a graphics processor or GPU can be a specialized electronic circuit that is designed for processing graphics content. As an additional example, a graphics processor or GPU can be a general purpose processor that is configured to process graphics content. [0051] FIG. 1 is a block diagram that illustrates an example content generation system 1configured to implement one or more techniques of this disclosure. The content generation system 100 includes a device 104. The device 104 may include one or more components or circuits for performing various functions described herein. In some examples, one or more components of the device 104 may be components of a SOC. The device 104 may include one or more components configured to perform one or more techniques of this disclosure. In the example shown, the device 104 may include a processing unit 120, a content encoder/decoder 122, and a system memory 124. In some aspects, the device 104 may include a number of components (e.g., a communication interface 126, a transceiver 132, a receiver 128, a transmitter 130, a display processor 127, and one or more displays 131). Display(s) 131 may refer to one or more displays 131. For example, the display 131 may include a single display or multiple displays, which may include a first display and a second display. The first display may be a left-eye display and the second display may be a right-eye display. In some examples, the first display and the second display may receive different frames for presentment thereon. In other examples, the first and second display may receive the same frames for presentment thereon. In further examples, the results of the graphics processing may not be displayed on the device, e.g., the first display and the second display may not receive any frames for presentment thereon. Instead, the frames or graphics processing results may be transferred to another device. In some aspects, this may be referred to as split-rendering. [0052] The processing unit 120 may include an internal memory 121. The processing unit 120 may be configured to perform graphics processing using a graphics processing pipeline 107. The content encoder/decoder 122 may include an internal memory 123. In some examples, the device 104 may include a processor, which may be configured to perform one or more display processing techniques on one or more frames generated by the processing unit 120 before the frames are displayed by the one or more displays 131. While the processor in the example content generation system 1is configured as a display processor 127, it should be understood that the display processor 127 is one example of the processor and that other types of processors, controllers, etc., may be used as substitute for the display processor 127. The display processor 127 may be configured to perform display processing. For example, the display processor 127 may be configured to perform one or more display processing techniques on one or more frames generated by the processing unit 120. The one or more displays 131 may be configured to display or otherwise present frames processed by the display processor 127. In some examples, the one or more displays 131 may include one or more of a liquid crystal display (LCD), a plasma display, an organic light emitting diode (OLED) display, a projection display device, an augmented reality display device, a virtual reality display device, a head-mounted display, or any other type of display device. [0053] Memory external to the processing unit 120 and the content encoder/decoder 122, such as system memory 124, may be accessible to the processing unit 120 and the content encoder/decoder 122. For example, the processing unit 120 and the content encoder/decoder 122 may be configured to read from and/or write to external memory, such as the system memory 124. The processing unit 120 may be communicatively coupled to the system memory 124 over a bus. In some examples, the processing unit 120 and the content encoder/decoder 122 may be communicatively coupled to the internal memory 121 over the bus or via a different connection. [0054] The content encoder/decoder 122 may be configured to receive graphical content from any source, such as the system memory 124 and/or the communication interface 126. The system memory 124 may be configured to store received encoded or decoded graphical content. The content encoder/decoder 122 may be configured to receive encoded or decoded graphical content, e.g., from the system memory 124 and/or the communication interface 126, in the form of encoded pixel data. The content encoder/decoder 122 may be configured to encode or decode any graphical content. [0055] The internal memory 121 or the system memory 124 may include one or more volatile or non-volatile memories or storage devices. In some examples, internal memory 1or the system memory 124 may include RAM, static random access memory (SRAM), dynamic random access memory (DRAM), erasable programmable ROM (EPROM), EEPROM, flash memory, a magnetic data media or an optical storage media, or any other type of memory. The internal memory 121 or the system memory 124 may be a non-transitory storage medium according to some examples. The term “non-transitory” may indicate that the storage medium is not embodied in a carrier wave or a propagated signal. However, the term “non-transitory” should not be interpreted to mean that internal memory 121 or the system memory 124 is non-movable or that its contents are static. As one example, the system memory 124 may be removed from the device 104 and moved to another device. As another example, the system memory 124 may not be removable from the device 104. [0056] The processing unit 120 may be a CPU, a GPU, GPGPU, a microprocessor, or any other processing unit that may be configured to perform graphics processing, such as configuration of attributes for a set of rendered layers (e.g., a render rate of a rendering unit, a render resolution of a rendering unit, a remote compression setting for transmission of a set of rendered layers). In some examples, the processing unit 1may be integrated into a motherboard of the device 104. In further examples, the processing unit 120 may be present on a graphics card that is installed in a port of the motherboard of the device 104, or may be otherwise incorporated within a peripheral device configured to interoperate with the device 104. The processing unit 120 may include one or more processors, such as one or more microprocessors, GPUs, ASICs, FPGAs, arithmetic logic units (ALUs), DSPs, discrete logic, software, hardware, firmware, other equivalent integrated or discrete logic circuitry, or any combinations thereof. If the techniques are implemented partially in software, the processing unit 120 may store instructions for the software in a suitable, non-transitory computer-readable storage medium, e.g., internal memory 121, and may execute the instructions in hardware using one or more processors to perform the techniques of this disclosure. Any of the foregoing, including hardware, software, a combination of hardware and software, etc., may be considered to be one or more processors.
[0057] The content encoder/decoder 122 may be any processing unit configured to perform content decoding. In some examples, the content encoder/decoder 122 may be integrated into a motherboard of the device 104. The content encoder/decoder 1may include one or more processors, such as one or more microprocessors, application specific integrated circuits (ASICs), field programmable gate arrays (FPGAs), arithmetic logic units (ALUs), digital signal processors (DSPs), video processors, discrete logic, software, hardware, firmware, other equivalent integrated or discrete logic circuitry, or any combinations thereof. If the techniques are implemented partially in software, the content encoder/decoder 122 may store instructions for the software in a suitable, non-transitory computer-readable storage medium, e.g., internal memory 123, and may execute the instructions in hardware using one or more processors to perform the techniques of this disclosure. Any of the foregoing, including hardware, software, a combination of hardware and software, etc., may be considered to be one or more processors. [0058] In some aspects, the content generation system 100 may include a communication interface 126. The communication interface 126 may include a receiver 128 and a transmitter 130. The receiver 128 may be configured to perform any receiving function described herein with respect to the device 104. Additionally, the receiver 128 may be configured to receive information, e.g., eye or head position information, rendering commands, and/or location information, from another device. The transmitter 130 may be configured to perform any transmitting function described herein with respect to the device 104. For example, the transmitter 130 may be configured to transmit information to another device, which may include a request for content. The receiver 128 and the transmitter 130 may be combined into a transceiver 132. In such examples, the transceiver 132 may be configured to perform any receiving function and/or transmitting function described herein with respect to the device 104. [0059] Referring again to FIG. 1, in certain aspects, the processing unit 120 may include a render configuration engine 198 configured to obtain a first indicator of a plurality of layers of graphics content. The render configuration engine 198 may be configured to obtain a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device. The render configuration engine 198 may be configured to configure at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device. The render configuration engine 198 may be configured to output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. Although the following description may be focused on graphics processing, the concepts described herein may be applicable to other similar processing techniques. [0060] A device, such as the device 104, may refer to any device, apparatus, or system configured to perform one or more techniques described herein. For example, a device may be a server, a base station, a user equipment, a client device, a station, an access point, a computer such as a personal computer, a desktop computer, a laptop computer, a tablet computer, a computer workstation, or a mainframe computer, an end product, an apparatus, a phone, a smart phone, a server, a video game platform or console, a handheld device such as a portable video game device or a personal digital assistant (PDA), a wearable computing device such as a smart watch, an augmented reality device, or a virtual reality device, a non-wearable device, a display or display device, a television, a television set-top box, an intermediate network device, a digital media player, a video streaming device, a content streaming device, an in-vehicle computer, any mobile device, any device configured to generate graphical content, or any device configured to perform one or more techniques described herein. Processes herein may be described as performed by a particular component (e.g., a GPU) but in other embodiments, may be performed using other components (e.g., a CPU) consistent with the disclosed embodiments. [0061] GPUs can process multiple types of data or data packets in a GPU pipeline. For instance, in some aspects, a GPU can process two types of data or data packets, e.g., context register packets and draw call data. A context register packet can be a set of global state information, e.g., information regarding a global register, shading program, or constant data, which can regulate how a graphics context will be processed. For example, context register packets can include information regarding a color format. In some aspects of context register packets, there can be a bit or bits that indicate which workload belongs to a context register. Also, there can be multiple functions or programming running at the same time and/or in parallel. For example, functions or programming can describe a certain operation, e.g., the color mode or color format. Accordingly, a context register can define multiple states of a GPU.
[0062] Context states can be utilized to determine how an individual processing unit functions, e.g., a vertex fetcher (VFD), a vertex shader (VS), a shader processor, or a geometry processor, and/or in what mode the processing unit functions. In order to do so, GPUs can use context registers and programming data. In some aspects, a GPU can generate a workload, e.g., a vertex or pixel workload, in the pipeline based on the context register definition of a mode or state. Certain processing units, e.g., a VFD, can use these states to determine certain functions, e.g., how a vertex is assembled. As these modes or states can change, GPUs may need to change the corresponding context. Additionally, the workload that corresponds to the mode or state may follow the changing mode or state. [0063] FIG. 2 illustrates an example GPU 200 in accordance with one or more techniques of this disclosure. As shown in FIG. 2, GPU 200 includes command processor (CP) 210, draw call packets 212, VFD 220, VS 222, vertex cache (VPC) 224, triangle setup engine (TSE) 226, rasterizer (RAS) 228, Z process engine (ZPE) 230, pixel interpolator (PI) 232, fragment shader (FS) 234, render backend (RB) 236, L2 cache (UCHE) 238, and system memory 240. Although FIG. 2 displays that GPU 2includes processing units 220-238, GPU 200 can include a number of additional processing units. Additionally, processing units 220-238 are merely an example and any combination or order of processing units can be used by GPUs according to the present disclosure. GPU 200 also includes command buffer 250, context register packets 260, and context states 261. [0064] As shown in FIG. 2, a GPU can utilize a CP, e.g., CP 210, or hardware accelerator to parse a command buffer into context register packets, e.g., context register packets 260, and/or draw call data packets, e.g., draw call packets 212. The CP 210 can then send the context register packets 260 or draw call data packets 212 through separate paths to the processing units or blocks in the GPU. Further, the command buffer 2can alternate different states of context registers and draw calls. For example, a command buffer can simultaneously store the following information: context register of context N, draw call(s) of context N, context register of context N+1, and draw call(s) of context N+1. [0065] GPUs can render images in a variety of different ways. In some instances, GPUs can render an image using direct rendering and/or tiled rendering. In tiled rendering GPUs, an image can be divided or separated into different sections or tiles. After the division of the image, each section or tile can be rendered separately. Tiled rendering GPUs can divide computer graphics images into a grid format, such that each portion of the grid, i.e., a tile, is separately rendered. In some aspects of tiled rendering, during a binning pass, an image can be divided into different bins or tiles. In some aspects, during the binning pass, a visibility stream can be constructed where visible primitives or draw calls can be identified. A rendering pass may be performed after the binning pass. In contrast to tiled rendering, direct rendering does not divide the frame into smaller bins or tiles. Rather, in direct rendering, the entire frame is rendered at a single time (i.e., without a binning pass). Additionally, some types of GPUs can allow for both tiled rendering and direct rendering (e.g., flex rendering). [0066] In some aspects, GPUs can apply the drawing or rendering process to different bins or tiles. For instance, a GPU can render to one bin, and perform all the draws for the primitives or pixels in the bin. During the process of rendering to a bin, the render targets can be located in GPU internal memory (GMEM). In some instances, after rendering to one bin, the content of the render targets can be moved to a system memory and the GMEM can be freed for rendering the next bin. Additionally, a GPU can render to another bin, and perform the draws for the primitives or pixels in that bin. Therefore, in some aspects, there might be a small number of bins, e.g., four bins, that cover all of the draws in one surface. Further, GPUs can cycle through all of the draws in one bin, but perform the draws for the draw calls that are visible, i.e., draw calls that include visible geometry. In some aspects, a visibility stream can be generated, e.g., in a binning pass, to determine the visibility information of each primitive in an image or scene. For instance, this visibility stream can identify whether a certain primitive is visible or not. In some aspects, this information can be used to remove primitives that are not visible so that the non-visible primitives are not rendered, e.g., in the rendering pass. Also, at least some of the primitives that are identified as visible can be rendered in the rendering pass. [0067] In some aspects of tiled rendering, there can be multiple processing phases or passes. For instance, the rendering can be performed in two passes, e.g., a binning, a visibility or bin-visibility pass and a rendering or bin-rendering pass. During a visibility pass, a GPU can input a rendering workload, record the positions of the primitives or triangles, and then determine which primitives or triangles fall into which bin or area. In some aspects of a visibility pass, GPUs can also identify or mark the visibility of each primitive or triangle in a visibility stream. During a rendering pass, a GPU can input the visibility stream and process one bin or area at a time. In some aspects, the visibility stream can be analyzed to determine which primitives, or vertices of primitives, are visible or not visible. As such, the primitives, or vertices of primitives, that are visible may be processed. By doing so, GPUs can reduce the unnecessary workload of processing or rendering primitives or triangles that are not visible. [0068] In some aspects, during a visibility pass, certain types of primitive geometry, e.g., position-only geometry, may be processed. Additionally, depending on the position or location of the primitives or triangles, the primitives may be sorted into different bins or areas. In some instances, sorting primitives or triangles into different bins may be performed by determining visibility information for these primitives or triangles. For example, GPUs may determine or write visibility information of each primitive in each bin or area, e.g., in a system memory. This visibility information can be used to determine or generate a visibility stream. In a rendering pass, the primitives in each bin can be rendered separately. In these instances, the visibility stream can be fetched from memory and used to remove primitives which are not visible for that bin. [0069] Some aspects of GPUs or GPU architectures can provide a number of different options for rendering, e.g., software rendering and hardware rendering. In software rendering, a driver or CPU can replicate an entire frame geometry by processing each view one time. Additionally, some different states may be changed depending on the view. As such, in software rendering, the software can replicate the entire workload by changing some states that may be utilized to render for each viewpoint in an image. In certain aspects, as GPUs may be submitting the same workload multiple times for each viewpoint in an image, there may be an increased amount of overhead. In hardware rendering, the hardware or GPU may be responsible for replicating or processing the geometry for each viewpoint in an image. Accordingly, the hardware can manage the replication or processing of the primitives or triangles for each viewpoint in an image. [0070] FIG. 3 illustrates image or surface 300, including multiple primitives divided into multiple bins in accordance with one or more techniques of this disclosure. As shown in FIG. 3, image or surface 300 includes area 302, which includes primitives 321, 322, 323, and 324. The primitives 321, 322, 323, and 324 are divided or placed into different bins, e.g., bins 310, 311, 312, 313, 314, and 315. FIG. 3 illustrates an example of tiled rendering using multiple viewpoints for the primitives 321-324. For instance, primitives 321-324 are in first viewpoint 350 and second viewpoint 351. As such, the GPU processing or rendering the image or surface 300 including area 3can utilize multiple viewpoints or multi-view rendering. [0071] As indicated herein, GPUs or graphics processors can use a tiled rendering architecture to reduce power consumption or save memory bandwidth. As further stated above, this rendering method can divide the scene into multiple bins, as well as include a visibility pass that identifies the triangles that are visible in each bin. Thus, in tiled rendering, a full screen can be divided into multiple bins or tiles. The scene can then be rendered multiple times, e.g., one or more times for each bin. [0072] In aspects of graphics rendering, some graphics applications may render to a single target, i.e., a render target, one or more times. For instance, in graphics rendering, a frame buffer on a system memory may be updated multiple times. The frame buffer can be a portion of memory or random access memory (RAM), e.g., containing a bitmap or storage, to help store display data for a GPU. The frame buffer can also be a memory buffer containing a complete frame of data. Additionally, the frame buffer can be a logic buffer. In some aspects, updating the frame buffer can be performed in bin or tile rendering, where, as discussed above, a surface is divided into multiple bins or tiles and then each bin or tile can be separately rendered. Further, in tiled rendering, the frame buffer can be partitioned into multiple bins or tiles. [0073] As indicated herein, in some aspects, such as in bin or tiled rendering architecture, frame buffers can have data stored or written to them repeatedly, e.g., when rendering from different types of memory. This can be referred to as resolving and unresolving the frame buffer or system memory. For example, when storing or writing to one frame buffer and then switching to another frame buffer, the data or information on the frame buffer can be resolved from the GMEM at the GPU to the system memory, i.e., memory in the double data rate (DDR) RAM or dynamic RAM (DRAM). [0074] In some aspects, the system memory can also be system-on-chip (SoC) memory or another chip-based memory to store data or information, e.g., on a device or smart phone. The system memory can also be physical data storage that is shared by the CPU and/or the GPU. In some aspects, the system memory can be a DRAM chip, e.g., on a device or smart phone. Accordingly, SoC memory can be a chip-based manner in which to store data. [0075] In some aspects, the GMEM can be on-chip memory at the GPU, which can be implemented by static RAM (SRAM). Additionally, GMEM can be stored on a device, e.g., a smart phone. As indicated herein, data or information can be transferred between the system memory or DRAM and the GMEM, e.g., at a device. In some aspects, the system memory or DRAM can be at the CPU or GPU. Additionally, data can be stored at the DDR or DRAM. In some aspects, such as in bin or tiled rendering, a small portion of the memory can be stored at the GPU, e.g., at the GMEM. In some instances, storing data at the GMEM may utilize a larger processing workload and/or consume more power compared to storing data at the frame buffer or system memory. [0076] FIG. 4 is a block diagram 400 illustrating an example of a reprojection engine 4configured to process a set of rendered layers 408. The rendered layers may be rendered by both a local rendering unit 406 and a remote rendering unit 404. The reprojection engine 410 may be co-located on a device 414 with the local rendering unit 406. In other words, both the reprojection engine 410 and the local rendering unit 406 may communicate with one another via a local BUS or a wire with a high throughput. In contrast, the remote rendering unit may be located on a device remote from the device 414, and may communicate with the device via a communication means with a low throughput, such as a wireless connection (e.g., Bluetooth™, WiFi, or a radio access technology (RAT)), as compared with the connection between the local rendering unit 406 and the reprojection engine 410. While the block diagram 400 illustrates one local rendering unit and one remote rendering unit, the reprojection engine 410 may process rendered layers from a plurality of local rendering units and/or a plurality of remote rendering units in other aspects. [0077] The local rendering unit 406 and the remote rendering unit 404 may be hardware devices configured to render layers for the reprojection engine 410. For example, the local rendering unit 406 may be a GPU of the device 414, and the remote rendering unit 404 may be a GPU of a remote device in wireless communication with the device 414. The reprojection engine 410 may be configured to perform reprojection on a set of rendered layers 408 for a display, for example a display of an augmented reality (AR) device or an extended reality (XR) device. The device 414 may include such an AR device, which may augment a real (physical) world with virtual content for a set of tasks. For example, an AR device may augment a real world room with virtual furniture to virtually shop for furniture, may augment a real world table with virtual table top games, may augment a real world path with virtual directions to a destination, may augment real world food with virtual health and nutrition information, may augment a real world setting with virtual screens and displays, or may augment a real world conference room with virtual meeting attendees. An AR device may be wearable over a head of a user, and may be configured to map a physical world about the AR device, localize the AR device in that physical world, and may position and/or render virtual content on a near-eye display visible to a user wearing the AR device. In some aspects, an AR device may utilize hand and/or fingertip tracking sensors to allow users of the AR device to control a user interface of a program running on the AR device in augmented reality. [0078] The degree of freedom (DoF) input 402 may be a set of sensors that provide DoF information of the device 414 to the remote rendering unit 404 and the local rendering unit 406. For example, the DoF input 402 may provide head pose information to the local rendering unit 406, which may then render a set of locally rendered layers 4based on the head pose information. The DoF input 402 may provide the head pose information via a local system bus or a wire of the device 414. Similarly, the DoF input 402 may provide head pose information to the remote rendering unit 404, which may then render a set of remotely rendered layers 405 based on the head pose information. The DoF input 402 may provide the head pose information via a wireless transceiver of the device 414. The remote rendering unit 404 may transmit the set of remotely rendered layers 405 to the device 414, which may receive the set of remotely rendered layers 405 via a wireless transceiver. The set of rendered layers 408 may include the set of remotely rendered layers 405 and the set of locally rendered layers 407. [0079] The reprojection engine 410 may be configured to adjust the set of rendered layers 408. The set of rendered layers 408 may include, for example, low-framerate rendered content for the latest head-pose of the device 414 to minimize motion-to-photon latency. The reprojection engine 410 may increase the framerate of a layer to match a framerate of a set of displays (e.g., a display for a right eye and a display for a left eye) for a user of the device 414. In one example, the set of rendered layers 408 may include a set of layers rendered at 1 FPS to 60 FPS, and the reprojection engine 4may generate a set of reprojected layers 412 having a framerate of 360 Hz per eye. For example, the set of rendered layers 408 may include a first layer of a fast-moving virtual object rendered at a framerate of 60 FPS, a second layer of a virtual video screen playing a video rendered to match a content framerate of 30 FPS, and a third layer of a virtual clock on a wall with a minute hand that is rendered at 1 frame per minute. The reprojection engine 410 may process the set of rendered layers 408 to generate the set of reprojected layers 412 having a framerate of 60 FPS, or 120 FPS.
In some aspects, the reprojected layers 412 may be processed using a composition engine before being displayed to a set of displays. [0080] In other aspects, the set of rendered layers 408 may include, for example, low-resolution rendered content for the latest head-pose of the device 414. The reprojection engine 410 may increase the resolution of a layer to match a resolution of a set of displays for a user of the device 414. In one example, the set of rendered layers 408 may include a set of layers rendered at 1K to 4K, and the reprojection engine 410 may generate a set of reprojected layers 412 having a resolution of 8K per eye. For example, the set of rendered layers 408 may include a first layer of a first virtual object rendered at a resolution of 1K, a second layer of a second virtual object rendered at a resolution of 2K, and a third layer of a third virtual object rendered at a resolution of 4K. The reprojection engine 410 may process the set of rendered layers 408 to generate the set of reprojected layers 412 having a resolution of 4K or 8K. In some aspects, the reprojected layers 412 may be processed using a composition engine before being displayed to a set of displays. [0081] The reprojection engine 410 may be configured to process the set of rendered layers 408 in one or more ways. For example, the reprojection engine 410 may be configured to upscale at least some of the set of rendered layers 408, downscale at least some of the set of rendered layers 408, rotate at least some of the set of rendered layers 408, or warp at least some of the set of rendered layers 408 (e.g., matrix warp, gridded warp). In some aspects, a system may define the capabilities of the reprojection engine 410 based on a set of characteristics of the reprojection engine 410. For example, a system may define a table that indicates the performance of the reprojection engine 410 for a specific reprojection type. Such a table may include indicators of a downscale capability of the reprojection engine 410 (e.g., a maximum downscale ratio), an upscale capability of the reprojection engine 410 (e.g., a maximum upscale ratio), a rotation capability of the reprojection engine 410 (e.g., a maximum rotation), or a warp capability of the reprojection engine 410 (e.g., whether the reprojection is configured to perform a matrix warp or a gridded warp, maximum number of pixels that can be warped in a given time). [0082] The render configuration engine 198 may be configured to configure a set of render attributes, for example a render rate, a render resolution, or a remote compression setting, to optimize performance based on the limits of the reprojection system, thereby reducing power consumed by the reprojection engine 410 due to re-rendering without introducing visual artifacts. For example, the render configuration engine 1may be configured to obtain a first indicator of a plurality of layers of graphics content (e.g., render the reprojected layers 412 to have a specified framerate and a specified resolution). The render configuration engine 198 may be configured to obtain a first set of characteristics of the reprojection engine 410 of the device 414 and at least one of a second set of content statistics (e.g., time stamp and head pose of previously rendered layers by the remote rendering unit 404 and/or time stamp and head pose of previously rendered layers by the local rendering unit 406) or a third set of wireless characteristics of the device 414 (e.g., current data usage, available wireless capacity, available data transmission windows). The render configuration engine 198 may be configured to configure at least one of a render rate of the set of remotely rendered layers 405 by the remote rendering unit 404, a render resolution of the set of remotely rendered layers 405 by the remote rendering unit 404, or a remote compression setting of the set of remotely rendered layers 405 before transmission based on at least one of the first set of characteristics of the reprojection engine 410 of the device 414, the set of content statistics of the remote rendering unit 404, or the set of wireless characteristics of the device 414. The render configuration engine 198 may be configured to configure at least one of a render rate of the set of locally rendered layers 407 by the local rendering unit 406, or a render resolution of the set of locally rendered layers 407 by the local rendering unit 406, based on at least one of the first set of characteristics of the reprojection engine 410 of the device 414, the set of content statistics of the local rendering unit 406, or the set of wireless characteristics of the device 414. The render configuration engine 198 may be configured to output a second indicator of at least one of the configured render rate for the remote rendering unit 404, the configured render resolution for the remote rendering unit 404, the configured remote compression setting for the transmission of the remotely rendered layers 405, the configured render rate for the local rendering unit 406, and/or the configured render resolution for the local rendering unit 406. [0083] FIG. 5 is a block diagram 500 illustrating an example of a reprojection engine 5configured to process a set of rendered layers 508. The reprojection engine 510 may be a fixed hardware reprojection engine. In some aspects, the reprojection engine 5may be a flexible software reprojection process. The rendered layers may be rendered by both a local rendering unit 506 and a remote rendering unit 504. The reprojection engine 510 may be co-located on a device 514 with the local rendering unit 506. In other words, both the reprojection engine 510 and the local rendering unit 506 may communicate with one another via a local BUS or a wire with a high throughput. In contrast, the remote rendering unit may be located on a device remote from the device 514, and may communicate with the device via a communication means with a low throughput, such as a wireless connection via the wireless device 516 and/or the wireless device 518, as compared with the connection between the local rendering unit 506 and the reprojection engine 510. While the block diagram 500 illustrates one local rendering unit and one remote rendering unit, the reprojection engine 510 may process rendered layers from a plurality of local rendering units and/or a plurality of remote rendering units in other aspects. While the block diagram 500 illustrates the remote rendering unit 504 receiving DoF input 502 via the wireless device 516 and transmitting the set of remotely rendered layers 505 via the wireless device 518, the remote rendering unit 504 may receive DoF input 502 and may transmit the set of remotely rendered layers 505 via the same wireless device (e.g., a transceiver of the remote rendering unit 504, a wireless base station in communication with the device 514). The wireless device 516 and/or the wireless device 518 may be a wireless tethered device, a wireless connected device, a universal serial bus (USB)-tethered device, or a peripheral component interconnect express (PCIe)-tethered device. [0084] The local rendering unit 506 and the remote rendering unit 504 may be hardware devices configured to render layers for the reprojection engine 510. For example, the local rendering unit 506 may be a GPU of the device 514, and the remote rendering unit 504 may be a GPU of a remote device in wireless communication with the device 514. The reprojection engine 510 may be configured to perform reprojection on a set of rendered layers 508 for a display, for example a display of an AR device or an XR device. The device 514 may include such an AR device, which may augment a real (physical) world with virtual content for a set of tasks. For example, an AR device may augment a real world room with virtual furniture to virtually shop for furniture, may augment a real world table with virtual table top games, may augment a real world path with virtual directions to a destination, may augment real world food with virtual health and nutrition information, may augment a real world setting with virtual screens and displays, or may augment a real world conference room with virtual meeting attendees. An AR device may be wearable over a head of a user, and may be configured to map a physical world about the AR device, localize the AR device in that physical world, and may position and/or render virtual content on a near-eye display visible to a user wearing the AR device. In some aspects, an AR device may utilize hand and/or fingertip tracking sensors to allow users of the AR device to control a user interface of a program running on the AR device in augmented reality. [0085] The DoF input 502 may be a set of sensors that provide DoF information of the device 514 to the remote rendering unit 504 and the local rendering unit 506. For example, the DoF input 502 may provide head pose information to the local rendering unit 506, which may then render a set of locally rendered layers 507 based on the head pose information. The DoF input 502 may provide the head pose information via a local system bus or a wire of the device 514. Similarly, the DoF input 502 may provide head pose information to the remote rendering unit 504 via the wireless device 516, which may then render a set of remotely rendered layers 505 based on the head pose information. The DoF input 502 may provide the head pose information via a wireless transceiver of the device 514. The remote rendering unit 504 may transmit the set of remotely rendered layers 505 to the device 514 via the wireless device 518. The device 514 may receive the set of remotely rendered layers 505 via a wireless transceiver that receives wireless signals from the wireless device 518. The set of rendered layers 5may include the set of remotely rendered layers 505 and the set of locally rendered layers 507. The wireless device 518 may collect wireless characteristics, such as current data usage for transmissions of the set of remotely rendered layers 505 to the device 514, available wireless capacity for transmitting the set of remotely rendered layers 505 to the device 514, and/or available data transmission windows for transmitting the set of remotely rendered layers 505 to the device 514. In some aspects, the current data usage may be indicated as a function of channel conditions, for example congestion. An indicator of current data usage may be used to calculate a projected performance of receiving a remotely rendered layer for use with the reprojection engine 510. The render configuration engine 198 may collect the wireless characteristics from the wireless device 518. [0086] The reprojection engine 510 may be configured to adjust the set of rendered layers 508. The set of rendered layers 508 may include, for example, low-framerate rendered content for the latest head-pose of the device 514 to minimize motion-to-photon latency. The reprojection engine 510 may increase the framerate of a layer to match a framerate of a set of displays (e.g., a display for a right eye and a display for a left eye) for a user of the device 514. In one example, the set of rendered layers 508 may include a set of layers rendered at 1 FPS to 60 FPS, and the reprojection engine 510 may generate a set of reprojected layers 512 having a framerate of 360 Hz per eye. For example, the set of rendered layers 508 may include a first layer of a fast-moving virtual object rendered at a framerate of 60 FPS, a second layer of a virtual video screen playing a video rendered to match a content framerate of 30 FPS, and a third layer of a virtual clock on a wall with a minute hand that is rendered at 1 frame per minute. The reprojection engine 510 may process the set of rendered layers 508 to generate the set of reprojected layers 512 having a framerate of 60 FPS, or 120 FPS. In some aspects, the reprojected layers 512 may be processed using a composition engine before being displayed to a set of displays. [0087] In other aspects, the set of rendered layers 508 may include, for example, low-resolution rendered content for the latest head-pose of the device 514. The reprojection engine 510 may increase the resolution of a layer to match a resolution of a set of displays for a user of the device 514. In one example, the set of rendered layers 508 may include a set of layers rendered at 1K to 5K, and the reprojection engine 510 may generate a set of reprojected layers 512 having a resolution of 8K per eye. For example, the set of rendered layers 508 may include a first layer of a first virtual object rendered at a resolution of 1K, a second layer of a second virtual object rendered at a resolution of 2K, and a third layer of a third virtual object rendered at a resolution of 5K. The reprojection engine 510 may process the set of rendered layers 508 to generate the set of reprojected layers 512 having a resolution of 5K or 8K. In some aspects, the reprojected layers 512 may be processed using a composition engine before being displayed to a set of displays. [0088] The reprojection engine 510 may be configured to process the set of rendered layers 508 in one or more ways. For example, the reprojection engine 510 may be configured to upscale at least some of the set of rendered layers 508, downscale at least some of the set of rendered layers 508, rotate at least some of the set of rendered layers 508, or warp at least some of the set of rendered layers 508 (e.g., matrix warp, gridded warp). In some aspects, a system may define the capabilities of the reprojection engine 510 based on a set of characteristics of the reprojection engine 510. For example, a system may define a table that indicates the performance of the reprojection engine 510 for a specific reprojection type. Such a table may include indicators of a downscale capability of the reprojection engine 510 (e.g., a maximum downscale ratio), an upscale capability of the reprojection engine 510 (e.g., a maximum upscale ratio), a rotation capability of the reprojection engine 510 (e.g., a maximum rotation), or a warp capability of the reprojection engine 510 (e.g., whether the reprojection is configured to perform a matrix warp or a gridded warp, maximum number of pixels that can be warped in a given time). [0089] The render configuration engine 198 may be configured to configure a set of render attributes, for example a render rate, a render resolution, or a remote compression setting, to optimize performance based on the limits of the reprojection system, thereby reducing power consumed by the reprojection engine 510 due to re-rendering without introducing visual artifacts. For example, the render configuration engine 1may be configured to obtain a first indicator of a plurality of layers of graphics content (e.g., render the reprojected layers 512 to have a specified framerate and a specified resolution). The render configuration engine 198 may be configured to obtain a first set of characteristics of the reprojection engine 510 of the device 514 and at least one of a second set of content statistics (e.g., time stamp and head pose of previously rendered layers by the remote rendering unit 504 and/or time stamp and head pose of previously rendered layers by the local rendering unit 506) or a third set of wireless characteristics of the device 514 (e.g., current data usage, available wireless capacity, available data transmission windows). The render configuration engine 198 may be configured to configure at least one of a render rate of the set of remotely rendered layers 505 by the remote rendering unit 504, a render resolution of the set of remotely rendered layers 505 by the remote rendering unit 504, or a remote compression setting of the set of remotely rendered layers 505 before transmission based on at least one of the first set of characteristics of the reprojection engine 510 of the device 514, the set of content statistics of the remote rendering unit 504, or the set of wireless characteristics of the device 514. The render configuration engine 198 may be configured to configure at least one of a render rate of the set of locally rendered layers 507 by the local rendering unit 506, or a render resolution of the set of locally rendered layers 507 by the local rendering unit 506, based on at least one of the first set of characteristics of the reprojection engine 510 of the device 514, the set of content statistics of the local rendering unit 506, or the set of wireless characteristics of the device 514. The render configuration engine 198 may be configured to output a second indicator of at least one of the configured render rate for the remote rendering unit 504, the configured render resolution for the remote rendering unit 504, the configured remote compression setting for the transmission of the remotely rendered layers 505, the configured render rate for the local rendering unit 506, and/or the configured render resolution for the local rendering unit 506. [0090] FIG. 6 is a call flow diagram 600 illustrating example communications between a processor 604 (e.g., a CPU, a microprocessor), a local rendering unit 602, and a remote rendering unit 606 in accordance with one or more techniques of this disclosure. While the call flow diagram 600 illustrates a single local rendering unit and a single remote rendering unit 606, the local rendering unit 602 may represent a set of local rendering units (e.g., a set of GPUs on a device) and/or the remote rendering unit 606 may represent a set of remote rendering units (e.g., a set of remote GPUs configured to assist a device in rendering layers). [0091] At 608, the processor 604 may obtain characteristics of a reprojection engine. For example, the processor 604 may retrieve a table that indicates the performance of a reprojection engine for a specific reprojection type. Such a table may include indicators of a downscale capability of the reprojection engine (e.g., a maximum downscale ratio), an upscale capability of the reprojection engine (e.g., a maximum upscale ratio), a rotation capability of the reprojection engine (e.g., a maximum rotation), or a warp capability of the reprojection engine (e.g., whether the reprojection is configured to perform a matrix warp or a gridded warp, maximum number of pixels that can be warped in a given time). [0092] At 610, the processor 604 may obtain wireless characteristics for wirelessly transmitting a set of remotely rendered layers from the remote rendering unit 606 to a device having the reprojection engine. The wireless characteristics may include, for example current data usage, available wireless capacity, and/or available data transmission windows. [0093] At 612, the local rendering unit 602 may collect content statistics. For example, the local rendering unit 602 may save a log, or table, of timestamps and head poses for each rendered layer over time. The local rendering unit 602 may output the set of content statistics 614 to the processor 604. In some aspects, the processor 604 may operate on the same device as the local rendering unit 602, and may receive the set of content statistics 614 via a local transmission pathway (e.g., a system bus, a wire). In some aspects, the processor 604 may operate on a different device as the local rendering unit 602, and may receive the set of content statistics 614 via a wireless transceiver.
[0094] At 616, the remote rendering unit 606 may collect content statistics. For example, the remote rendering unit 606 may save a log, or table, of timestamps and head poses for each rendered layer over time. The remote rendering unit 606 may output the set of content statistics 618 to the processor 604. In some aspects, the processor 604 may operate on the same device as the remote rendering unit 606, and may receive the set of content statistics 618 via a local transmission pathway (e.g., a system bus, a wire). In some aspects, the processor 604 may operate on a different device as the remote rendering unit 606, and may receive the set of content statistics 618 via a wireless transceiver. [0095] At 620, the processor 604 may obtain graphics content for a set of layers to be processed by the reprojection engine. For example, the graphics content may define a framerate and a resolution for the set of reprojected layers. At 622, the processor 6may configure a set of rendering attributes to ensure that the rendered GPU content does not violate the limitations of the system. For example, a reprojection engine may have a maximum rotation of 45 degrees, and the graphics content may include a rotation of greater than 45 degrees. The processor 604 may configure a set of rendering attributes to rotate a rendered layer such that the reprojection engine does not rotate the rendered layer by more than 45 degrees. In another example, a wireless device used to transmit sets of rendered layers may have a max data throughput of megabits per second (mbps) with 5 mbps currently in use. The processor 604 may configure a set of remote compression settings to ensure that the transmission of the sets of rendered layers does not cause the data throughput of the wireless device to exceed 10 mbps. Without knowledge of such limitations, the processor 604 may violate the limitations, resulting in performance degradation, power impact, or visual artifacts when generating reprojected layers using a reprojection engine. [0096] The processor 604 may output a render configuration 624 to the local rendering unit 602 based on the configuration at 622. The render configuration 624 may include a configuration for a render rate for the local rendering unit 602 to render a set of rendered layers. The render configuration 624 may include a configuration for a resolution for the local rendering unit 602 to use to render a set of rendered layers. [0097] The processor 604 may output a render configuration 626 to the local rendering unit 602 based on the configuration at 622. The render configuration 626 may include a configuration for a render rate for the remote rendering unit 606 to render a set of rendered layers. The render configuration 626 may include a configuration for a resolution for the remote rendering unit 606 to use to render a set of rendered layers. The render configuration 626 may include a configuration for a remote compression setting to use to transmit the set of rendered layers generated by the remote rendering unit 606 to the device with the reprojection engine. [0098] FIG. 7 is a flowchart 700 of an example method of configuring a render rate for a rendered layer by a rendering unit in accordance with one or more techniques of this disclosure. The method may be performed by an apparatus, such as an apparatus for graphics processing, a CPU, a microprocessor, a GPU, a GPU driver, a wireless communication device, and the like, as used in connection with the aspects of FIGs. 1-6. Moreover, the method may be performed by the render configuration engine 1in FIGs. 1, 4, or 5. [0099] At 704, a processor may determine whether the framerate of the content triggers a re-render of a set of layers. If the framerate of the content triggers a re-render of a set of layers, then at 708, the processor may configure re-rendering a layer with the latest head pose. If the framerate of the content does not trigger a re-render of a set of layers, at 706 the processor may determine whether a reprojection engine is sufficient to render the content based on the previously rendered layer. If the reprojection engine is sufficient to render the content based on the previously rendered layer, then at 702, the processor does not re-render the layer, and the reprojection engine uses the previously rendered layer to generate reprojected layers. If the reprojection engine is not sufficient to render the content based on the previously rendered layer, then at 708, the processor may configure re-rendering a layer with the latest head pose. [0100] FIG. 8 is a flowchart 800 of an example method of configuring a render resolution for a rendered layer by a rendering unit in accordance with one or more techniques of this disclosure. The method may be performed by an apparatus, such as an apparatus for graphics processing, a CPU, a microprocessor, a GPU, a GPU driver, a wireless communication device, and the like, as used in connection with the aspects of FIGs. 1-6. Moreover, the method may be performed by the render configuration engine 1in FIGs. 1, 4, or 5. [0101] At 804, a processor may determine whether the resolution of the content triggers a re-render of a set of layers. If the resolution of the content triggers a re-render of a set of layers, then at 808, the processor may configure re-rendering a layer with the latest head pose based on the resolution. If the resolution of the content does not trigger a re-render of a set of layers, at 806 the processor may determine whether a reprojection engine is sufficient to render the content at the resolution based on the previously rendered layer. If the reprojection engine is sufficient to render the content based on the previously rendered layer, then at 802, the processor does not re-render the layer, and the reprojection engine uses the previously rendered layer to generate reprojected layers at the given resolution. If the reprojection engine is not sufficient to render the content based on the previously rendered layer, then at 808, the processor may configure re-rendering a layer at the given resolution. [0102] FIG. 9 is a flowchart 900 of an example method of configuring a remote compression setting for a transmission of a rendered layer in accordance with one or more techniques of this disclosure. The method may be performed by an apparatus, such as an apparatus for graphics processing, a CPU, a microprocessor, a GPU, a GPU driver, a wireless communication device, and the like, as used in connection with the aspects of FIGs. 1-6. Moreover, the method may be performed by the render configuration engine 198 in FIGs. 1, 4, or 5. [0103] At 904, a processor may determine whether the wireless characteristics for receiving a set of remotely rendered layers has been updated (e.g., bandwidth has changed, available time windows have changed). If the wireless characteristics has been updated, then at 908, the processor may configure remote compression settings based on the updated wireless characteristics. If the wireless characteristics have not been updated, at 906 the processor may determine whether a reprojection engine is sufficient to render the content based on the previously rendered layer and previously used compression settings. If the reprojection engine is sufficient to render the content based on the previously used compression settings, then at 902, the processor does not configure compression settings, and the wireless transceivers/devices use the previously configured compression settings. If the reprojection engine is not sufficient to render the content based on the previously used compression settings, then at 908, the processor may configure remote compression settings based on the updated wireless characteristics. [0104] FIG. 10 is a flowchart 1000 of an example method of graphics processing in accordance with one or more techniques of this disclosure. The method may be performed by an apparatus, such as an apparatus for graphics processing, a CPU, a microprocessor, a GPU, a wireless communication device, and the like, as used in connection with the aspects of FIGs. 1-6.
[0105] At 1002, the apparatus may obtain a first indicator of a plurality of layers of graphics content. For example, 1002 may be performed by the processor 604 in FIG. 6, which may, at 620, obtain graphics content for a set of layers to be processed by a reprojection engine. The processor 604 may obtain that graphics content by obtaining an indicator of the set of layers of the graphics content to be rendered. The indicator may indicate for the set of layers to be rendered to be composed into a frame. Moreover, 1002 may be performed by the render configuration engine 198 in FIGs. 1, 4, and 5. [0106] At 1004, the apparatus may obtain a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device. For example, 1002 may be performed by the processor 604 in FIG. 6, which may obtain a first set of characteristics of a reprojection engine of a device (e.g., downscale/upscale capability, rotation capability, warp capability). The processor 604 may obtain the set of content statistics 614 from the local rendering unit 602 and/or may receive the set of content statistics 618 from the remote rendering unit 606. At 610, the processor 604 may obtain wireless characteristics of the device, for example from a wireless device. Moreover, 1004 may be performed by the render configuration engine 198 in FIGs. 1, 4, and 5. [0107] At 1006, the apparatus may configure at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device. For example, 10may be performed by the processor 604 in FIG. 6, which may, at 622, configure at least one of a render rate (e.g., rendering framerate at a local and/or remote GPU) or a render resolution (e.g., resolution of rendered layer at a local and/or remote GPU) of at least one of the plurality of layers based on the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device. Moreover, 1006 may be performed by the render configuration engine 198 in FIGs. 1, 4, and 5. [0108] At 1008, the apparatus may output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. For example, 1002 may be performed by the processor 604 in FIG. 6, which may output the render configuration 624 to the local rendering unit 602 and/or the render configuration 626 to the remote rendering unit 606. Moreover, 1008 may be performed by the render configuration engine 198 in FIGs. 1, 4, and 5. [0109] In configurations, a method or an apparatus for graphics processing is provided. The apparatus may be a GPU, a CPU, or some other processor that may perform graphics processing. In aspects, the apparatus may be the processing unit 120 within the device 104, or may be some other hardware within the device 104 or another device. The apparatus may include means for obtaining a first indicator of a plurality of layers of graphics content. The apparatus may include means for obtaining a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics, or a third set of wireless characteristics of the device. The apparatus may include means for configuring at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device. The apparatus may include means for outputting a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. The apparatus may include means for configuring at least one of the render rate or the render resolution by configuring a remote compression setting. The apparatus may include means for outputting a third indication of the configured remote compression setting. The apparatus may include means for outputting the third indication of the configured remote compression setting by transmitting, to a second device remote from the device, the third indication of the configured remote compression setting. The apparatus may include means for outputting the third indication of the configured remote compression setting by outputting, to a compression unit of the device, the third indication of the configured remote compression setting. The first set of characteristics may include at least one of: a third indicator of a downscale capability for the reprojection engine; a fourth indicator of an upscale capability for the reprojection engine; a fifth indicator of a rotation capability for the reprojection engine; or a sixth indicator of a warp capability for the reprojection engine. The second set of content statistics may include at least one of: a third indicator of a timestamp associated with a layer rendered by a rendering unit; or a fourth indicator of a head pose associated with the layer rendered by the rendering unit. The third set of wireless characteristics may include at least one of: a third indicator of an amount of data wirelessly transmitted to the device; a fourth indicator of a fourth set of transmission windows used to transmit wirelessly transmit the amount of data to the device; a fifth indicator of a wireless transmission capacity associated with transmission of rendered layers to the device; or a sixth indicator of a fifth set of available data transmission windows associated with the transmission of rendered layers to the device. The apparatus may include means for configuring at least one of the render rate or the render resolution of the plurality of layers by configuring at least one of the render rate or the render resolution for a rendering unit at the device. The apparatus may include means for outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content by outputting, to the rendering unit at the device, the second indicator of at least one of the configured render rate or the configured resolution. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device by obtaining, from the rendering unit of the device, a third indicator of the second set of content statistics. The apparatus may include means for outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content by transmitting, to the device, the second indicator of at least one of the configured render rate or the configured resolution. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device by receiving, from the device, a third indicator of the second set of content statistics. The apparatus may include means for configuring at least one of the render rate or the render resolution of the plurality of layers by configuring at least one of the render rate or the render resolution for a rendering unit at a second device remote from the device. The apparatus may include means for outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content by outputting, to the rendering unit at the second device, the second indicator of at least one of the configured render rate or the configured resolution. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device by obtaining, from the rendering unit at the second device, a third indicator of the second set of content statistics. The apparatus may include means for outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content by transmitting, to the second device, the second indicator of at least one of the configured render rate or the configured resolution. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device by receiving, from the second device, a third indicator of the second set of content statistics. The apparatus may include means for obtaining the first indicator of the plurality of layers of graphics content by obtaining, from a set of sensors at the device, the first indicator of the plurality of layers of graphics content. The apparatus may include means for obtaining the first indicator of the plurality of layers of graphics content by receiving, from the device, the first indicator of the plurality of layers of graphics content. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device by obtaining, from the reprojection engine of the device, a third indicator of the first set of characteristics. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of a device, the second set of content statistics, or the third set of wireless characteristics of the device by receiving, from the device, a third indicator of the first set of characteristics. The apparatus may include means for obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device by receiving, from a transceiver, a third indicator of the third set of wireless characteristics. The transceiver may be configured to transmit at least a subset of the plurality of layers of graphics content. [0110] It is understood that the specific order or hierarchy of blocks/steps in the processes, flowcharts, and/or call flow diagrams disclosed herein is an illustration of example approaches. Based upon design preferences, it is understood that the specific order or hierarchy of the blocks/steps in the processes, flowcharts, and/or call flow diagrams may be rearranged. Further, some blocks/steps may be combined and/or omitted. Other blocks/steps may also be added. The accompanying method claims present elements of the various blocks/steps in a sample order, and are not meant to be limited to the specific order or hierarchy presented.
[0111] The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects. Thus, the claims are not intended to be limited to the aspects shown herein, but is to be accorded the full scope consistent with the language of the claims, where reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any aspect described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects. [0112] Unless specifically stated otherwise, the term “some” refers to one or more and the term “or” may be interpreted as “and/or” where context does not dictate otherwise. Combinations such as “at least one of A, B, or C,” “one or more of A, B, or C,” “at least one of A, B, and C,” “one or more of A, B, and C,” and “A, B, C, or any combination thereof” include any combination of A, B, and/or C, and may include multiples of A, multiples of B, or multiples of C. Specifically, combinations such as “at least one of A, B, or C,” “one or more of A, B, or C,” “at least one of A, B, and C,” “one or more of A, B, and C,” and “A, B, C, or any combination thereof” may be A only, B only, C only, A and B, A and C, B and C, or A and B and C, where any such combinations may contain one or more member or members of A, B, or C. All structural and functional equivalents to the elements of the various aspects described throughout this disclosure that are known or later come to be known to those of ordinary skill in the art are expressly incorporated herein by reference and are intended to be encompassed by the claims. Moreover, nothing disclosed herein is intended to be dedicated to the public regardless of whether such disclosure is explicitly recited in the claims. The words “module,” “mechanism,” “element,” “device,” and the like may not be a substitute for the word “means.” As such, no claim element is to be construed as a means plus function unless the element is expressly recited using the phrase “means for.” Unless stated otherwise, the phrase “a processor” may refer to “any of one or more processors” (e.g., one processor of one or more processors, a number (greater than one) of processors in the one or more processors, or all of the one or more processors) and the phrase “a memory” may refer to “any of one or more memories” (e.g., one memory of one or more memories, a number (greater than one) of memories in the one or more memories, or all of the one or more memories). [0113] In one or more examples, the functions described herein may be implemented in hardware, software, firmware, or any combination thereof. For example, although the term “processing unit” has been used throughout this disclosure, such processing units may be implemented in hardware, software, firmware, or any combination thereof. If any function, processing unit, technique described herein, or other module is implemented in software, the function, processing unit, technique described herein, or other module may be stored on or transmitted over as one or more instructions or code on a computer-readable medium. [0114] Computer-readable media may include computer data storage media or communication media including any medium that facilitates transfer of a computer program from one place to another. In this manner, computer-readable media generally may correspond to: (1) tangible computer-readable storage media, which is non-transitory; or (2) a communication medium such as a signal or carrier wave. Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code, and/or data structures for implementation of the techniques described in this disclosure. By way of example, and not limitation, such computer-readable media may include RAM, ROM, EEPROM, compact disc-read only memory (CD-ROM), or other optical disk storage, magnetic disk storage, or other magnetic storage devices. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk, and Blu-ray disc, where disks usually reproduce data magnetically, while discs usually reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media. A computer program product may include a computer-readable medium. [0115] The techniques of this disclosure may be implemented in a wide variety of devices or apparatuses, including a wireless handset, an integrated circuit (IC) or a set of ICs, e.g., a chip set. Various components, modules or units are described in this disclosure to emphasize functional aspects of devices configured to perform the disclosed techniques, but do not necessarily need realization by different hardware units. Rather, as described above, various units may be combined in any hardware unit or provided by a collection of inter-operative hardware units, including one or more processors as described above, in conjunction with suitable software and/or firmware.
Accordingly, the term “processor,” as used herein may refer to any of the foregoing structure or any other structure suitable for implementation of the techniques described herein. Also, the techniques may be fully implemented in one or more circuits or logic elements. [0116] The following aspects are illustrative only and may be combined with other aspects or teachings described herein, without limitation. [0117] In some aspects, the techniques described herein relate to a method of graphics processing, including: obtaining a first indicator of a plurality of layers of graphics content; obtaining a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device; configuring at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device; and outputting a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content. [0118] In some aspects, the techniques described herein relate to a method, where configuring at least one of the render rate or the render resolution includes configuring a remote compression setting, where the method further includes outputting a third indication of the configured remote compression setting. [0119] In some aspects, the techniques described herein relate to a method, where outputting the third indication of the configured remote compression setting includes: transmitting, to a second device remote from the device, the third indication of the configured remote compression setting. [0120] Aspect 1 is a method of graphics processing, comprising: obtaining a first indicator of a plurality of layers of graphics content; obtaining at least one of a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device; configuring at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device; and outputting a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content.
[0121] Aspect 2 is the method of aspect 1, wherein configuring at least one of the render rate or the render resolution comprises configuring a remote compression setting, the method further comprising: outputting a third indication of the configured remote compression setting. [0122] Aspect 3 is the method of aspect 2, wherein outputting the third indication of the configured remote compression setting comprises transmitting, to a second device remote from the device, the third indication of the configured remote compression setting. [0123] Aspect 4 is the method of either of aspects 2 or 3, wherein outputting the third indication of the configured remote compression setting comprises outputting, to a compression unit of the device, the third indication of the configured remote compression setting. [0124] Aspect 5 is the method of any of aspects 1 to 4, wherein the first set of characteristics comprises at least one of: a third indicator of a downscale capability for the reprojection engine; a fourth indicator of an upscale capability for the reprojection engine; a fifth indicator of a rotation capability for the reprojection engine; or a sixth indicator of a warp capability for the reprojection engine. [0125] Aspect 6 is the method of any of aspects 1 to 5, wherein the second set of content statistics comprises at least one of: a third indicator of a timestamp associated with a layer rendered by a rendering unit; or a fourth indicator of a head pose associated with the layer rendered by the rendering unit. [0126] Aspect 7 is the method of any of aspects 1 to 6, wherein the third set of wireless characteristics comprises at least one of: a third indicator of an amount of data wirelessly transmitted to the device; a fourth indicator of a fourth set of transmission windows used to transmit wirelessly transmit the amount of data to the device; a fifth indicator of a wireless transmission capacity associated with transmission of rendered layers to the device; or a sixth indicator of a fifth set of available data transmission windows associated with the transmission of rendered layers to the device. [0127] Aspect 8 is the method of any of aspects 1 to 7, wherein configuring at least one of the render rate or the render resolution of the plurality of layers comprises configuring at least one of the render rate or the render resolution for a rendering unit at the device. [0128] Aspect 9 is the method of aspect 8, wherein outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content comprises outputting, to the rendering unit at the device, the second indicator of at least one of the configured render rate or the configured resolution. [0129] Aspect 10 is the method of either of aspects 8 or 9, wherein obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device comprises obtaining, from the rendering unit of the device, a third indicator of the second set of content statistics. [0130] Aspect 11 is the method of any of aspects 8 to 10, wherein outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content comprises transmitting, to the device, the second indicator of at least one of the configured render rate or the configured resolution. [0131] Aspect 12 is the method of any of aspects 8 to 11, wherein obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device comprises receiving, from the device, a third indicator of the second set of content statistics. [0132] Aspect 13 is the method of any of aspects 1 to 12, wherein configuring at least one of the render rate or the render resolution of the plurality of layers comprises configuring at least one of the render rate or the render resolution for a rendering unit at a second device remote from the device. [0133] Aspect 14 is the method of aspect 13, wherein outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content comprises outputting, to the rendering unit at the second device, the second indicator of at least one of the configured render rate or the configured resolution. [0134] Aspect 15 is the method of either of aspects 13 or 14, wherein obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device comprises obtaining, from the rendering unit at the second device, a third indicator of the second set of content statistics. [0135] Aspect 16 is the method of any of aspects 13 to 15, wherein outputting the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of graphics content comprises transmitting, to the second device, the second indicator of at least one of the configured render rate or the configured resolution. [0136] Aspect 17 is the method of any of aspects 13 to 16, wherein obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device comprises receiving, from the second device, a third indicator of the second set of content statistics. [0137] Aspect 18 is the method of any of aspects 1 to 17, wherein obtaining the first indicator of the plurality of layers of graphics content comprises obtaining, from a set of sensors at the device, the first indicator of the plurality of layers of graphics content. [0138] Aspect 19 is the method of any of aspects 1 to 18, wherein obtaining the first indicator of the plurality of layers of graphics content comprises receiving, from the device, the first indicator of the plurality of layers of graphics content. [0139] Aspect 20 is the method of any of aspects 1 to 19, wherein obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device comprises obtaining, from the reprojection engine of the device, a third indicator of the first set of characteristics. [0140] Aspect 21 is the method of any of aspects 1 to 20, wherein obtaining at least one of the first set of characteristics of the reprojection engine of a device, the second set of content statistics, or the third set of wireless characteristics of the device comprises receiving, from the device, a third indicator of the first set of characteristics. [0141] Aspect 22 is the method of any of aspects 1 to 21, further comprising a transceiver coupled to the processor, wherein obtaining at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device comprises receiving, from a transceiver, a third indicator of the third set of wireless characteristics, wherein the transceiver is configured to transmit at least a subset of the plurality of layers of graphics content. [0142] Aspect 23 is an apparatus for graphics processing including at least one processor coupled to a memory and configured to implement a method as in any of aspects 1-22. [0143] Aspect 24 may be combined with aspect 23 and includes that the apparatus is a wireless communication device.
[0144] Aspect 25 is an apparatus for graphics processing including means for implementing a method as in any of aspects 1-22. [0145] Aspect 26 is a computer-readable medium (e.g., a non-transitory computer-readable medium) storing computer executable code, the code when executed by at least one processor causes the at least one processor to implement a method as in any of aspects 1-22. [0146] Various aspects have been described herein. These and other aspects are within the scope of the following claims.
ABSTRACT This disclosure provides systems, devices, apparatus, and methods, including computer programs encoded on storage media, for configuring power-saving configurations for reprojection systems. A processor (e.g., a central processing unit (CPU) or a customized microprocessor) may obtain a first indicator of a plurality of layers of graphics content and a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device. The processor may configure at least one of a render rate or a render resolution of the plurality of layers based on the obtained information. The processor may output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of graphics content.

Claims (20)

  1. WHAT IS CLAIMED IS: 1. An apparatus for graphics processing, comprising: a memory; and a processor coupled to the memory and, based at least in part on information stored in the memory, the processor is configured to: obtain a first indicator of a plurality of layers of graphics content; obtain a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device; configure at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device; and output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of the graphics content. 2. The apparatus of claim 1, wherein, to configure at least one of the render rate or the render resolution, the processor is configured to: configure a remote compression setting, wherein the processor is configured to: output a third indication of the configured remote compression setting.
  2. 3. The apparatus of claim 2, wherein, to output the third indication of the configured remote compression setting, the processor is configured to: transmit, to a second device remote from the device, the third indication of the configured remote compression setting.
  3. 4. The apparatus of claim 2, wherein, to output the third indication of the configured remote compression setting, the processor is configured to: output, to a compression unit of the device, the third indication of the configured remote compression setting.
  4. 5. The apparatus of claim 1, wherein the first set of characteristics comprises at least one of: a third indicator of a downscale capability for the reprojection engine; a fourth indicator of an upscale capability for the reprojection engine; a fifth indicator of a rotation capability for the reprojection engine; or a sixth indicator of a warp capability for the reprojection engine.
  5. 6. The apparatus of claim 1, wherein the second set of content statistics comprises at least one of: a third indicator of a timestamp associated with a layer rendered by a rendering unit; or a fourth indicator of a head pose associated with the layer rendered by the rendering unit.
  6. 7. The apparatus of claim 1, wherein the third set of wireless characteristics comprises at least one of: a third indicator of an amount of data wirelessly transmitted to the device; a fourth indicator of a fourth set of transmission windows used to transmit wirelessly transmit the amount of the data to the device; a fifth indicator of a wireless transmission capacity associated with transmission of rendered layers to the device; or a sixth indicator of a fifth set of available data transmission windows associated with the transmission of the rendered layers to the device.
  7. 8. The apparatus of claim 1, wherein, to configure at least one of the render rate or the render resolution of the plurality of layers, the processor is configured to: configure at least one of the render rate or the render resolution for a rendering unit at the device.
  8. 9. The apparatus of claim 8, wherein, to output the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of the graphics content, the processor is configured to: output, to the rendering unit at the device, the second indicator of at least one of the configured render rate or the configured resolution, wherein, to obtain at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device, the processor is configured to: obtain, from the rendering unit of the device, a third indicator of the second set of content statistics.
  9. 10. The apparatus of claim 8, wherein, to output the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of the graphics content, the processor is configured to: transmit, to the device, the second indicator of at least one of the configured render rate or the configured resolution, wherein, to obtain at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device, the processor is configured to: receive, from the device, a third indicator of the second set of content statistics.
  10. 11. The apparatus of claim 1, wherein, to configure at least one of the render rate or the render resolution of the plurality of layers, the processor is configured to: configure at least one of the render rate or the render resolution for a rendering unit at a second device remote from the device.
  11. 12. The apparatus of claim 11, wherein, to output the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of the graphics content, the processor is configured to: output, to the rendering unit at the second device, the second indicator of at least one of the configured render rate or the configured resolution, wherein, to obtain at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device, the processor is configured to: obtain, from the rendering unit at the second device, a third indicator of the second set of content statistics.
  12. 13. The apparatus of claim 11, wherein, to output the second indicator of at least one of the configured render rate or the configured resolution for the plurality of layers of the graphics content, the processor is configured to: transmit, to the second device, the second indicator of at least one of the configured render rate or the configured resolution, wherein, to obtain at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device, the processor is configured to: receive, from the second device, a third indicator of the second set of content statistics.
  13. 14. The apparatus of claim 1, wherein, to obtain the first indicator of the plurality of layers of the graphics content, the processor is configured to: obtain, from a set of sensors at the device, the first indicator of the plurality of layers of the graphics content.
  14. 15. The apparatus of claim 1, wherein, to obtain the first indicator of the plurality of layers of the graphics content, the processor is configured to: receive, from the device, the first indicator of the plurality of layers of the graphics content.
  15. 16. The apparatus of claim 1, wherein, to obtain the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device, the processor is configured to: obtain, from the reprojection engine of the device, a third indicator of the first set of characteristics.
  16. 17. The apparatus of claim 1, wherein, to obtain the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device, the processor is configured to: receive, from the device, a third indicator of the first set of characteristics.
  17. 18. The apparatus of claim 1, further comprising a transceiver coupled to the processor, wherein, to obtain the first set of characteristics of the reprojection engine of the device and at least one of the second set of content statistics or the third set of wireless characteristics of the device, the processor is configured to: receive, from the transceiver, a third indicator of the third set of wireless characteristics, wherein the transceiver is configured to transmit at least a subset of the plurality of layers of the graphics content.
  18. 19. A method of graphics processing, comprising: obtaining a first indicator of a plurality of layers of graphics content; obtaining a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device; configuring at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device; and outputting a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of the graphics content.
  19. 20. A computer-readable medium storing computer executable code, the code when executed by a processor, causes the processor to: obtain a first indicator of a plurality of layers of graphics content; obtain a first set of characteristics of a reprojection engine of a device and at least one of a second set of content statistics or a third set of wireless characteristics of the device; configure at least one of a render rate or a render resolution of the plurality of layers based on at least one of the first set of characteristics of the reprojection engine of the device, the second set of content statistics, or the third set of wireless characteristics of the device; and output a second indicator of at least one of the configured render rate or the configured render resolution for the plurality of layers of the graphics content.
IL313019A 2024-05-22 2024-05-22 Adaptive energy-saving configurations for outsourcing systems IL313019A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
IL313019A IL313019A (en) 2024-05-22 2024-05-22 Adaptive energy-saving configurations for outsourcing systems
PCT/US2025/027792 WO2025244825A1 (en) 2024-05-22 2025-05-05 Adaptive power-saving configurations for reprojection systems

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
IL313019A IL313019A (en) 2024-05-22 2024-05-22 Adaptive energy-saving configurations for outsourcing systems

Publications (1)

Publication Number Publication Date
IL313019A true IL313019A (en) 2025-12-01

Family

ID=95895715

Family Applications (1)

Application Number Title Priority Date Filing Date
IL313019A IL313019A (en) 2024-05-22 2024-05-22 Adaptive energy-saving configurations for outsourcing systems

Country Status (2)

Country Link
IL (1) IL313019A (en)
WO (1) WO2025244825A1 (en)

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2612795B (en) * 2021-11-11 2024-10-30 Sony Interactive Entertainment Inc Graphics rendering
US11917011B2 (en) * 2022-01-10 2024-02-27 Meta Platforms Technologies, Llc Resilient rendering for augmented-reality devices

Also Published As

Publication number Publication date
WO2025244825A1 (en) 2025-11-27

Similar Documents

Publication Publication Date Title
US20240370967A1 (en) Visibility generation improvements in tile based gpu architectures
US20230335049A1 (en) Display panel fps switching
US12136166B2 (en) Meshlet shading atlas
US11893654B2 (en) Optimization of depth and shadow pass rendering in tile based architectures
US11748842B2 (en) Methods and apparatus for content shifting in foveated rendering
WO2025244825A1 (en) Adaptive power-saving configurations for reprojection systems
US20260004542A1 (en) Centering content layers
WO2026020405A1 (en) Regional render refresh
WO2024221309A1 (en) Ddic fallback scaling for partial frame update
US20250259389A1 (en) Delivering stored objects for xr applications
US20250252524A1 (en) Reordering chains of producing and consuming workloads to exploit gpu memory residency
US20250252648A1 (en) Optimization of graphics workloads through future frames
WO2025160792A1 (en) Dynamic video/camera power framework
US11869115B1 (en) Density driven variable rate shading
US12100335B2 (en) Power efficient display architecture
US20250348969A1 (en) Distinct visible event tokens for distinct execute commands and visible draw call primitives
WO2024238103A1 (en) Adaptive reprojection in split xr systems
US11373267B2 (en) Methods and apparatus for reducing the transfer of rendering information
WO2024238041A1 (en) Rendering multi-level image quality in multi-user split xr systems
WO2024249364A1 (en) Power efficient display color processing for static regions on screen
WO2023055655A1 (en) Meshlet shading atlas
WO2024228848A1 (en) Content migration in split xr systems