Frame timing in VRVirtual Reality More info
See in Glossary mode works exactly like it does in VSync-enabled non-VR mode (see documentation on the Execution order of event functions). The only difference is that Unity does not depend on the underlying 3D SDK’s VSyncVertical synchronization (VSync) is a display setting that caps a game’s frame rate to match the refresh rate of a monitor, to prevent image tearing.
See in Glossary, but instead whichever VR SDK it currently renders with. There is no benefit to renderingThe process of drawing graphics to the screen (or to a render texture). By default, the main camera in Unity renders its view to the screen. More info
See in Glossary faster than the display can refresh. Instead, Unity focuses on utilizing the time most efficiently.
In multithreaded rendering, Unity synchronizes the simulation thread and the rendering thread on the CPU. This reduces latency. The simulation thread primarily executes the scriptsA piece of code that allows you to create your own Components, trigger game events, modify Component properties over time and respond to user input in any way you like. More info
See in Glossary, sounds, AI, and other tasks needed for the simulation to run. The render thread focuses on submitting draw calls to the graphics driver, which in turn validates them and sends them to the GPU. To achieve the greatest parallelism between threads, Unity executes graphics commands as they come in, and simulates the next frame at the same time.
Unity waits for the GPU to completely render and display the last frame, then submits rendering commands for the next frame.
Before Unity can submit the first rendering command that depends on the view transformation matrix, it must first get the view matrix from the VR SDK. To keep latency as low as possible, the VR SDK predicts the head transform twice per frame:
One prediction to render the current frame. This prediction corresponds with where your head actually is, in real space, when the frame arrives on the screen.
One prediction to simulate the following frame.
Unity applies the rendering prediction for the current frame to camerasA component which creates an image of a particular viewpoint in your scene. The output is either drawn to the screen or captured as a texture. More info
See in Glossary, controllers, and anything that needs information for SceneA Scene contains the environments and menus of your game. Think of each unique Scene file as a unique level. In each Scene, you place your environments, obstacles, and decorations, essentially designing and building your game in pieces. More info
See in Glossary rendering. Unity uses the simulation prediction for the following frame if it is unable to render the following frame.
For more information on how the hardware handles rendering, see the manufacturer documentation for your head-mounted display.
If Unity doesn’t submit the fully rendered frame in time for it to be ready for the next display refresh, a few things might happen, depending on which VR SDK is active. The VR SDK might:
Display the previously submitted frame. This looks like judder, and greatly reduces the quality of the experience.
Rotationally reproject the previously submitted frame based on the current head pose. This is a crude approximation that works well as a fallback for static content, but any content with animation or positional movement does not look correct.
Apply some form of positional reprojection. This might be temporal to fill in missing details.
For more information on reprojection (sometimes called time warping), see the XinReality wiki entry on Timewarp.
This all happens automatically. At this point, it’s usually too late for Unity to successfully finish the next frame that it should already be rendering. Instead, the VR SDK takes the last frame it received and reprojects it to the latest predicted pose, and Unity waits until the next opportunity to render a full frame to get back on track. If your app continually drops frames, Unity only renders every other frame.
For more information on how the hardware handles rendering, see the manufacturer documentation for your head-mounted display.
Each display has a specific refresh rate that Unity is bound to. You can get this at run time with VRDevice.refreshRate. A value of 1 divided by the refreshRate
gives you the time allocated to a single frame.
For example, this might be 11.1 milliseconds in the case of a refresh rate of 90 hz.
If your app is GPU-bound, the Unity ProfilerA window that helps you to optimize your game. It shows how much time is spent in the various areas of your game. For example, it can report the percentage of time spent rendering, animating, or in your game logic. More info
See in Glossary displays XR.WaitForGPU
in excess of one frame’s time in milliseconds.
If your app is CPU-bound, a frame takes longer than the designated frame time, but the Unity Profiler displays XR.WaitForGPU
shorter than the one frame.