What Is Ray Tracing And Is It Worth It For Gaming?

What is ray tracing exactly, and is it worth enabling for gaming? Does Nvidia, AMD and Intel GPUs support ray tracing and at what performance cost?

Ray tracing, simply put, is the simulation of realistic light interactions with virtual objects as if it was in the actual world.

When done effectively, this enhances the detail quality for rendered graphics at extremely high levels, making them look even more realistic than ever before.

However, as many already understand, there is substantially more to the idea of live light directions than merely improving aesthetics.

Especially for games, where the technology’s merits and drawbacks make its practicality a perennial hot topic of debate.

Let’s dive right into it!

Table of ContentsShow

How Does Ray Tracing Work?

Ray Tracing Explained

Ray tracing attempts to simulate natural light interaction with mathematically calculated paths along an imaginary eye and the virtual environment that it “observes”.

Each pixel displayed on the screen arises from a simulated light beam that calculates where it should be reflected from where the eye (or any observing medium) is looking.

As anticipated, the color properties of the surfaces affected by this process can dynamically change simply by switching viewing angles or moving objects around. Just like how objects in the real world change their shadow properties or luminosity when moved nearer or away from a light source.

Why Is Ray Tracing Important?

Ray Tracing On And Off

Ray tracing is crucial because it theoretically skips the manual labor of pre-set or engine-based graphics lighting, and instead uses those calculations for real-time, dynamic lighting effects.

Because the light ray calculations optimize for the interaction of these vectors at each surface point available, it also creates much more lifelike reflections and vivid color tones.

This is unlike traditional lighting techniques in graphics design, where each pixel change has to be painstakingly modified manually. Lighting effects on regular graphical assets would only work for a pre-set number of changes built by its designer.

Even when using contemporary graphics engines, there is an ultimate limit on how rendering software could “assume” light interactions, leaving what should be perfectly reflective surfaces in games as generalized mirrored backdrops.

Rasterization vs. Ray Tracing

Ray Tracing

Rasterization is the technique of converting a collection of points and curves into a series of pixels with visible images (and depth, if 3D) using calculations.

In simpler terms, this is the traditional manual way for graphics hardware to render images on the screen. This includes the manual color shifting of pixels when lighting is simulated.

Because rasterization is predetermined, calculations are relatively straightforward, provided that the graphics card has enough power to crunch all the requisite numbers in the shortest amount of time.

Ray tracing, while it should theoretically look superior (or more lifelike), presents a much higher level of hardware computational challenge.

First, it has to dynamically calculate all surface color property changes using simulated light rays. The light rays also require extra processing power since their paths also need to be computed.

On top of that, the graphics card still needs to use a substantial chunk of rasterization under ray tracing tasks. After all, it still has to account for the objects inside the game, as well as the overall visual quality of the environment (material textures, surface tessellation, etc.).

Even at our current adoption level, ray tracing still reduces a considerable amount of rendered frames.

Without using upscaling technologies like DLSS, FSR, and XeSS, it requires a greater number of dedicated cores and enhanced pure rasterization performance to overcome the frame rate dips.

Ray Tracing On Nvidia GPUs

Nvidia RTX 4090

Nvidia may not have originated the concept, but the company has spearheaded the technological trend of ray tracing support with the launch of the RTX lineup of GPUs.

RTX (Ray Tracing Texel eXtreme) started with the 20-series (Turing), evolving from the conventional GTX (Giga Texel Shader eXtreme) line of graphics cards, which concluded with the GTX 10-series (Pascal).

The inaugural ray tracing cores of the 20 series were scarcely impressive. But with DLSS at least, Nvidia was able to demonstrate the promising future of fully detailed ray-traced environments without making the game completely unplayable.

By the time the RTX 40-series (Lovelace) arrived, ray tracing hardware technology had improved enough that 1080p gaming was effectively “conquered” by Nvidia’s technology. And even at 1440p, the latest hardware can still provide exceptional ray tracing features with enhanced graphical settings without relying too much on DLSS.

Ray Tracing On AMD GPUs

AMD Radeon GPU

AMD started quite late into the ray tracing game, with the company’s focus on RDNA 2 architecture being more on rasterization effectiveness than providing novel technological features.

But at least when the RX 6000 series arrived, Team Red could offer a baseline level of ray tracing performance that can still be reasonably enjoyed at adequate frame rates. That was if you have the premier models.

As such, owners of anything below a Radeon RX 6700 XT are generally advised not to use ray tracing options.

They can still provide top-level detail with maxed-out conventional graphics settings for contemporary triple-A titles. But with the restricted ray tracing cores of these lower-tier graphics cards, frame rates take a much bigger hit, sometimes even with the use of FSR 2.0.

Ray Tracing On Intel GPUs

Intel ARC GPUs

Intel surprised the world with its inaugural venture into the GPU market with the Intel Arc series and the first Alchemist architecture.

Despite lacking any top-level offerings, Team Blue was at least eager on providing feature parity by delivering exceptional compute-based performance. The GPUs also had a slew of other offerings like XeSS and native (hardware-based) AV1 encoding.

Very notably, the ray tracing abilities of the Intel Arc A770 and Arc A750 generally exceed what comparable RTX cards can provide.

The condition, of course, is that the game in question is already functioning properly and effortlessly with the Intel GPUs in the first place. Otherwise, the ray tracing-supported game never runs at all or performs marginally below the Geforce RTX 3060.

If you are playing on a natively DX12-supported game, turning on ray tracing is certainly a delight for both Intel graphics cards. Experience enhanced frame rates than its competitors without relying on FSR and XeSS!

Should You Turn On Ray Tracing?

Fortnite Ray Tracing

If you have a substantially high-end or upper-mid-range GPU that has dedicated ray tracing features, then you can most likely turn on ray tracing and still enjoy sufficient frame rates. However, there are a few caveats, namely:

  • Basic graphics settings take a higher priority. Try to max out other graphic settings first before turning ray tracing on. Ray tracing enhances the experience of regular rasterized graphics. It would not feel as amazing or as stunning if other graphical quality settings were sacrificed for it.
  • Remember that ray tracing implementation varies wildly depending on the game. Cyberpunk 2077 takes a far harder hit to frame rates when ray tracing is turned on, for example, than other games like Call of Duty: Warzone or Doom Eternal. Some games even require ray tracing to be turned on all the time, such as Metro Exodus: Enhanced Edition.
  • Don’t turn on ray tracing if you can’t get respectable frame rates from reasonable graphics settings. The Radeon RX 6400 may theoretically support ray tracing with its dedicated hardware, but it is far from even being able to play any ray tracing-supported game at a usable FPS with it.
  • If the game demands high-level real-time action, ray tracing might not be as beneficial. Ray tracing works more practically for games where you can appreciate the visuals without getting too busy with the environment. Something like No Man’s Sky, or even a slower-paced game like Resident Evil: Village.

Recommended Reads

Should I Overclock My GPU
Should I Overclock My GPU? Is Overclocking Worth It?
Christopher Lee

Christopher is currently the editor-in-chief for GPU Mag. He has worked as a computer technician for 3 years before landing a job at Dell where he is now currently working full-time remotely.