If you’ve ever tinkered with the video settings of a game, you have probably encountered the option to turn VSync on or off.
In the majority of cases, this is not accompanied by any explanation as to what VSync is, and the name itself does little to dispel the confusion.
VSync is short for Vertical Synchronization, and it is used to combat the problem of screen tearing.
In essence, this issue occurs when your graphics card produces more frames than your monitor is capable of displaying on your visible screen.
For example, if your GPU can output 80 frames per second and your monitor has a 60hz refresh rate, the monitor will attempt to squeeze those 80 frames into its 60 refresh cycles. This causes some parts of the screen to be displayed separately, resulting in screen tearing.

Table of ContentsShow
What Is VSync?
Although PC gaming has now existed for a few decades, it wasn’t until the swift technological advancement of graphics cards in the 2000s that we encountered the issue of screen tearing. This is because GPU technology quickly moved forward while monitor technology progressed far more gradually.
Screen tearing was accepted by some gamers as simply playing a waiting game until monitors with faster refresh rates were developed. Fortunately, some innovative people had other ideas.
VSync is the original solution developed to address the screen tearing issue. It involves limiting the GPU output software to match the monitor’s maximum refresh rate. Theoretically, this sounds like a perfect solution, but there are several concerns and these are the reason why VSync is largely considered obsolete in 2025.
One of these issues occurs when the GPU is unable to perform up to the monitor’s refresh rate. As a result, the monitor will leave the previous image on display until the next one is ready, which causes visible stuttering.
Fortunately, there are now technologies far superior suited to handle this issue, but more on those later.
Likely the largest issue with VSync is its input lag. This is particularly frustrating in games where quick reactions are necessary, such as a shooter game. It’s an even more significant issue if the game in question is a multiplayer game, where your opponent could end up beating you simply because they use a different syncing solution.
Should You Use VSync?
From what has been said so far, the simplest answer would be yes, you should use VSync. The benefits are obvious, while the drawbacks are less likely to hinder you.
However, the more nuanced answer is that you should only use VSync if it’s absolutely necessary. The reason for this is simple: there are better alternatives.
As mentioned previously, VSync technology is quite simple and it didn’t take long for other GPU giants to provide their own solutions.
VSync Alternatives
First and foremost, we must acknowledge the deserving credit; VSync was the original solution. It was a suitable solution back then and grew to be a crucial aspect of graphics settings in the following decade.
While we can commend VSync for its achievement, we have to admit that its time has largely passed and that there are now superior alternatives available. Let’s take a look at which one is currently the best.
AdaptiveSync
First up, we have AdaptiveSync, which is distinct from NVIDIA’s Adaptive VSync (notice the extra ‘V’). AdaptiveSync is the sole technology mentioned here that wasn’t developed by either AMD or NVIDIA. It was developed by VESA, the organization responsible for the DisplayPort standard, which is frequently used today.
To avoid confusion, both AMD and NVIDIA are members of VESA, but they weren’t part of the development process. AdaptiveSync is also a supplementary standard, which means any member of VESA can use it. Both AMD and NVIDIA used it to develop their own brand-specific screen-tearing solutions.
Probably the finest thing about this standard is the way it smooths out stuttering when the FPS drops below the monitor’s refresh rate. It’s so remarkable that it can feel like there are more frames than there genuinely are.
AdaptiveSync accomplishes this by enabling an image to be displayed once it’s fully rendered while maintaining the previous image up in the interim. How does that differ from VSync?
It’s quite straightforward, really. AdaptiveSync adjusts the refresh rate of the monitor and forces it to wait until the frame is prepared before loading it up. You can take a look at the image below to better understand this.

FastSync
**FastSync** is NVIDIA’s version of AdaptiveSync and an upgrade of its own Adaptive VSync, which was considered a bit of a catastrophe. As the industry leader that it is, NVIDIA rapidly responded to its own failure with FastSync.
FastSync attempts to achieve the same thing as the AdaptiveSync standard but runs into some issues where stuttering and jerking are more noticeable. It does its job, but we can only recommend using it if you’re gaming online.
It’s worth pointing out that, much like the tech it’s trying to emulate, it’s still a superior option than VSync.
Enhanced Sync

While the previous entry was NVIDIA’s endeavor at fixing the vertical synchronization issues, this is AMD’s effort at it.
There isn’t much more to say about the way AMD addressed this. Enhanced Sync can mainly be considered the same as FastSync, but with AMD’s unique branding all over it.
The same issue that affects FastSync, intermittent stuttering, is also noticeable here. Although both technologies were developed to eliminate stuttering, neither succeeded entirely. However, they were both superb attempts, considering their respective release dates. All in all, Enhanced Sync does manage to reduce the stuttering level of VSync.
G-Sync

We’re doing well now. G-Sync is NVIDIA’s response to screen tearing and it functions superbly.
For NVIDIA, being the first company to release a groundbreaking product has its advantages. NVIDIA was determined to take full advantage of this and sought to leave no money on the table.
That’s right, as much as G-Sync is the ideal solution for screen tearing, it also comes at a cost.
For NVIDIA to be able to solve the issue of adjusting the monitor’s refresh rate, they had to invent a module that is implemented directly into the monitors themselves. This module, in conjunction with the software solution, is able to dynamically alter the monitor’s refresh rate to match the frames generated by the GPU.
This remarkable and innovative solution requires the installation of the module by monitor manufacturers. As such, NVIDIA chose to charge money for it, hence the additional cost for consumers.
The key reason why G-Sync is a superior technology than VSync is that it eliminates the input lag. The main drawback since the inception of VSync is finally a thing of the past.
G-Sync Ultimate

This is the latest addition to the G-Sync brand and can finally somewhat justify the additional cost. NVIDIA made sure that this was noticeable and required a comprehensive and thorough inspection to be passed by monitor manufacturers in order to implement their G-Sync module.
The reason for this is simple: G-Sync Ultimate brings a lot to the table. NVIDIA managed to add 152 backlight zones, which means the IPS panel is capable of producing HDR images with improved precision.
FreeSync

As you might have guessed, this is AMD’s screen-tearing solution. However, AMD decided to allow their module to be completely free, hence the name FreeSync.
Its reasoning is probably twofold. NVIDIA was the initial company to reach the market and AMD lagged behind. NVIDIA also chose to charge money for their product, so AMD making its solution complimentary could give it a tactical edge.
To avoid any uncertainty, we have to admit that FreeSync is equally as excellent as G-Sync and the fact that it’s free makes it a more appealing option. However, things aren’t always that simple and that’s also the case here.
One of the reasons why NVIDIA charges monitor manufacturers for the use of its G-Sync module is to ensure the uniqueness of its cards. G-Sync monitors can only operate with NVIDIA cards.
One might think that AMD would do the same but the reality is that FreeSync monitors can be used with both AMD and NVIDIA GPUs. There is also a restriction here, as not all FreeSync monitors are compatible with NVIDIA GPUs.
In 2019, NVIDIA shocked the world when they announced that their G-Sync cards would have support for FreeSync monitors. Unsurprisingly, there are strict standards to be met, but the fact remains that the perfect option is currently a FreeSync monitor.
Much like G-Sync technology, FreeSync reduces the frustrating input lag. Something else that AMD can boast about is FreeSync’s reduced flicker, which really comes in handy in those prolonged gaming sessions.
FreeSync Premium

AMD attempted to improve upon their FreeSync model with FreeSync Premium, and we can say that they did a truly exceptional job.
Premium brings a host of enhancements, with its most significant being the Low Framerate Compensation. This innovation responds to the frame rate dropping below the monitor’s range by doubling the FPS number and using it as the monitor’s refresh rate. This eliminates screen tearing and, although the frame rate will be reduced, it will at least be stable.
FreeSync Premium Pro

As its name suggests, Premium Pro is another upgrade of FreeSync technology. In this case, it aims to preserve balance as AMD’s ray-tracing-capable cards are released.
Much like G-Sync Ultimate, Premium Pro aims to maintain the suitable refresh rate of the monitor to display HDR images that are produced by the recent graphics cards.
Why You Shouldn’t Use VSync
As suggested, VSync simply isn’t a perfect solution. The worst part is that it causes input lag, which no serious gamer wants to deal with.
If available, we would recommend either G-Sync or FreeSync, depending on which graphics card you have. As mentioned earlier, FreeSync appears to be a superior option at this stage with even NVIDIA adapting to it.
It’s crucial to point out that, if your monitor is neither G-Sync nor FreeSync compatible, you’ll have to go for VSync, which should eliminate display tearing anyway.



