Skip to main content

AMD is losing the AI battle, and it’s time it started to worry

Both AMD and Nvidia make some of the best graphics cards on the market, but it’s hard to deny that Nvidia is usually in the lead. I don’t just mean the vast difference in market share. In this generation, it’s Nvidia that has the behemoth GPU that’s better than all the other cards, while AMD doesn’t have an answer to the RTX 4090 just yet.

Another thing that AMD doesn’t have a strong answer to right now is artificial intelligence. Even though I’m switching to AMD for personal use, it’s difficult to ignore the facts: Nvidia is winning the AI battle. Why is there such a marked difference, and will this become more of a problem for AMD down the line?

Recommended Videos

It’s not all about gaming

The RX 7900 XTX.
AMD / AMD

Most of us buy graphics cards based on two things — budget and gaming capabilities. AMD and Nvidia both know that the vast majority of their high-end consumer cards end up in gaming rigs of some sort, although professionals pick them up too. Still, gamers and casual users make up the biggest part of this segment of the market.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

For years, the GPU landscape was all about Nvidia, but in the last few generations, AMD made big strides — so much so that it trades blows with Nvidia now. Although Nvidia leads the market with the RTX 4090, AMD’s two RDNA 3 flagships (the RX 7900 XTX and RX 7900 XT) are powerful graphics cards that often outperform similar offerings from Nvidia, while being cheaper than the RTX 4080.

If we pretend that the RTX 4090 doesn’t exist, then comparing the RTX 4080 and 4070 Ti to the RX 7900 XTX and XT tells us that things are pretty even right now; at least as far as gaming is concerned.

And then, we get to ray tracing and AI workloads, and this is where AMD drops off a cliff.

GeForce RTX logo is shown on the side of a graphics card.
Image used with permission by copyright holder

There’s no way to sugarcoat this — Nvidia is simply better at running AI-generated tasks than AMD is right now. It’s not really an opinion, it’s more of a fact. This is also not the only ace up its sleeve.

Tom’s Hardware recently tested AI inference on Nvidia, AMD, and Intel cards, and the results were not favorable to AMD at all.

To compare the GPUs, the tester benchmarked them in Stable Diffusion, which is an AI image creator tool. Read the source article if you want to know all the technical details that went into setting up the benchmarks, but long story short, Nvidia outperformed AMD, and Intel Arc A770 did so poorly that it barely warrants a mention.

Even getting Stable Diffusion to run outside of an Nvidia GPU seems to be quite the challenge, but after some trial and error, the tester was able to find projects that were somewhat suited to each GPU.

After the testing, the end result was that Nvidia’s RTX 30-series and RTX 40-series both did fairly well (albeit after some tweaking for the latter). AMD’s RDNA 3 line also held up well, but the last-gen RDNA 2 cards were fairly mediocre. However, even AMD’s best card was miles behind Nvidia in these benchmarks, showing that Nvidia is simply faster and better at tackling AI-related tasks.

Nvidia cards are the go-to for professionals in need of a GPU for AI or machine learning workloads. Some people may buy one of the consumer cards and others may pick up a workstation model instead, such as the confusingly named RTX 6000, but the fact remains that AMD is often not even on the radar when such rigs are being built.

RX 7900 XT and RX 7900 XTX performance in Cyberpunk 2077 with ray tracing.
Jacob Roach / Digital Trends

Let’s not gloss over the fact that Nvidia also has a strong lead over AMD in things like ray tracing and Deep Learning Super Sampling (DLSS). In our own benchmarks, we found that Nvidia still leads the charge in ray tracing over AMD, but at least Team Red seems to be making steps in the right direction.

This generation of GPUs is the first one where the ray tracing gap is closing. In fact, AMD’s RX 7900 XTX outperforms Nvidia’s RTX 4070 Ti in that regard. However, Nvidia’s Ada Lovelace GPUs have another edge in the form of DLSS 3, a technology that copies entire frames, instead of just pixels, using AI. Once again, AMD is falling behind.

Nvidia has a long history of AI

Nvidia GeForce RTX 4090 GPU.
Jacob Roach / Digital Trends

AMD and Nvidia graphics cards are vastly different on an architectural level, so it’s impossible to compare them completely. However, one thing we do know is that Nvidia’s cards are optimized for AI in terms of their very structure, and this has been the case for years.

Nvidia’s latest GPUs are equipped with Compute Unified Device Architecture (CUDA) cores, whereas AMD cards have Compute Units (CUs) and Stream Processors (SPs). Nvidia also has Tensor Cores that aid the performance of deep learning algorithms, and with Tensor Core Sparsity, they also help the GPU skip unnecessary computations. This reduces the time the GPU needs to perform certain tasks, such as training deep neural networks.

CUDA cores are one thing, but Nvidia has also created a parallel computing platform by the same name, which is only accessible to Nvidia graphics cards. CUDA libraries allow programmers to harness the power of Nvidia GPUs in order to run machine learning algorithms much faster.

The development of CUDA is what really sets Nvidia apart from AMD. While AMD didn’t really have a good alternative, Nvidia invested heavily in CUDA, and in turn, most of the AI progress in the last years was made using CUDA libraries.

AMD has done some work on its own alternatives, but it’s fairly recent when you compare it to the years of experience Nvidia has had. AMD’s Radeon Open Compute platform (ROCm) lets developers accelerate compute and machine learning workloads. Under that ecosystem, it has launched a project called GPUFORT.

GPUFORT is AMD’s effort to help developers transition away from Nvidia cards and onto AMD’s own GPUs. Unfortunately for AMD, Nvidia’s CUDA libraries are much more widely supported by some of the most popular deep learning frameworks, such as TensorFlow and PyTorch.

Despite AMD’s attempts to catch up, the gap only grows wider each year as Nvidia continues to dominate the AI and ML landscape.

Time is running out

Nvidia and AMD CEOs are shown side by side in a split-screen view.
Nvidia/AMD

Nvidia’s investment in AI was certainly sound. It left Nvidia with a booming gaming GPU lineup alongside a powerful range of cards capable of AI- and ML-related tasks. AMD is not quite there yet.

Although AMD seems to be trying to optimize its cards on the software side with yet-unused AI cores on its latest GPUs, it doesn’t have the software ecosystem that Nvidia has built up.

AMD plays a crucial role as the only serious competitor to Nvidia, though. I can’t deny that AMD has made great strides both in the GPU and CPU markets over the past years. It managed to climb back out of irrelevance and become a strong alternative to Intel, making some of the best processors available right now. Its graphics cards are now also competitive, even if it’s just for gaming. On a personal level, I’ve been leaning toward AMD instead of Nvidia because I’m against Nvidia’s pricing approach in the last couple of generations. Still, that doesn’t make up for AMD’s lack of AI presence.

It’s very visible in programs such as ChatGPT that AI is here to say, but it’s also present in countless other things that go unnoticed by most PC users. In a gaming PC, AI works in the background performing tasks such as real-time optimization and anti-cheat measures in gaming. Non-gamers see plenty of AI on a daily basis too, because AI is found in ever-present chatbots, voice-based personal assistants, navigation apps, and smart home devices.

As AI permeates our daily lives more and more, and computers are needed to perform tasks that only increase in complexity, GPUs are also expected to keep up. AMD has a tough task ahead, but if it doesn’t get serious about AI, it may be doomed to never catch up.

Monica J. White
Monica is a computing writer at Digital Trends, focusing on PC hardware. Since joining the team in 2021, Monica has written…
The numbers are in. Is AMD abandoning gamers for AI?
AMD's RX 7700 XT in a test bench.

The data for the first quarter of 2024 is in, and it's bad news for the giants behind some of the best graphics cards. GPU shipments have decreased, and while every GPU vendor experienced this, AMD saw the biggest drop in shipments. Combined with the fact that AMD's gaming revenue is down significantly, it's hard not to wonder about the company's future in the gaming segment.

The report comes from the analyst firm Jon Peddie Research, and the news is not all bad. The PC-based GPU market hit 70 million units in the first quarter of 2024, and from year to year, total GPU shipments (which includes all types of graphics cards) increased by 28% (desktop GPU shipments dropped by -7%, and CPU shipments grew by 33.3%). Comparing the final quarter of 2023 to the beginning of this year looks much less optimistic, though.

Read more
New Nvidia update suggests DLSS 4.0 is closer than we thought
A hand holding the RTX 4090 GPU.

Nvidia might be gearing up for DLSS 4.0. A new update for Nvidia's Streamline pipeline includes updated files for DLSS Super Resolution and Frame Generation that bring the version to 3.7.

This is a fairly small update aimed at developers. I haven't had a chance to try out DLSS 3.7, but like previous updates, I assume this includes some small tweaks to performance and image quality. Nvidia commits these updates pretty frequently, usually centered around reducing visual artifacts in games.

Read more
How 8GB VRAM GPUs could be made viable again
Screenshot of full ray tracing in Cyberpunk 2077.

Perhaps there is still some hope for GPUs with low VRAM. According to a new patent published by Microsoft, the company worked out a method that could make ray tracing and path tracing more viable in terms of how much video memory (VRAM) they use. As of right now, without using upscaling techniques, seamless ray tracing requires the use of one of the best graphics cards—but this might finally change if this new method works out as planned.

This new patent, first spotted by Tom's Hardware, describes how Microsoft hopes to reduce the impact of ray tracing on GPU memory. It addresses the level of detail (LOD) philosophy, which is already something that's used in games but not in relation to ray tracing, and plans to use LOD to adjust ray tracing quality dynamically, thus lowering the load that the GPU -- particularly its memory -- has to bear.

Read more