Skip to main content

Intel’s Arc graphics cards have quietly become excellent

Intel’s Arc A770 and A750 were decent at launch, but over the past few months, they’ve started to look like some of the best graphics cards you can buy if you’re on a budget. Disappointing generational improvements from AMD and Nvidia, combined with high prices, have made it hard to find a decent GPU around $200 to $300 — and Intel’s GPUs have silently filled that gap.

They don’t deliver flagship performance, and in some cases, they’re just straight-up worse than the competition at the same price. But Intel has clearly been improving the Arc A770 and A750, and although small driver improvements don’t always make a splash, they’re starting to add up.

Recommended Videos

Silently improving

The backs of the Arc A770 and Arc A750 graphics cards.
Jacob Roach / Digital Trends

If you keep up with the world of graphics cards, you’ve probably heard about Intel doubling performance with a driver update earlier this year. That was only for DirectX 9 games, however, which were all but broken at launch. Intel doubled their performance, sure, but that only brought the A770 and A750 up to par.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

Since then, Intel has focused more attention on recent DirectX 11 and DirectX 12 games, and it’s clear some titles are receiving specific optimization. When I retested Cyberpunk 2077, for example, the Arc A750 went from 56 frames per second (fps) at 1080p to 76 fps — nearly a 36% increase. Similarly, the Arc A770 has gone from 59 fps to 83 fps since launch.

Previously, Cyberpunk 2077 was one of the worst showcases of the A750 and A770, falling short of even Nvidia’s RTX 3060. Now, it’s one of the best, with the A750 outperforming even the RTX 3060 Ti at 1080p and 1440p.

It’s not just Cyberpunk 2077. The A750 went from 86 fps in Horizon Zero Dawn at 1080p to 95 fps, while the A770 climbed from 98 fps to 106 fps. Again, the cards are now competing with GPUs like the RTX 3060 Ti and RX 6600 XT after taking a clear back seat.

The Arc A770 graphics card running in a PC.
Jacob Roach / Digital Trends

Those are impressive improvements, but they’re targeted. In a Vulkan-based game like Red Dead Redemption 2, Intel’s slew of driver updates didn’t move the needle at all. And in Assassin’s Creed Valhalla, there are some minor improvements, but they only account for a few frames.

What’s becoming clear, however, is that Intel’s claim that the Arc A750 and A770 have more fuel in the tank is holding up. Targeted optimization through drivers for specific games has brought some stark improvements. They aren’t universal, but if Intel keeps up its driver pace, the A750 and A770 could be a force to be reckoned with.

Competitive on price

Nvidia RTX 3060 Ti Founders Edition on a pink background.
Jacob Roach / Digital Trends

All of this comes in the context of price, though. For as impressive as Intel’s driver improvements are, the A750 and A770 are competing with last-gen GPUs from AMD and Nvidia in performance. Even the RX 7600 can blow past the A770 in most games, and it’s nearly $100 cheaper. Intel is starting to become competitive, though.

Take the Arc A750 at $250. It’s not a great option now that the RX 7600 is here at $270 and the last-gen RX 6600 XT is selling for around the same price. However, it’s been marked down to only $200 a handful of times recently, and at that price, it’s a steal.

That becomes apparent when you look at ray tracing. In Hogwarts Legacy, the RX 6700 XT  (around $350) squarely beats the A770 and A750 at 1080p (by about 18% and 37%, respectively). Flip on ray tracing, though, and suddenly even the A750 is matching AMD while the A770 claims a lead of 30%. The A770 is even competitive with the RTX 4060 Ti here.

Front of the AMD RX 7600.
Jacob Roach / Digital Trends

Similarly, AMD’s crop of GPUs at this price point can’t even maintain 30 fps in Cyberpunk 2077’s ray tracing mode at 1080p without the assistance of upscaling (you’ll need to jump up to a $500 RX 6800 to hit that mark), while the A750 and A770 are comfortably above 30 fps. That was a strength of Intel’s Arc GPUs at launch, being competitive with Nvidia at ray tracing, and even AMD’s new RX 7600 doesn’t change that.

There are other considerations here as well, specifically VRAM. The A770 is outfitted with 16GB of VRAM, while options like the RTX 3060 Ti are only outfitted with 8GB. In Horizon Zero Dawn, the A770 is a few frames slower than the RTX 3060 Ti. In more recent VRAM-limited games like Resident Evil 4 and The Last of Us Part 1, however, the A770 beats out the RTX 3060 Ti by close to 10%.

Intel Arc A770 GPU installed in a test bench.
Jacob Roach / Digital Trends

Intel still has a long road ahead, but it has made a ton of progress very quickly. At list price, the A750 and A770 are competitive, even if they aren’t the best option for everyone. On sale, it’s hard justifying anything from AMD or Nvidia at the same price. If you can find the A750 at $200, it’s suddenly competing with a GPU like the RTX 3050 on price, while offering around a 30% boost in performance.

My biggest hope is that Intel will stick with it. The A770 and A750 have proven over the last few months that Team Blue has a fighting chance in the world of GPUs. The A770 and A750 are a solid swing out of the gate, but the future of Arc really hinges on what the upcoming Battlemage, Celestial, and Druid generations can offer.

Jacob Roach
Lead Reporter, PC Hardware
Jacob Roach is the lead reporter for PC hardware at Digital Trends. In addition to covering the latest PC components, from…
I’m worried Intel is making a mistake with Arrow Lake
Someone holding the Core i9-12900KS processor.

For the last several years, every new generation from Intel has felt like a make-or-break moment. Now, with Arrow Lake CPUs, the stakes are even higher. Intel is facing unprecedented financial troubles, and although it still makes some of the best processors, the silicon giant that used to loom over the PC industry isn’t as strong as it once was.

Arrow Lake is yet another major shift. The CPUs kill Intel’s long-standing Hyper-Threading feature. They introduce two new core architectures. And they debut the Core Ultra branding on desktop, along with the new LGA 1851 socket. I’m worried that Intel’s strategy won’t work with Arrow Lake, though.

Read more
Sorry, gamers — Intel’s new CPUs won’t deliver any gains
A render for an Intel Arrow Lake CPU.

Intel is setting expectations for its upcoming Arrow Lake-S desktop CPUs. Although the company is holding strong that the new generation will be competitive with the best processors when they release on October 24, the new range of CPUs won't deliver much, if any, performance gains for gamers -- and that's coming from Intel itself.

To kick off the Arrow Lake generation, now called Intel Core Ultra 200S, Intel is releasing five processors. You can see the standard Core Ultra 9, 7, and 5 models in the table below, along with Core Ultra 7 and 5 models that cut the integrated graphics for a slightly lower price. All five of the processors are unlocked for overclocking with the new LGA 1851 socket. Unlike AMD's new Zen 5 CPUs, Core Ultra 200S chips require a new motherboard as Intel retires its LGA 1700 socket.

Read more
Intel did the unthinkable with its new Arrow Lake CPUs
A render of an Intel Core Ultra 200-S chip.

It finally happened. Intel killed Hyper-Threading on its desktop CPUs. The new Arrow Lake range, called Core Ultra 200S, ditches the simultaneous multi-threading (SMT) feature that Intel has held onto for more than a decade. And according to Intel, it doesn't need the extra threads to still deliver a generational performance improvement, even up against the best processors.

Intel says the new range, which we break down in detail in our post focused its gaming potential, can deliver an 8% performance improvement in single-threaded workloads over the previous generation, and a 4% uplift compared to the Ryzen 9 9950X. Those are pretty small margins, but the real impressive stuff comes in multi-threaded performance.

Read more