Skip to main content

Nvidia and Microsoft are solving a big problem with Copilot+

The Surface Laptop running local AI models.
Luke Larsen / Digital Trends

When Microsoft announced Copilot+ PCs a few weeks back, one question reigned supreme: Why can’t I just run these AI applications on my GPU? At Computex 2024, Nvidia finally provided an answer.

Nvidia and Microsoft are working together on an Application Programming Interface (API) that will allow developers to run their AI-accelerated apps on RTX graphics cards. This includes the various Small Language Models (SLMs) that are part of the Copilot runtime, which are used as the basis for features like Recall and Live Captions.

With the toolkit, developers can allow apps to run locally on your GPU instead of the NPU. This opens up the door to not only more powerful AI applications, as the AI capabilities of GPUs are generally higher than NPUs, but also the ability to run on PCs that don’t currently fall under the Copilot+ umbrella.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

It’s a great move. Copilot+ PCs currently require a Neural Processing Unit (NPU) that’s capable of at least 40 Tera Operations Per Second (TOPS). At the moment, only the Snapdragon X Elite satisfies that criteria. Despite that, GPUs have much higher AI processing capabilities, with even low-end models reaching to 100 TOPS, and higher-end options scaling even higher.

In addition to running on the GPU, the new API adds retrieval-augmented generation (RAG) capabilities to the Copilot runtime. RAG gives the AI model access to specific information locally, allowing it to provide more helpful solutions. We saw RAG on full display with Nvidia’s Chat with RTX earlier this year.

Performance comparison with the RTX AI toolkit.
Nvidia

Outside of the API, Nvidia announced the RTX AI Toolkit at Computex. This developer suite, arriving in June, combines various tools and SDKs that allow developers to tune AI models for specific applications. Nvidia says that by using the RTX AI Toolkit, developers can make models four times faster and three times smaller compared to using open-source solutions.

We’re seeing a wave of tools that enable developers to build specific AI applications for end users. Some of that is already showing up in Copilot+ PCs, but I suspect we’ll see far more AI applications at this point next year. We have the hardware to run these apps, after all; now we just need the software.

Jacob Roach
Lead Reporter, PC Hardware
Jacob Roach is the lead reporter for PC hardware at Digital Trends. In addition to covering the latest PC components, from…
AMD and Intel are finally getting the Copilot+ treatment
The Asus Zenbook S 16 sitting on a coffee table.

Microsoft confirmed at IFA 2024 that its slate of Copilot+ features are coming to AMD and Intel laptops later this year. In a blog post, Microsoft revealed that AMD and Intel PCs that meet the minimum Copilot+ requirements will receive the AI features in November through free Windows 11 updates.

The inclusion of AMD and Intel has been up in the air since Microsoft released Copilot+ laptops exclusively with Qualcomm CPUs earlier this year. Although AMD and Intel held strong that the AI features would be available on Ryzen AI 300 and Lunar Lake CPUs, respectively, neither company would provide a definitive time frame.

Read more
Microsoft: AI is no replacement for human expertise
Microsoft Copilot allows you to ask an AI assistant questions within Office apps.

Microsoft has updated the terms of service going into effect at the end of September and is clarifying that its Copilot AI services should not be used as a replacement for advice from actual humans.

AI-based agents are popping up across industries as chatbots are increasingly used for customer service calls, health and wellness applications, and even doling out legal advice. However, Microsoft is once again reminding its customers that its chatbots responses should not be taken as gospel. “AI services are not designed, intended, or to be used as substitutes for professional advice,” the updated Service Agreement reads.

Read more
Nvidia’s next-gen GPUs may be delayed due to ‘design flaws’
Nvidia introducing its Blackwell GPU architecture at GTC 2024.

Dark clouds are looming over the future of Nvidia's best graphics cards. According to a new report, Nvidia told some of its partners that it will be delaying its upcoming Blackwell GPUs, and is now aiming for an early 2025 release instead. Delays are one thing, but the cause is perhaps the most worrying part of it all -- design flaws. What does this mean for Nvidia's RTX 50-series?

This worrying report originates from The Information, which cites two sources who helped produce the Blackwell chip, as well as its server hardware. Bloomberg recounts that the chips may be delayed by over three months at this point. Nvidia is preparing B100 and B200 chips for some of the world's most prominent tech companies, including Meta, Google, and Microsoft, so these delays could hit pretty hard.

Read more