Skip to main content

Nvidia's Drive PX 2 to be used in autonomous racecars for the Roborace Championship

nvidia roborace gtc2016 image by chief design officer daniel simon  ltd
Image used with permission by copyright holder
During the GPU Technology Conference keynote on Tuesday, Nvidia CEO Jen-Hsun Huang said that the company’s Drive PX 2 AI supercomputer will be used in autonomous racecars that will compete in the Roborace Championship. He also revealed that the company is working on an end-to-end mapping system using Drive PX 2-based cars and Tesla GPUs, the latter of which are typically installed in the data center.

For starters, the Roborace Championship, first reported on Digital Trends here, is part of the Formula E ePrix “electric” racing series, meaning the racecars are driverless and utilize alternative “earth-friendly” energy sources. These one-hour races consist of ten teams that use two identical cars packing Nvidia’s Drive PX 2. Because these cars don’t rely on driver intuition to win, the teams must program the AI to be highly strategic.

According to Nvidia, the Drive PX 2 supercomputer is the size of a lunchbox, keeping with the overall compact design of the cars. The Drive PX 2 is also capable of up to 24 trillion operations a second for AI applications, thus providing “supercomputer-class” performance, or the processing power of 150 MacBook Pro laptops. And because of its deep learning capability, the cars will become smarter — and thus faster — the more they actually race.

Get your weekly teardown of the tech behind PC gaming
Check your inbox!

“Since the cars don’t need human drivers, these racecars are incredibly compact, and the designs — conceived by auto designer Daniel Simon, the man behind Tron: Legacy’s light cycles — are like nothing that’s been seen on a road, or a racetrack, before,” Nvidia’s Danny Shapiro said in a blog. “There’s no room in these racers for the trunk full of PCs that powered earlier generations of autonomous vehicles.”

Nvidia says the Drive PX 2 is capable of incorporating input from a number of sensors installed in the racecar such as GPS, cameras, radar, and more. Ultimately, the Drive PX 2 supercomputer and the Roborace Championship races should lead to smarter and safer driverless and standard cars for public consumption.

As for those high-definition maps mentioned in Tuesday’s keynote, the system will enable the rapid development of HD maps and frequent updates through the use of the Drive PX 2 and Tesla GPUs. These maps are important to driverless cars because they reduce the amount of processing the supercomputer performs as it incorporates inputs from multiple sensors. Just imagine how easy driving can become when you know exactly what’s up the street or around the corner.

The HD mapping system is an open platform based on Nvidia’s DriveWorks SDK.  It’s a “highly efficient” system that pushes most of the data processing onto the Drive PX 2 so that communication to the cloud is minimal. It also uses a technique that Nvidia calls “visual simultaneous localization and mapping” along with deep learning that handles the mapping process.

Nvidia says that the deep learning aspect helps detect important features during the mapmaking process, such as road signs, lanes, and landmarks. It can also recognize changes in the environment, thus the system is capable of recording and updating maps to be used by autonomous vehicles.

So why is this good news? Up until now, mapping was done by cars with numerous sensors that gathered huge volumes of data. In turn, this data was recorded and then processed offline. GPS alone is also old news, as autonomous vehicles require the use of exact details on the road. Precision is achieved when combining GPS with the car’s internal sensors and what’s called motion algorithms, which convert 2D data into 3D information.

Nvidia launched the Drive PX 2 during CES 2016 in January, billed as the world’s first in-car artificial intelligence supercomputer. Additional information regarding Nvidia’s Drive solutions can be found here.

Editors' Recommendations

Kevin Parrish
Former Digital Trends Contributor
Kevin started taking PCs apart in the 90s when Quake was on the way and his PC lacked the required components. Since then…
The Meta Quest 3 will get an exciting new type of app
A Windows app extends into 3D space via a Meta Quest 3 VR headset.

A Windows app extends into 3D space via a Meta Quest 3 VR headset. Microsoft

At Build 2024, Microsoft announced it partnered with Meta to extend Windows apps into 3D space with the help of a Quest VR headset. When working on physical objects, it’s important to have spatial awareness of components.

Read more
How to build a PC from scratch: A beginner’s guide
Installing RAM in a desktop PC.

Building a PC for the first time, or even the second or third time, can feel a little intimidating. But one of the best parts about building a computer is that, for the most part, the parts fit where they should, and don't fit where they shouldn't. A graphics card will fit in the graphics card slot, and good luck putting the CPU in the wrong socket.

With a little care, time, and this handy guide, you can build a PC without hassle. We're here to walk you through it.

Read more
Microsoft just made Paint relevant again
Person using Windows 11 laptop on their lap by the window.

The controversial Recall feature has grabbed all the headlines from Microsoft’s Copilot+ announcements yesterday, but this new AI feature is also making Paint relevant again.

It’s called Cocreator, and it’s a new AI feature that can turn your quick sketch, augmented by text, into a much more realistic and impressive image. The exciting thing is that it does all this in real time. It might not get it right the first time, so you'll need patience, and the more details you give about what you want in the image, the better.

Read more