Skip to main content

IBM is cutting deep-learning processing times from days down to hours

deep learning
Daniel Kaesler/123RF
Deep learning uses algorithms inspired by the way human brains operate to put computers to work on tasks too big for organic gray matter. On Monday, IBM announced that a new record for the performance of a large neural network working with a large data set.

The company’s new deep-learning software brings together more than 256 graphics processing units across 64 IBM Power systems. The speed improvements brought about by the research come as a result of better communication between the array of GPUs.

Faster GPUs provide the necessary muscle to take on the kind of large scale problems today’s deep-learning systems are capable of tackling. However, the faster the components are, the more difficult it is to ensure that they are all working together as one cohesive unit.

As individual GPUs work on a particular problem, they share their learning with the other processors that make up the system. Conventional software is not capable of keeping up with the speed of current GPU technology, which means that time is wasted as they wait around for one another’s results.

Hillery Hunter, IBM’s director of systems acceleration and memory, compared the situation to the well-known parable of the blind men and the elephant. The company’s distributed deep-learning project has resulted in an API that developers can be used in conjunction with deep-learning frameworks to scale to multiple servers, making sure that their GPUs remain synchronized.

IBM recorded image recognition accuracy of 33.8 percent on a test run using 7.5 million images from the ImageNet-22K database. The previous best-published result was 29.8 percent, which was posted by Microsoft in October 2014 — in the past, accuracy has typically edged forward at a rate of about one percent in new implementations, so an improvement of four percent is considered to be a very good result.

Crucially, IBM’s system managed to achieve this in seven hours; the process that allowed Microsoft to set the previous record took 10 days to complete.

“Speed and scalability, which means higher accuracy, means that we can quickly retrain an AI model after there is a new cyber-security hack or a new fraud situation,” Hunter told Digital Trends. “Waiting for days or weeks to retrain the model is not practical — so being able to train accurately and within hours makes a big difference.”

These massive improvements in terms of speed, combined with advances in terms of accuracy make IBM’s distributed deep-learning software a major boon for anyone working with this technology. A technical preview of the API is available now as part of the company’s PowerAI enterprise deep-learning software.

Editors' Recommendations

Brad Jones
Former Digital Trends Contributor
Brad is an English-born writer currently splitting his time between Edinburgh and Pennsylvania. You can find him on Twitter…
Japanese researchers use deep learning A.I. to get driftwood robots moving
driftwood ai robots move mzmzmjaymw

walk

Did you ever make sculptures out of found objects like driftwood? Researchers at the University of Tokyo have taken this same idea and applied it to robots. In doing so, they’ve figured out a way to take everyday natural objects like pieces of wood and get deep reinforcement learning algorithms to figure out how to make them move. Using just a few basic servos, they’ve opened up a whole new way of building robots -- and it’s pretty darn awesome.

Read more
What’s that liquid? IBM’s flavor-identifying ‘e-tongue’ will tell you
ibm e tongue project 48055705116 bdaf7b65de o

IBM Hypertaste: An AI-assisted e-tongue for fast and portable fingerprinting of complex liquids

With its Watson technology, IBM has helped create a pretty convincing artificial brain. But now it’s seemingly ready to move onto other body parts as well -- and it’s settled on the tongue as a next step. As developed by computer scientists at IBM Research, the A.I.-assisted e-tongue is a portable device, equipped with special sensors, that allow it to taste and identify different liquids.

Read more
Get ready to waste your day with this creepily accurate text-generating A.I.
ai spots writing by fake news feature

Whether you believe it was one of the most dangerous versions of artificial intelligence created or dismiss it as a massive unnecessary PR exercise, there’s no doubt that the GPT-2 algorithm created by research lab OpenA.I. caused a lot of buzz when it was announced earlier this year.

Revealed in February, OpenA.I. said it developed an algorithm too dangerous to release to the general public. Although only a text generator, GPT-2 supposedly generated text so crazily humanlike that it could convince people that they were reading a real text written by an actual person. To use it, all a user had to do would be to feed in the start of the document, and then let the A.I. take over to complete it. Give it the opening of a newspaper story, and it would even manufacture fictitious “quotes.” Predictably, news media went into overdrive describing this as the terrifying new face of fake news. And for potentially good reason.

Read more