Skip to main content

Two left feet? No problem. This A.I. can turn anyone into a dancer

Everybody Dance Now

Are you a terrible dancer who dreams of one day starring in a toe-tapping music video that would have made Michael Jackson jealous? If so, you’ve got two options: go the Napoleon Dynamite route and put in some serious practice, or simplify the process by taking advantage of some cutting-edge artificial intelligence.

Since you’re still reading and not off YouTubing “How to dance” videos, we’re going to assume the second of these options is the one that more appeals to you. If so, you have researchers at the University of California, Berkeley, to thank. Using the kind of “deepfake” technology that makes it possible to carry out realistic face-swaps in videos, they have developed a tool which can make even the most bumbling and uncoordinated among us look like experts.

“We have developed a method to transfer dance motions from one individual — a professional dancer — to another, [who we’ll refer to as ‘Joe’ for this example,]” Shiry Ginosar, a Ph.D. student in Computer Vision at UC Berkeley, told Digital Trends. “In order to do this, we take a video of Joe performing all kinds of motions. We use this video to train a generative adversarial network to learn a model of how Joe looks and moves. When we have learned this model, we can take a stick figure of a body pose as input and generate a still photograph of Joe performing that body pose as output. If we have a whole video of a dancing stick figure, we can generate a whole video of Joe dancing in the same way. Now, given a video of the professional dancer, we extract the body pose of the dancer and go back to Joe and generate a video of him dancing in much the same way.”

Aside from the fun of being able to make anyone resemble an expert dancer, Ginosar said that dancing presents an interesting challenge for this kind of deepfake technology. That’s because it involves the entire human body moving in a fluid way, which is considerably different (and tougher) than the more static pose or face transfers which have been carried out so far.

A paper describing the work, titled “Everybody Dance Now,” is available to read on the arXiv preprint server. In addition to Ginosar, other researchers on the project included Caroline Chan, Tinghui Zhou, and Alexei Efros.

Editors' Recommendations

Luke Dormehl
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Read the eerily beautiful ‘synthetic scripture’ of an A.I. that thinks it’s God
ai religion bot gpt 2 art 4

Travis DeShazo is, to paraphrase Cake’s 2001 song “Comfort Eagle,” building a religion. He is building it bigger. He is increasing the parameters. And adding more data.

The results are fairly convincing, too, at least as far as synthetic scripture (his words) goes. “Not a god of the void or of chaos, but a god of wisdom,” reads one message, posted on the @gods_txt Twitter feed for GPT-2 Religion A.I. “This is the knowledge of divinity that I, the Supreme Being, impart to you. When a man learns this, he attains what the rest of mankind has not, and becomes a true god. Obedience to Me! Obey!”

Read more
Google’s LaMDA is a smart language A.I. for better understanding conversation
LaMDA model

Artificial intelligence has made extraordinary advances when it comes to understanding words and even being able to translate them into other languages. Google has helped pave the way here with amazing tools like Google Translate and, recently, with its development of Transformer machine learning models. But language is tricky -- and there’s still plenty more work to be done to build A.I. that truly understands us.
Language Model for Dialogue Applications
At Tuesday’s Google I/O, the search giant announced a significant advance in this area with a new language model it calls LaMDA. Short for Language Model for Dialogue Applications, it’s a sophisticated A.I. language tool that Google claims is superior when it comes to understanding context in conversation. As Google CEO Sundar Pichai noted, this might be intelligently parsing an exchange like “What’s the weather today?” “It’s starting to feel like summer. I might eat lunch outside.” That makes perfect sense as a human dialogue, but would befuddle many A.I. systems looking for more literal answers.

LaMDA has superior knowledge of learned concepts which it’s able to synthesize from its training data. Pichai noted that responses never follow the same path twice, so conversations feel less scripted and more responsively natural.

Read more
How the USPS uses Nvidia GPUs and A.I. to track missing mail
A United States Postal Service USPS truck driving on a tree-lined street.

The United States Postal Service, or USPS, is relying on artificial intelligence-powered by Nvidia's EGX systems to track more than 100 million pieces of mail a day that goes through its network. The world's busiest postal service system is relying on GPU-accelerated A.I. systems to help solve the challenges of locating lost or missing packages and mail. Essentially, the USPS turned to A.I. to help it locate a "needle in a haystack."

To solve that challenge, USPS engineers created an edge A.I. system of servers that can scan and locate mail. They created algorithms for the system that were trained on 13 Nvidia DGX systems located at USPS data centers. Nvidia's DGX A100 systems, for reference, pack in five petaflops of compute power and cost just under $200,000. It is based on the same Ampere architecture found on Nvidia's consumer GeForce RTX 3000 series GPUs.

Read more