Skip to main content

Baidu’s new A.I. can mimic your voice after listening to it for just one minute

Image used with permission by copyright holder

We’re not in the business of writing regularly about “fake” news, but it’s hard not to be concerned about the kind of mimicry technology is making possible. First, researchers developed deep learning-based artificial intelligence (A.I.) that can superimpose one person’s face onto another person’s body. Now, researchers at Chinese search giant Baidu have created an A.I. they claim can learn to accurately mimic your voice — based on less than a minute’s worth of listening to it.

“From a technical perspective, this is an important breakthrough showing that a complicated generative modeling problem, namely speech synthesis, can be adapted to new cases by efficiently learning only from a few examples,” Leo Zou, a member of Baidu’s communications team, told Digital Trends. “Previously, it would take numerous examples for a model to learn. Now, it takes a fraction of what it used to.”

Recommended Videos

Baidu Research isn’t the first to try and create voice-replicating A.I. Last year, we covered a project called Lyrebird, which used neural networks to replicate voices including President Donald Trump and former President Barack Obama with a relatively small number of samples. Like Lyrebird’s work, Baidu’s speech synthesis technology doesn’t sound completely convincing, but it’s an impressive step forward — and way ahead of a lot of the robotic A.I. voice assistants that existed just a few years ago.

The work is based around Baidu’s text-to-speech synthesis system Deep Voice, which was trained on upwards of 800 hours of audio from a total of 2,400 speakers. It needs just 100 5-second sections of vocal training data to sound its best, but a version trained on only 10 5-second samples was able to trick a voice-recognition system more than 95 percent of the time.

“We see many great use cases or applications for this technology,” Zou said. “For example, voice cloning could help patients who lost their voices. This is also an important breakthrough in the direction of personalized human-machine interfaces. For example, a mom can easily configure an audiobook reader with her own voice. The method [additionally] allows creation of original digital content. Hundreds of characters in a video game would be able to have unique voices because of this technology. Another interesting application is speech-to-speech language translation, as the synthesizer can learn to mimic the speaker identity in another language.”

For a deeper dive into this subject, you can listen to a sample of the voices or read a paper describing the work.

Luke Dormehl
Former Digital Trends Contributor
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Alexa and Siri can’t understand the tone of your voice, but Oto can
oto voice intonation ai call center mouthpiece

Speech-recognition technology is everywhere these days, most notably in A.I. smart assistants such as Amazon’s Alexa, Apple’s Siri, and Google's Assistant. But as anyone who has ever had a conversation IRL (in real life) will know, speech isn’t just about the words that a person says, but the tone of voice in which they say them. It’s one reason that text-based conversations online can be such a nightmare, since the basic words themselves don’t allow for sufficient nuance to always convey a person’s meaning.

One exciting startup looking to inject more understanding into speech recognition is Oto, a spinoff from the prestigious SRI International, which helped spawn Siri more than a decade ago. Oto is working on voice-intonation technology that will, at least initially, enable call centers to better understand the vocal emotions of callers and sales agents alike.

Read more
New cardiology A.I. knows if you’ll die soon. Doctors can’t explain how it works
cardiology ai predicts death toe tag

Here’s a scenario scary enough to warrant a horror movie: An artificial intelligence that is able to accurately predict your chances of dying in the next year by looking at heart test results, despite the fact that the results may look totally fine to trained doctors. The good news: The technology might just wind up saving your life one day.

“We have developed two different artificial intelligence algorithms that can automatically analyze electrical tracings from the heart and make predictions about the likelihood of a future important clinical event,” Brandon Fornwalt, from Pennsylvania-based healthcare provider Geisinger, told Digital Trends.

Read more
The Tesla Supercharger network is expanding again
Teslas parked using the Tesla Superchargers.

EV owners worried about getting stranded before their next charge may breathe a bit easier now: The Tesla Supercharger network has resumed its expansion over the last three months, after massive layoffs at Tesla had led to a slide in new stalls earlier this year.Tesla opened 2,800 new Supercharger stalls globally during the three-month period ending September 30, an increase of 23% from the year earlier, according to the company’s official Tesla Charging account on X (formerly Twitter).Tesla also delivered 1.4 terawatt hours (TWh) of energy during the third quarter, which translates to 27% year-over-year growth. The Tesla Charging account said the amount of energy delivered enabled Tesla owners to save over 150 million gallons of gasoline, offsetting more than 3 billion pounds of CO2.According to the website Supercharge.info, which relies on user contributions to track the opening of new stalls, there were 2,677 Supercharger stations in North America at the end of September, 125 more than at the end of the third quarter.The additions of new stalls, however, comes in the wake of a 31% slide in the second quarter, after a massive round of firings at the company removed 500 employees from the Supercharger team in April.The third-quarter expansion should still go some ways to appease the concerns of Tesla and non-Tesla electric vehicle (EV) owners, as EV manufacturers are lining up in droves to be able to use the Supercharger network.In September, General Motors electric vehicles made by Chevrolet, GMC and Cadillac joined the growing list of vehicles to have adopted Tesla’s North American Charging Standard (NACS). The network already allows access, via adapters, to EVs made by Ford, Rivian, Honda and Volvo. The likes of Nissan, Hyundai, Toyota, BMW, Volkswagen, Volvo and Jaguar have also signed agreements to start allowing access in 2025.
The Tesla map of Superchargers across the U.S. reveals they are concentrated in major cities and across East-West transit corridors.According to Tesla, a Supercharger can add up to 200 miles of range in 15 minutes.

Read more