Skip to main content

Researchers are programming robots to learn as human babies do

Google is betting big on artificial intelligence. Robots have come a long way over the past decade or so, but they’re still not as good at interacting with humans as they could be. In fact, robots still struggle to do some basic tasks.

A team at Carnegie Mellon, however, is trying to fix that. The team, led by assistant professor Abhinav Gupta, is taking a new approach — allowing robots to play with everyday physical objects and explore the world to help them learn — exactly as a human baby would.

Recommended Videos

“Psychological studies have shown that if people can’t affect what they see, their visual understanding of that scene is limited,” said Lerrel Pinto, a PhD student in the group, in a report by The Verge. “Interaction with the real world exposes a lot of visual dynamics.”

The group first showed off its tech last year, and the demo helped it land a three-year, $1.5 million award from Google, which will be used to expand on the number of robots that are being used in the study. More robots allows the researchers to gather data more quickly, which basically helps the group build increasingly advanced robots.

But the team isn’t just looking toward more robots to help speed up data gathering. It’s also trying to teach robots skills that will, in turn, help the robot learn other skills. The team also uses adversarial learning — which, according to the Verge report, is akin to a parent teaching a child how to catch a ball by pitching increasingly difficult throws. Apparently, taking this approach results in significantly faster learning than alternative methods.

It will certainly be interesting to see what comes of the project, and we’ll likely hear more about it as time goes on. Check out the video below to see the robots in action.

Robot Adversaries for Grasp Learning
Christian de Looper
Christian de Looper is a long-time freelance writer who has covered every facet of the consumer tech and electric vehicle…
I let Gemini turn complex research into podcasts. I’ll never go back
Audio Overview in Gemini.

The shift away from Google Assistant, and into the Gemini era, is nearly in its last stages. One can feel nostalgic about the eponymous virtual assistant, but it’s undeniable that the arrival of Gemini has truly changed what an AI agent can do for us.

The language understanding chops are far better with Gemini. Conversations are natural, app interactions are fluid, integration with other Google products is rewarding, and even in its free state, Gemini takes Siri to the cleaners even on an iPhone.

Read more
Microsoft 365 Copilot gets an AI Researcher that everyone will love
Researcher agent in action inside Microsoft 365 Copilot app.

Microsoft is late to the party, but it is finally bringing a deep research tool of its own to the Microsoft 365 Copilot platform across the web, mobile, and desktop. Unlike competitors such as Google Gemini, Perplexity, or OpenAI’s ChatGPT, all of which use the Deep Research name, Microsoft is going with the Researcher agent branding.
The overarching idea, however, isn’t too different. You tell the Copilot AI to come up with thoroughly researched material on a certain topic or create an action plan, and it will oblige by producing a detailed document that would otherwise take hours of human research and compilation. It’s all about performing complex, multi-step research on your behalf as an autonomous AI agent.
Just to avoid any confusion early on, Microsoft 365 Copilot is essentially the rebranded version of the erstwhile Microsoft 365 (Office) app. It is different from the standalone Copilot app, which is more like a general purpose AI chatbot application.
Researcher: A reasoning agent in Microsoft 365 Copilot
How Researcher agent works?
Underneath the Researcher agent, however, is OpenAI’s Deep Research model. But this is not a simple rip-off. Instead, the feature’s implementation in Microsoft 365 Copilot runs far deeper than the competition. That’s primarily because it can look at your own material, or a business’ internal data, as well.
Instead of pulling information solely from the internet, the Researcher agent can also take a look at internal documents such as emails, chats, internal meeting logs, calendars, transcripts, and shared documents. It can also reference data from external sources such as Salesforce, as well as other custom agents that are in use at a company.
“Researcher’s intelligence to reason and connect the dots leads to magical moments,” claims Microsoft. Researcher agent can be configured by users to reference data from the web, local files, meeting recordings, emails, chats, and sales agent, on an individual basis — all of them, or just a select few.

Why it stands out?

Read more
Samsung might put AI smart glasses on the shelves this year
Google's AR smartglasses translation feature demonstrated.

Samsung’s Project Moohan XR headset has grabbed all the spotlights in the past few months, and rightfully so. It serves as the flagship launch vehicle for a reinvigorated Android XR platform, with plenty of hype from Google’s own quarters.
But it seems Samsung has even more ambitious plans in place and is reportedly experimenting with different form factors that go beyond the headset format. According to Korea-based ET News, the company is working on a pair of smart glasses and aims to launch them by the end of the ongoing year.
Currently in development under the codename “HAEAN” (machine-translated name), the smart glasses are reportedly in the final stages of locking the internal hardware and functional capabilities. The wearable device will reportedly come equipped with camera sensors, as well.

What to expect from Samsung’s smart glasses?
The Even G1 smart glasses have optional clip-on gradient shades. Photo by Tracey Truly / Digital Trends
The latest leak doesn’t dig into specifics about the internal hardware, but another report from Samsung’s home market sheds some light on the possibilities. As per Maeil Business Newspaper, the Samsung smart glasses will feature a 12-megapixel camera built atop a Sony IMX681 CMOS image sensor.
It is said to offer a dual-silicon architecture, similar to Apple’s Vision Pro headset. The main processor on Samsung’s smart glasses is touted to be Qualcomm’s Snapdragon AR1 platform, while the secondary processing hub is a chip supplied by NXP.
The onboard camera will open the doors for vision-based capabilities, such as scanning QR codes, gesture recognition, and facial identification. The smart glasses will reportedly tip the scales at 150 grams, while the battery size is claimed to be 155 mAh.

Read more