Skip to main content

CAPTCHAs may be a thing of the past, thanks to new machine learning research

CAPTCHAs
Vicarious
CAPTCHA is an acronym for Completely Automated Public Turing test to tell Computers and Humans Apart. The term was coined in 2003, when the use of automated bots was becoming commonplace, and it refers to those annoying squiggly distorted letters that you have to type in when creating an online account. Although some companies have found ways around them, CAPTCHAs are still ubiquitous online.

Researchers at the AI company Vicarious may have just made CAPTCHAs obsolete, however, by creating a machine-learning algorithm that mimics the human brain. To simulate the human capacity for what is often described as “common sense,” the scientists built a computer vision model dubbed the Recursive Cortical Network.

Recommended Videos

“For common sense to be effective it needs to be amenable to answer a variety of hypotheticals — a faculty that we call imagination,” they noted in a post at their blog.

Please enable Javascript to view this content

The ability to decipher CAPTCHAs has become something of a benchmark for artificial intelligence research. The new Vicarious model, published in the journal Science, cracks the fundamentals of the CATCHPA code by parsing the text using techniques that are derived from human reasoning. We can easily recognize the letter A for example, even if it’s partly obscured or turned upside down.

As Dileep George, the co-founder of the company explained to NPR, the RCN takes far less training and repetition to learn to recognize characters by building its own version of a neural network. “So if you expose it to As and Bs and different characters, it will build its own internal model of what those characters are supposed to look like,” he said. “So it would say, these are the contours of the letter, this is the interior of the letter, this is the background, etc.”

These various features get put into groups, creating a hierarchal “tree” of related features. After several passes, the data is given a score for evaluation. CAPTCHAs can be identified with a high degree of accuracy. The RCN was able to crack the BotDetect system with 57 percent accuracy with far less training than conventional “deep learning” algorithms, which rely more on brute force and require tens of thousands of images before they can understand CAPTCHAs with any degree of accuracy.

Solving CATCHPAs is not the goal of the research, but it provides insight into how our brains work and how computers can replicate it, NYU’s Brenden Lake told Axios. “It’s an application that not everybody needs,” he said. “Whereas object recognition is something that our minds do every second of every day.”

“Biology has put a scaffolding in our brain that is suitable for working with this world. It makes the brain learn quickly,” George said. “So we copy those insights from nature and put it in our model. Similar things can be done in neural networks.”

Mark Austin
Former Digital Trends Contributor
Mark’s first encounter with high-tech was a TRS-80. He spent 20 years working for Nintendo and Xbox as a writer and…
ChatGPT’s latest model may be a regression in performance
chatGPT on a phone on an encyclopedia

According to a new report from Artificial Analysis, OpenAI's flagship large language model for ChatGPT, GPT-4o, has significantly regressed in recent weeks, putting the state-of-the-art model's performance on par with the far smaller, and notably less capable, GPT-4o-mini model.

This analysis comes less than 24 hours after the company announced an upgrade for the GPT-4o model. "The model’s creative writing ability has leveled up–more natural, engaging, and tailored writing to improve relevance & readability," OpenAI wrote on X. "It’s also better at working with uploaded files, providing deeper insights & more thorough responses." Whether those claims continue to hold up is now being cast in doubt.

Read more
ChatGPT already listens and speaks. Soon it may see as well
ChatGPT meets a dog

ChatGPT's Advanced Voice Mode, which allows users to converse with the chatbot in real time, could soon gain the gift of sight, according to code discovered in the platform's latest beta build. While OpenAI has not yet confirmed the specific release of the new feature, code in the ChatGPT v1.2024.317 beta build spotted by Android Authority suggests that the so-called "live camera" could be imminently forthcoming.

OpenAI had first shown off Advanced Voice Mode's vision capabilities for ChatGPT in May, when the feature was first launched in alpha. During a demo posted at the time, the system was able to identify that it was looking at a dog through the phone's camera feed, identify the dog based on past interactions, recognize the dog's ball, and associate the dog's relationship to the ball (i.e. playing fetch).

Read more
Perplexity takes aim at Google and Amazon with new shopping tool
An image of someone using Perplexity to shop online.

AI startup Perplexity has launched a new shopping feature that it describes as a “one-stop solution where you can research and purchase products.”

Available initially for U.S.-based users of its paid Pro subscription tier, with plans to expand internationally soon, Perplexity Shopping is a shot across the bows of rival services like Google and Amazon as the company seeks to attract more users to its AI chatbot while building out related services like search and online shopping.

Read more