Skip to main content

Anthropic aims to fix one of the biggest problems in AI right now

the Anthropic logo
Anthropic

Hot on the heels of the announcement that its Claude 3.5 Sonnet large language model beat out other leading models, including GPT-4o and Llama-400B, AI startup Anthropic announced Monday that it plans to launch a new program to fund the development of independent, third-party benchmark tests against which to evaluate its upcoming models.

Per a blog post, the company is willing to pay third-party developers to create benchmarks that can “effectively measure advanced capabilities in AI models.”

“Our investment in these evaluations is intended to elevate the entire field of AI safety, providing valuable tools that benefit the whole ecosystem,” Anthropic wrote in a Monday blog post. “Developing high-quality, safety-relevant evaluations remains challenging, and the demand is outpacing the supply.”

The company wants submitted benchmarks to help measure the relative “safety level” of an AI based on a number of factors, including how well it resists attempts to coerce responses that might include cybersecurity; chemical, biological, radiological, and nuclear (CBRN); and misalignment, social manipulation, and other national security risks. Anthropic is also looking for benchmarks to help evaluate models’ advanced capabilities and is willing to fund the “development of tens of thousands of new evaluation questions and end-to-end tasks that would challenge even graduate students,” essentially testing a model’s ability to synthesize knowledge from a variety of sources, its ability to refuse cleverly worded malicious user requests, and its ability to respond in multiple languages.

Anthropic is looking for “sufficiently difficult,” high-volume tasks that can involve as many as “thousands” of testers across a diverse set of test formats that help the company inform its “realistic and safety-relevant” threat modeling efforts. Any interested developers are welcome to submit their proposals to the company, which plans to evaluate them on a rolling basis.

Andrew Tarantola
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
Meta and Google made AI news this week. Here were the biggest announcements
Ray-Ban Meta Smart Glasses will be available in clear frames.

From Meta's AI-empowered AR glasses to its new Natural Voice Interactions feature to Google's AlphaChip breakthrough and ChromaLock's chatbot-on-a-graphing calculator mod, this week has been packed with jaw-dropping developments in the AI space. Here are a few of the biggest headlines.

Google taught an AI to design computer chips
Deciding how and where all the bits and bobs go into today's leading-edge computer chips is a massive undertaking, often requiring agonizingly precise work before fabrication can even begin. Or it did, at least, before Google released its AlphaChip AI this week. Similar to AlphaFold, which generates potential protein structures for drug discovery, AlphaChip uses reinforcement learning to generate new chip designs in a matter of hours, rather than months. The company has reportedly been using the AI to design layouts for the past three generations of Google’s Tensor Processing Units (TPUs), and is now sharing the technology with companies like MediaTek, which builds chipsets for mobile phones and other handheld devices.

Read more
This was a huge week for AI. Here are the 5 biggest announcements you need to know
Apple Intelligence and Camera Control.

The race toward achieving AGI (artificial general intelligence) continued apace with what felt like a monumental week in the rapid development of AI.

From Apple giving us a taste of its Intelligence to huge advances in AI-generated video, let's take a look at some of the top AI stories from this week.
Apple Intelligence soft launches

Read more
YouTube can now generate a 6-second video Short using AI
The Digital Trends YouTube channel as seen on an iPhone and on a TV in the background.

YouTube content creators will soon have a slew of new AI-empowered tools at their disposal, the company announced Wednesday at its Made on YouTube event in New York City, including the ability to generate a complete six-second YouTube Shorts video clip with a text prompt.

The new capability arrives thanks to the integration of Google Deep Mind's Veo video generation model into YouTube Shorts. Built to compete with the likes of OpenAI's Sora, Runway's Gen-3 Alpha, Adobe's Firefly or Kuaishou Technology's Kling, Veo can generate six-second clips at 1080p resolution across a wide range of cinematic themes and styles.

Read more