Skip to main content

We may have just learned how Apple will compete with ChatGPT

An iPhone on a table with the Siri activation animation playing on the screen.
Omid Armin / Unsplash

As we approach Apple’s Worldwide Developers Conference (WWDC) in June, the rumor mill has been abuzz with claims over Apple’s future artificial intelligence (AI) plans. Well, there have just been a couple of major developments that shed some light on what Apple could eventually reveal to the world, and you might be surprised at what Apple is apparently working on.

According to Bloomberg, Apple is in talks with Google to infuse its Gemini generative AI tool into Apple’s systems and has also considered enlisting ChatGPT’s help instead. The move with Google has the potential to completely change how the Mac, iPhone, and other Apple devices work on a day-to-day basis, but it could come under severe regulatory scrutiny.

Recommended Videos

The partnership comes as something of a surprise, given recent rumors about Apple’s AI efforts. For months, leaks and rumors have claimed that Apple has been working on its own generative AI model, dubbed Ajax, that would potentially power all manner of features inside Apple’s operating systems. There’s been talk of how many of Apple’s apps and services, from Pages to Siri, could be smartened up with its own AI. Yet, that idea seems to have taken a blow after Bloomberg’s revelation.

That’s not the only reason why this rumored deal may seem startling. For one thing, we know that Apple likes to have control over its systems, and that often means developing new features in-house. Will Apple be happy to sign over such an important new development to another company?

In addition, Apple and Google have been rivals for years. The two companies have very different philosophies: Google often prefers to use the cloud to increase speed and performance. For Apple, the priority is doing things on-device to boost user security and privacy. The question many people will be asking is how two very different tech firms can find a compromise that works for both of them and boosts your iPhone’s capabilities.

Siri, powered by Gemini

The Google Gemini AI logo.
Google

Yet Apple and Google are not averse to working together. The two have a deal for Google to be the default search engine on Apple devices, after all, with many people complaining that it makes it difficult to change your default search engine in iOS and macOS.

However, I think what this mooted deal really suggests is that Apple’s own AI is nowhere near ready for prime time. If it were, it wouldn’t need to enlist the services of Gemini or ChatGPT. Apple doesn’t want to fall behind its AI rivals any further, but it looks like it doesn’t have time to finish its own model.

So, if Gemini is coming to an iPhone near you, what kind of features will it offer? Bloomberg reporter Mark Gurman believes Gemini’s AI features will “theoretically be baked into Siri and other apps.” Apple’s in-house AI, meanwhile, would be “focused on proactively providing users with information and conducting tasks on their behalf in the background.”

If Apple picks Gemini to power its generative AI efforts, that could be a major boost to Siri, potentially giving Apple’s assistant the shot in the arm it needs to catch up with rival services. Right now, the gulf between Siri and generative AIs like ChatGPT is absolutely enormous.

How long will the deal last?

Blue Titanium iPhone 15 Pro in hand.
Christine Romero-Chan / Digital Trends

Interestingly, we know that Apple is pouring resources into giving its generative AI model a leg up, perhaps so that it can one day take over from the likes of Gemini on your device. That’s because a recently published research paper from Apple shows how the company’s in-house generative AI has been trained on a mixture of text and images, which enabled it to achieve “state-of-the-art few-shot results” across a range of benchmarks.

The model, dubbed MM1, has been able to perform in-depth, multi-step reasoning without needing much prompting from the user. Apple claims it also exhibits impressive performance when it comes to image captioning and working out natural-language meanings.

According to VentureBeat, “This points to the potential for large multimodal models to tackle complex, open-ended problems that require grounded language understanding and generation.” Yet it looks like all of that will have to wait for another time, what with Apple seeking a deal with Google to integrate Gemini into its devices. Gurman doesn’t believe Apple will reveal its Gemini partnership at WWDC, but I wouldn’t be surprised to see it announced later this year. If that’s the case, it may be a long time before Ajax gets the go-ahead.

There’s another big question, though: How long Apple will want to stick with Gemini powering its services? With regulatory measures and privacy concerns possibly rearing their heads, I expect Apple will want to switch to its own Ajax AI model sooner rather than later. But if Ajax truly isn’t ready yet, when exactly that will happen is anyone’s guess.

Alex Blake
Alex Blake has been working with Digital Trends since 2019, where he spends most of his time writing about Mac computers…
ChatGPT’s latest model may be a regression in performance
chatGPT on a phone on an encyclopedia

According to a new report from Artificial Analysis, OpenAI's flagship large language model for ChatGPT, GPT-4o, has significantly regressed in recent weeks, putting the state-of-the-art model's performance on par with the far smaller, and notably less capable, GPT-4o-mini model.

This analysis comes less than 24 hours after the company announced an upgrade for the GPT-4o model. "The model’s creative writing ability has leveled up–more natural, engaging, and tailored writing to improve relevance & readability," OpenAI wrote on X. "It’s also better at working with uploaded files, providing deeper insights & more thorough responses." Whether those claims continue to hold up is now being cast in doubt.

Read more
ChatGPT just improved its creative writing chops
a phone displaying the ChatGPT homepage on a beige bbackground.

One of the great strengths of ChatGPT is its ability to aid in creative writing. ChatGPT's latest large language model, GPT-4o, has received a bit of a performance boost, OpenAI announced Wednesday. Users can reportedly expect "more natural, engaging, and tailored writing to improve relevance & readability" moving forward.

https://twitter.com/OpenAI/status/1859296125947347164

Read more
ChatGPT already listens and speaks. Soon it may see as well
ChatGPT meets a dog

ChatGPT's Advanced Voice Mode, which allows users to converse with the chatbot in real time, could soon gain the gift of sight, according to code discovered in the platform's latest beta build. While OpenAI has not yet confirmed the specific release of the new feature, code in the ChatGPT v1.2024.317 beta build spotted by Android Authority suggests that the so-called "live camera" could be imminently forthcoming.

OpenAI had first shown off Advanced Voice Mode's vision capabilities for ChatGPT in May, when the feature was first launched in alpha. During a demo posted at the time, the system was able to identify that it was looking at a dog through the phone's camera feed, identify the dog based on past interactions, recognize the dog's ball, and associate the dog's relationship to the ball (i.e. playing fetch).

Read more