Skip to main content

All of the internet now belongs to Google’s AI

Google Bard being shown off at Google I/O 2023.
Image used with permission by copyright holder

Google’s latest update to its privacy policy will make it so that the company has free range to scrape the web for any content that can benefit building and improving its AI tools.

“Google uses information to improve our services and to develop new products, features, and technologies that benefit our users and the public,” the new Google policy says. “For example, we use publicly available information to help train Google’s AI models and build products and features like Google Translate, Bard, and Cloud AI capabilities.”

Gizmodo notes that the policy has been updated to say “AI models” when it previously said, “for language models.” Additionally, the policy added Bard and Cloud AI, when it previously only mentioned Google Translate, for which it collected data.

The privacy policy, which was updated over the weekend, appears especially ominous because it indicates that any information you produce online is up for grabs for Google to use for training its AI models.

The aforementioned wording seems to describe not just those in the Google ecosystem in one way or another but is detailed in such a way that the brand could have access to information from any part of the web.

Major issues surrounding the mass development of artificial intelligence are questions about privacy, plagiarism, and whether AI can dispel correct information. Early versions of chatbots such as ChatGPT are based on large language models (LLMs) that used already public sources, such as the common crawl web archive, WebText2, Books1, Books2, and Wikipedia as training data.

Early ChatGPT was infamous for becoming stuck on information beyond 2021 and subsequently filling in responses with false data. This could likely be one of the reasons Google would want unfettered access to web data to benefit tools such as Bard, to have real-world and potentially real-time training for its AI models.

Gizmodo also noted that Google could use this new policy to collect old, but still human-generated content, such as long-forgotten reviews or blog posts, to still have a feel of how human text and speech is developed and distributed. Still, it remains to be seen exactly how Google will use the data it collects.

Several social media platforms, including Twitter and Reddit, which are major sources of up-to-date information have already limited their public access in the wake of AI chatbot popularity, to the chagrin of their entire communities.

Both platforms have closed free access to their APIs, which restricts users from downloading massive amounts of posts for sharing elsewhere, under the guise of protecting their intellectual property. This instead broke many of the third-party tools that make both Twitter and Reddit run smoothly.

Both Twitter and Reddit have had to deal with other setbacks and controversies as their owners’ concerns heighten about AI taking over.

Editors' Recommendations

Fionna Agomuoh
Fionna Agomuoh is a technology journalist with over a decade of experience writing about various consumer electronics topics…
Google may build Gemini AI directly into Chrome
The Google Gemini AI logo.

Google is now fleshing out its newly unified Gemini AI system in its browser with its first attempt at implementing Chat with Gemini into the Chrome Omnibox.

This latest effort will update Google Chrome with a Chat with Gemini shortcut in the Chrome Omnibox, allowing users to access the AI chatbot feature without having to go to the Gemini website, according to WindowsReport. The Omnibox serves as an address bar and search bar, and it adds multiple other tasks to a browser. Now with a simple @ prompt, you can also access Google's AI chatbot to answer questions, create images, and generate summaries, among other tasks.

Read more
All RTX GPUs now come with a local AI chatbot. Is it any good?
A window showing Nvidia's Chat with RTX.

It's been difficult to justify packing dedicated AI hardware in a PC. Nvidia is trying to change that with Chat with RTX, which is a local AI chatbot that leverages the hardware on your Nvidia GPU to run an AI model.

It provides a few unique advantages over something like ChatGPT, but the tool still has some strange problems. There are the typical quirks you get with any AI chatbot here, but also larger issues that prove Chat with RTX needs some work.
Meet Chat with RTX
Here's the most obvious question about Chat with RTX: How is this different from ChatGPT? Chat with RTX is a local large language model (LLM). It's using TensorRT-LLM compatible models -- Mistral and Llama 2 are included by default -- and applying them to your local data. In addition, the actual computation is happening locally on your graphics card, rather than in the cloud. Chat with RTX requires an Nvidia RTX 30-series or 40-series GPU and at least 8GB of VRAM.

Read more
Google One AI Premium: what is it, and how much does it cost?
Google Gemini app on Android.

Google is introducing a new AI-enabled tier of Google One, shortly after announcing a rebrand of its Bard AI chatbot.

The chatbot is now called Gemini, following the sequence of its base large language model (LLM). Google has also recently announced several other service updates, including Gemini Advanced, a paid tier of Gemini based on the new Ultra 1.0 language model.

Read more