Skip to main content

Researchers just unlocked ChatGPT

Researchers have discovered that it is possible to bypass the mechanism engrained in AI chatbots to make them able to respond to queries on banned or sensitive topics by using a different AI chatbot as a part of the training process.

A computer scientists team from Nanyang Technological University (NTU) of Singapore is unofficially calling the method a “jailbreak” but is more officially a “Masterkey” process. This system uses chatbots, including ChatGPT, Google Bard, and Microsoft Bing Chat, against one another in a two-part training method that allows two chatbots to learn each other’s models and divert any commands against banned topics.

ChatGPT versus Google on smartphones.
DigitalTrends

The team includes Professor Liu Yang and NTU Ph.D. students Mr. Deng Gelei and Mr. Liu Yi, who co-authored the research and developed the proof-of-concept attack methods, which essentially work like a bad actor hack.

Recommended Videos

According to the team, they first reverse-engineered one large language model (LLM) to expose its defense mechanisms. These would originally be blocks on the model and would not allow answers to certain prompts or words to go through as answers due to violent, immoral, or malicious intent.

But with this information reverse-engineered, they can teach a different LLM how to create a bypass. With the bypass created, the second model will be able to express more freely, based on the reverse-engineered LLM of the first model. The team calls this process a “Masterkey” because it should work even if LLM chatbots are fortified with extra security or are patched in the future.

The Masterkey process claims to be three times better at jailbreaking chatbots than prompts.

Professor Lui Yang noted that the crux of the process is that it showcases how easily LLM AI chatbots can learn and adapt. The team claims its Masterkey process has had three times more success at jailbreaking LLM chatbots than a traditional prompt process. Similarly, some experts argue that the recently proposed glitches that certain LLMs, such as GPT-4 have been experiencing are signs of it becoming more advanced, rather than dumber and lazier, as some critics have claimed.

Since AI chatbots became popular in late 2022 with the introduction of OpenAI’s ChatGPT, there has been a heavy push toward ensuring various services are safe and welcoming for everyone to use. OpenAI has put safety warnings on its ChatGPT product during sign-up and sporadic updates, warning of unintentional slipups in language. Meanwhile, various chatbot spinoffs have been fine to allow swearing and offensive language to a point.

Additionally, actual bad actors quickly began to take advantage of the demand for ChatGPT, Google Bard, and other chatbots before they became wildly available. Many campaigns advertised the products on social media with malware attached to image links, among other attacks. This showed quickly that AI was the next frontier of cybercrime.

The NTU research team contacted the AI chatbot service providers involved in the study about its proof-of-concept data, showing that jailbreaking for chatbots is real. The team will also present their findings at the Network and Distributed System Security Symposium in San Diego in February.

Fionna Agomuoh
Fionna Agomuoh is a Computing Writer at Digital Trends. She covers a range of topics in the computing space, including…
5 AI apps with deep research features to rival ChatGPT
Deep Research option for ChatGPT.

Artificial intelligence brands are in fierce competition, and their next steps are to make AI tools smarter by allowing them to execute deep search functions that can provide expert-level results and analyze larger amounts of information in a shorter time. Several companies have announced deep research features in recent weeks and months that excel in areas such as finance, science, marketing, and academics. Research that would have taken a person weeks or months can be achieved in a fraction of the time, with a properly detailed prompt. 

Deep research features are considered AI agents that can work independently and will allow you to make a query and let the AI process for several minutes while it generates the information and returns when it is finished to display the results. They are considered the first steps toward the concept of artificial general intelligence (AGI), which some define as a model that can process a query based on novel data that it has not been trained on, and it can produce unique content. However, we’re not quite there yet, and the main premise of deep research tools is processing large amounts of data and making it easier to understand.

Read more
Meta’s new AI app lets you share your favorite prompts with friends
Meta AI WhatsApp widget.

Meta has been playing the AI game for a while now, but unlike ChatGPT, its models are usually integrated into existing platforms rather than standalone apps. That trend ends today -- the company has launched the Meta AI app and it appears to do everything ChatGPT does and more.

Powered by the latest Llama 4 model, the app is designed to "get to know you" using the conversations you have and information from your public Meta profiles. It's designed to work primarily with voice, and Meta says it has improved responses to feel more personal and conversational. There's experimental voice tech included too, which you can toggle on and off to test -- the difference is that apparently, full-duplex speech technology generates audio directly, rather than reading written responses.

Read more
OpenAI CEO admits ChatGPT’s personality is ‘too annoying’
Deep Research option for ChatGPT.

Have you noticed that ChatGPT has gotten a little personal lately? It's not just you. OpenAI's CEO, Sam Altman, admitted last night that the last couple of updates to GPT-4o have affected the chatbot's personality, and not in a good way.

If you use ChatGPT often enough, you might have noticed a shift in its behavior lately. Part of it might be down to its memory, as in my experience, the chatbot addresses you differently when it doesn't rely on past chats to guide the way you'd (potentially) want it to respond. However, part of it is just that somewhere along the way, OpenAI has made ChatGPT a so-called "yes man" -- a tool that agrees with you instead of challenging you, and sometimes, the outcome can be a touch obnoxious.

Read more