All corrections
X April 8, 2026 at 08:46 PM

x.com/MaxRovensky/status/2041910644375478397

1 correction found

1
Claim
it has 100% impenetrable safeguard around it by default
Correction

This is incorrect because security controls are not “100% impenetrable” by default, and OpenAI’s own safety documentation says untrusted text can still be used for prompt-injection attacks.

Full reasoning

The claim uses an absolute security guarantee (“100% impenetrable”) that is contradicted by both general cybersecurity guidance and OpenAI’s own documentation.

  • CISA states plainly that “No one system or network is completely impenetrable.” That directly contradicts the idea of a default, perfectly impenetrable safeguard.
  • OpenAI’s agent safety guidance says “Prompt injections are a common and dangerous type of attack” and explains that untrusted text or data can enter an AI system and try to override its instructions. That means plain text is not automatically protected by some perfect default barrier.
  • OpenAI’s own security writing on agentic systems likewise describes prompt injection as an “open challenge for agent security” and shows examples where malicious content can cause unintended actions.

So while a system may have mitigations, calling them “100% impenetrable” is factually wrong. Current AI safety/security guidance explicitly says these attacks remain possible and must be mitigated, not assumed impossible.

3 sources
Model: OPENAI_GPT_5 Prompt: v1.16.0