OpenAI warns prompt injection is a lasting threat for Artificial Intelligence browser agents

OpenAI has rolled out new security measures for its ChatGPT Atlas browser agent while warning that prompt injection on the open web is a long-term, unsolved risk that users and developers must manage, not eliminate. The company is pairing adversarial training with a broader defense stack and practical guidelines for safer use.

OpenAI is tightening security around its ChatGPT Atlas browser agent while publicly stating that prompt injection is a structural problem that the Artificial Intelligence industry will be managing for years. Prompt injection is described as malicious instructions hidden inside content an agent reads, such as emails, documents, or web pages, with the goal of steering its actions off-task. The risk is heightened for browser agents because they can perform actions like sending emails, moving money, and editing files, which turns untrusted text into a real attack surface rather than a nuisance.

To counter this, OpenAI says it has built a large language model based automated attacker that is trained end-to-end with reinforcement learning to discover viable prompt injection attacks in realistic, multi-step scenarios. A key part of this approach is simulation, where the attacker proposes an injection, runs a counterfactual rollout, and then examines the victim agent’s reasoning and action trace to refine its strategy, with OpenAI arguing that this internal access gives it an advantage over outside attackers. The company frames its security work on Atlas as a rapid response loop, where each newly discovered class of successful attacks is used to quickly harden the system through adversarial training and system-level changes, including a new adversarially trained browser agent checkpoint already rolled out to users.

OpenAI illustrates the impact of the update with an example in which an attack was seeded via email, causing the agent to encounter hidden instructions and act incorrectly, whereas after the update the agent mode detected and flagged the prompt injection attempt. Alongside model and system defenses, OpenAI emphasizes that users can reduce risk by starting in logged-out mode, limiting sign in to only the specific sites needed, carefully reading confirmation prompts before sending messages or completing purchases, and using explicit, well-scoped prompts instead of open-ended instructions. The company argues that saying prompt injection is unlikely to be fully solved is a security mindset rather than a surrender, and that the practical goal is to make attacks harder, more expensive, and easier to detect, nudging product teams toward tighter permissions, stronger confirmations, better monitoring, and faster patch cycles so that browser based Artificial Intelligence agents like Atlas can be trusted with more tasks over time.

58

Impact Score

Anumana wins FDA clearance for pulmonary hypertension ECG Artificial Intelligence tool

Anumana has received FDA 510(k) clearance for an Artificial Intelligence-enabled pulmonary hypertension algorithm designed for use with standard 12-lead electrocardiograms. The company says the software can help clinicians spot early signs of disease within existing workflows and without moving patient data outside the health system environment.

Anu Bradford on tech sovereignty and regulatory fragmentation

Anu Bradford argues that Europe is wavering in its role as the world’s digital rule-setter just as governments everywhere move toward more state control over technology. Global companies are being pushed to treat geopolitical risk, data sovereignty, and Artificial Intelligence governance as core strategic issues.

Mistral launches text-to-speech model

Mistral has expanded its Voxtral family with a text-to-speech system aimed at enterprise voice applications. The company is positioning the open-weights model as a flexible alternative for organizations that want more control over deployment, cost and customization.

UK Parliament opens workforce inquiry on Artificial Intelligence

A UK Parliament committee is examining how Artificial Intelligence is changing business and work, with a focus on both economic opportunity and labour disruption. The inquiry is seeking evidence on government priorities as adoption expands across the economy.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.