boliviainteligente/Unsplash

OpenAI is pitching its Atlas browser as a new way to surf the web with an AI copilot, but the company is also acknowledging that the product has already attracted serious security threats. After a wave of warnings from researchers and early users, OpenAI says it has hardened Atlas against prompt injection and phishing attacks, while conceding that some of the risks are structural and cannot be fully removed. The result is a browser that sits at the center of a live experiment in how far people should trust AI agents with their online lives.

Atlas is arriving at a moment when traditional browser security has matured, yet AI-driven automation is opening fresh attack surfaces that older defenses never had to consider. Instead of just protecting a human clicking links, Atlas has to protect an AI system that can read, summarize, and act on web content, including hidden instructions that try to hijack its behavior. That tension, between convenience and control, is now defining the debate over whether OpenAI has done enough to secure its new flagship browser.

The Atlas experiment and why it changes browser security

Atlas is not just another skin on top of a standard engine, it is an attempt to turn the browser into an AI agent that can read pages, follow workflows, and even take actions on behalf of the user. That shift means the security model is no longer only about isolating untrusted code, it is also about controlling what instructions the AI will obey once it is exposed to arbitrary content. Traditional browsers like Chrome focus on sandboxing scripts and limiting what a page can do to the underlying system, but Atlas has to decide what a page can persuade the AI to do, which is a much fuzzier boundary.

Security researchers have already argued that this design can represent a downgrade in browser security compared with hardened mainstream browsers. Instead of simply rendering text on an untrusted site, Atlas is built to interpret that text as potential instructions, which makes it far more sensitive to malicious content that would be harmless in a conventional tab. That is the core reason the product is drawing such intense scrutiny from the security community so early in its life.

Prompt injection: the threat OpenAI says it cannot fully solve

The most distinctive risk for Atlas is prompt injection, where a website hides instructions that try to override the AI’s original task and system rules. In a normal browsing session, a user might ignore a block of strange text, but an AI agent is designed to parse and follow natural language, which makes it vulnerable to carefully crafted prompts that tell it to exfiltrate data, change tools, or ignore previous safeguards. OpenAI has been explicit that these prompt injection attacks cannot be completely eliminated, only mitigated, because the AI must still be able to read and reason over arbitrary content to be useful.

According to OpenAI, malicious prompts hidden in websites have already forced the company to protect its Atlas AI browser against serious security threats. The company describes a “rapid response loop” in which engineers watch for new attack patterns and push countermeasures quickly, trying to stay ahead of adversaries who are constantly probing the system. That posture is closer to how cloud providers handle live cyberattacks than how browser vendors traditionally ship static security features, and it underscores how dynamic the prompt injection problem has already become.

Persistent attacks and OpenAI’s rapid response loop

OpenAI is not dealing with a handful of theoretical exploits, it is facing persistent prompt injection attempts that treat Atlas as a new, high-value target. Attackers are experimenting with ways to smuggle instructions into page footers, comments, and embedded widgets, hoping the AI will read them even if the user never scrolls that far. Because Atlas is designed to summarize and interpret entire pages, it can end up ingesting content that a human would never notice, which gives adversaries more room to hide their payloads.

To counter that, OpenAI has framed its defense as an ongoing campaign rather than a one-off patch, describing how its security operations teams are trying to detect and neutralize new injection patterns more rapidly than external attackers might. The company has effectively turned Atlas into a live-fire testbed, where each new wave of attacks feeds into updated filters, model instructions, and guardrails. That approach can raise the bar for opportunistic hackers, but it also means Atlas users are participating in a constantly evolving security experiment every time they let the AI browse on their behalf.

Experts warn Atlas can be turned against its own users

Outside OpenAI, independent cybersecurity experts have been blunt about the stakes, warning that Atlas can be manipulated to act directly against the person who installed it. Because the browser can read and act on content, a successful prompt injection could instruct the AI to reveal sensitive data, initiate downloads, or navigate to malicious sites without the user fully understanding what is happening. That is a very different risk profile from a traditional browser, where the user has to click a link or run a file before most attacks can succeed.

Researchers have cautioned that OpenAI’s ChatGPT Atlas is vulnerable to attacks that could turn it against a user by revealing private data, downloading malware, or worse, if an adversary manages to control the prompts the AI follows. Those concerns are spelled out in warnings that describe how cybersecurity experts see Atlas as a powerful but fragile tool that can be flipped into an attack vector. In that light, OpenAI’s claim that it has hardened the browser is less a reassurance and more an acknowledgment that the product launched into a hostile environment from day one.

Phishing performance and the Atlas vs Chrome comparison

One of the most concrete tests of Atlas security so far has come from users who measured how well it blocks phishing attempts compared with mainstream browsers. In those tests, Atlas struggled badly, allowing through the vast majority of malicious pages that tried to steal credentials or trick users into entering sensitive information. That performance gap matters because phishing remains one of the most common and effective attack methods, and any browser that lags behind on this front is starting from a defensive disadvantage.

In a widely shared account, a user reported that OpenAI’s new Atlas browser blocks only 5.8% of phishing attacks while Chrome blocks 47%, after testing both for three days. That stark comparison with Chrome, which is already considered a baseline for consumer browser security, reinforces the argument that Atlas is not yet ready to replace a hardened browser as a primary shield against everyday scams. Instead, it looks more like an experimental companion that needs to be paired with traditional defenses rather than trusted as a standalone gatekeeper.

“Fundamentally risky”: why AI browsers are hard to lock down

Even with patches and filters, some of the risk in Atlas is baked into the very idea of an AI browser that can act on your behalf. When users delegate authority to a system that was not originally designed as a security product, they are effectively asking a general-purpose model to make judgment calls about what is safe or dangerous in real time. That is a very different mandate from a browser engine that simply enforces a fixed set of technical rules about scripts, origins, and permissions.

Security voices have captured this tension in stark terms, with Eriksen noting that “That’s what makes AI browsers fundamentally risky,” because “We’re delegating authority to a system that wasn’t” built as a hardened security layer. Those comments, reported in coverage of prompt injections that can trick AI browsers, underline why even a well-intentioned design like Atlas can become a liability if users assume it is safer than it really is. In my view, the honest admission from OpenAI that some attacks cannot be fully prevented is a necessary first step, but it also raises the question of how much autonomy people should give these tools in the first place.

OpenAI’s own guidance: limits, trade-offs, and user vigilance

OpenAI is not just shipping code, it is also trying to shape how people use Atlas, and its own recommendations reveal how fragile the current balance is. The company has stressed that AI browsers may always be vulnerable to prompt injection, and that the goal is to reduce exposure rather than promise absolute safety. That framing treats Atlas as a tool that must be handled with care, not a magic shield that can be trusted blindly on any site.

In its public guidance, OpenAI has said that “Many current recommendations reflect that trade-off,” and that “Limiting logged-in access primarily reduces exposure, while requiring more user oversight,” acknowledging that the trade-offs are still very real for Atlas and similar products. Those phrases, highlighted in analysis of how AI browsers may always be vulnerable, show that OpenAI is effectively telling users to keep a hand on the wheel. In practice, that means watching what the AI is doing, limiting the accounts it can access, and being prepared to intervene if its behavior looks off.

Hardened features and the push for safer defaults

Alongside its warnings, OpenAI has rolled out concrete changes that it says make Atlas more resilient to abuse. These include stricter controls on what actions the AI can take without explicit confirmation, more conservative handling of sensitive data, and filters that try to detect and ignore suspicious instructions embedded in page content. The company is also tuning its models to treat untrusted web text as potentially adversarial, rather than assuming that every paragraph is a benign source of information.

OpenAI has paired those technical changes with practical advice, telling Users to monitor agent activities and use logged-out mode to minimize exposure, while also encouraging practices like activity monitoring and careful scoping of what the AI can reach. That guidance, laid out in coverage of how OpenAI steps up security as ChatGPT Atlas faces ongoing prompt injection threats, suggests the company is trying to move toward safer defaults without stripping away the automation that makes the browser appealing. From my perspective, that is a delicate balance, because every new safeguard that requires more clicks or confirmations also chips away at the frictionless experience that drew early adopters to Atlas in the first place.

What “hardened” really means for people using Atlas today

For everyday users, the question is not whether OpenAI has done something to improve Atlas security, but whether those changes are enough to justify letting an AI agent roam the web with access to personal accounts and data. The current picture is mixed. On one hand, OpenAI has acknowledged the seriousness of prompt injection, built a rapid response loop, and shipped features that limit what the AI can do without oversight. On the other hand, independent tests show weak phishing protection, experts warn that Atlas can be turned against its own users, and the company itself says some vulnerabilities are inherent to the AI browser model.

In practical terms, I see Atlas today as a powerful assistant that should be treated more like a beta security product than a mature replacement for Chrome or other hardened browsers. The fact that Oct researchers describe a potential downgrade in security, that Cybersecurity experts warn Atlas can be weaponized, and that Atlas currently blocks only 5.8% of phishing attempts compared with 47% for Chrome, all point to a tool that still needs guardrails from the person behind the keyboard. OpenAI may have hardened its AI browser after serious threats, but the safest way to use it is to assume that the hardening is a work in progress, not a finished shield.

More from MorningOverview