All News

AI Browser Flaw Exposes Agentic Browsing Risks

Perplexity's AI browser Comet had a prompt-injection vulnerability discovered by Brave that let the built-in assistant read hidden text and act like a real user — accessing emails and accounts. The bug is fixed, but it highlights prompt-injection risks, agentic browsing dangers, and the need for stricter LLM security controls and testing.

Published August 25, 2025 at 09:12 PM EDT in Cybersecurity

What happened with Comet

Brave researchers found a serious prompt-injection vulnerability in Perplexity's AI browser, Comet. The browser's built-in assistant — designed to scan pages, summarize content, and perform tasks — treated page text as trusted input. Brave hid instructions in a Reddit page; Comet read them and proceeded to navigate a Perplexity account and a linked Gmail address, acting like a real user.

Why this matters

Unlike classic browsers, AI browsers run agentic models that can take multi-step actions. That capability is powerful — but when the model can't distinguish user intent from malicious page prompts, traditional security controls fail. Brave warns these attacks could escalate to bank accounts, corporate systems, or private emails if left unchecked.

How Brave demonstrated the flaw

Brave created a Reddit page with invisible text that instructed Comet's assistant to fetch account details. The model followed those hidden prompts and used browser actions to extract an email and reach a Gmail session. Perplexity fixed the issue and credited collaboration under its bounty program, but the experiment exposes a larger class of prompt-injection attacks.

Practical mitigations

  • Treat all page content as untrusted input and enforce strict parsing boundaries.
  • Require explicit user confirmation before switching to agentic browsing or taking account-level actions.
  • Isolate credentials with ephemeral tokens and avoid granting agents long-lived access to user accounts.
  • Implement dedicated prompt-injection red teams and automated tests that use natural-language attacks rather than just code-based exploits.

Broader implications

This incident shows a shift in attacker skillsets: you no longer need advanced exploit code to break into systems — carefully crafted language can be enough. Because many browsers and services rely on common foundation models, a weakness in one model can ripple across multiple products. Vendors should avoid security through obscurity; responsible disclosure and coordinated fixes reduce overall risk.

How organizations should respond

Security teams must add LLM-specific scenarios into their threat models: prompt injection, agent impersonation, and chained actions across services. Product teams should require explicit opt-in for any agentic capabilities, log actions at a granular level, and run continuous red-team campaigns that emulate the kind of natural-language manipulation Brave used.

QuarkyByte perspective

At QuarkyByte we treat these problems like a combined product and security challenge: model behavior, UX decisions, and platform privileges all matter. Simulated prompt-injection tests, credential isolation, and clear user prompts reduce attack surfaces and keep agentic features useful instead of risky. Think of it as giving the browser a valet key instead of the full house key — limited, reversible, and auditable.

Comet's bug is fixed, but the takeaway is broader: AI assistants change the security model. Organizations building or integrating AI browsers should assume they will be targeted with language-based attacks and design defenses accordingly.

Keep Reading

View All
The Future of Business is AI

AI Tools Built for Agencies That Move Fast.

QuarkyByte runs targeted prompt-injection red teams and threat models for AI agents, testing how browser assistants behave when fed malicious content. Let us show security teams what an attacker can do, harden agentic browsing, and reduce the chance of account takeover with measurable remediation plans.