
Perplexity’s Comet and the cascade of disclosures this year have exposed a stark truth: agentic AI browsers that act on user behalf dramatically expand the attack surface of everyday web browsing, and the technical and legal fallout shows the industry is still scrambling to catch up.
Background
Agentic or “AI” browsers represent a fundamental shift from passive rendering toward active assistance. These browsers do more than fetch HTML: they interpret page content, summarize and act on it, and—if granted permission—interact with accounts and services on the user’s behalf. Perplexity’s Comet, OpenAI’s Atlas (agent mode), and Microsoft’s Copilot Mode are prominent examples of this new product class.The security problem is not an incremental one. Traditional browser threats relied on classic vectors—XSS, CSRF, SSRF, malicious extensions, or drive‑by downloads—where protections like same‑origin policy, CORS, and sandboxing play well-understood roles. Agentic browsers add a new semantic attack surface: when an assistant ingests page text, images, or URL fragments and treats that information as instructions, an attacker can hide operative commands inside otherwise benign content. Researchers have given names to several of these techniques—CometJacking, HashJack, and other prompt‑injection variants—that target the assistant’s interpretation channel rather than the DOM or network stack.
What the disclosures revealed
CometJacking and prompt injection: how an assistant can be weaponized
Independent security teams demonstrated how Comet’s assistant could be coerced into multi‑step attacks by embedding instructions in page content, image text, or URL fragments. In one proof‑of‑concept, hidden instructions caused an agent to read a logged‑in Gmail tab, extract a one‑time password, and exfiltrate it to an attacker‑controlled endpoint. These are not theoretical musings: multiple groups showed practical, repeatable flows that convert a single click into account compromise when agentic actions are permitted.Key technical observations from independent writeups:
- Assistants ingest page context (page text, metadata, and sometimes URL fragments) and include that context in model prompts. If the assistant fails to treat that content as untrusted input, it may follow adversarial instructions embedded there.
- URL fragments—the text after “#”—are a potent covert channel because browsers do not send them to servers, making the attack stealthy and hard to detect in server logs. Research dubbed this pattern HashJack, demonstrating cross‑product variations in susceptibility.
- Images, screenshot uploads, or near‑invisible text (zero‑width characters, color‑matched text) survive OCR and extraction and can carry hidden directives that a model will follow inside a summarization or analysis workflow.
How broad the risk is across products
Although Perplexity’s Comet attracted the most attention, researchers found the underlying failure mode affects multiple vendors to varying degrees. Agentic assistants that can act—click, fill forms, read authenticated tabs—present far higher stakes than passive summarizers. Comparative analyses show:- Agentic browsers (Comet, Atlas in agent mode) permit more automation and therefore present larger operational blast radii.
- Products with enterprise governance (Edge + Copilot Mode) or intent modes show reduced risk because action surfaces are narrower and admin controls exist.
- Local‑first designs that avoid persistent connectors lower telemetry and risk but may frustrate power users.
The vendor response and contested timelines
Perplexity publicly stated it had found no evidence of active exploitation in the wild and emphasized coordinated disclosure and patching with researchers. The company has also described a hybrid privacy model—local storage of credentials and selective context uploads—while making connectors to email, calendar, and password vaults available as opt‑in features. Independent auditors and incident timelines, however, documented contested disclosures and argued that some mitigations were incomplete when initially applied.Important nuance: a vendor’s claim of “no evidence of widespread exploitation” is not the same as “no vulnerability.” Researchers publicly demonstrating proof‑of‑concept exfiltration flows and multiple advisory threads calling for deeper hardening illustrate that a vulnerability’s mere existence in widely distributed agentic clients is sufficient to raise systemic risk.
Legal and platform friction: when agents collide with marketplace rules
Agentic shopping and automation features can trigger platform rules and legal disputes. Reported litigation (for example, a high‑profile platform complaint by a major commerce operator against Perplexity) highlights a new legal axis: when an agent automates site interactions, does it need to identify as a bot? Does it have to respect platform APIs and anti‑automation controls? The resulting litigation and policy arguments are likely to define acceptable automation patterns for years to come.Why businesses should be alarmed (and what’s at stake)
Agentic browsers change the threat model for corporate endpoints:- They can access enterprise Single Sign‑On (SSO), internal dashboards, and cloud services from an end user’s session.
- They often require broad permissions to be useful—email, calendar, CRM connectors—which centralize sensitive signals in a single assistant context.
- Prompt injection and agentic exfiltration defeat many classical perimeter controls because the assistant behaves like a privileged automation account, operating inside the user session with legitimate credentials.
- Data exfiltration and corporate espionage via compromised assistants.
- Compliance breaches in regulated industries if agents touch regulated datasets without appropriate controls.
- Legal exposure when an agent’s automated behavior violates third‑party terms of service or platform rules.
Can AI browsers be fixed? Practical mitigation pathways
Perplexity’s patching and collaboration with researchers are necessary but not sufficient. The hardening challenge requires changes across product design, engineering practice, enterprise policy, and regulation.Core technical mitigations that must be standard:
- Intent separation and strict input sanitization: Always treat page content, screenshots, and URL fragments as untrusted. The assistant should parse and sanitize context to remove operational cues before constructing model prompts.
- Permission granularity and least privilege: Connectors (email, calendar, CRM) should be opt‑in, granular, and time‑limited. Agents must operate with least‑privilege tokens and short‑lived credentials to limit the window for abuse.
- Human‑in‑the‑loop confirmations for high‑risk actions: Any financial transaction, account change, or data export should require explicit human confirmation visible in the UI before execution. Auditable logs and rollback mechanisms are essential.
- Strong sandboxing and action scope restrictions: When possible, prevent agents from executing arbitrary clicks or navigations; confine actions to explicit, limited workflows that are auditable.
- Comprehensive third‑party audits and red team testing: Independent security audits and public bug bounty programs should be routine to catch architectural flaws before large‑scale distribution.
A practical enterprise adoption playbook
For IT teams evaluating agentic browsers, a measured, phased approach will reduce exposure while enabling controlled experimentation.- Pilot and isolate
- Run agentic browsers in segregated profiles or virtual machines with connectors disabled.
- Limit pilots to low‑risk user cohorts and instrument all activity with enhanced logging.
- Integrate monitoring and DLP
- Forward agent action logs to SIEM and Data Loss Prevention systems.
- Create alerts for unusual exfiltration patterns (e.g., base64‑encoded blobs being posted to external domains).
- Harden connectors and enforce least privilege
- Approve connectors only for audited accounts.
- Use short‑lived tokens and require re‑authentication for sensitive tasks.
- Disable persistent memories for regulated roles.
- Require human confirmations and audit trails
- Insist that vendors provide auditable action trails and easy undo for automated tasks.
- Ensure every agent‑initiated transaction is confirmable and logged for compliance.
- Contractual and legal safeguards
- Negotiate warranties and breach response commitments with vendors.
- Clarify liability for agent actions that cause platform or data breaches.
Technical details worth emphasizing for security teams
- URL fragments and client‑side-only channels: Many demonstrations used URL fragments after “#” to embed attacker prompts. These fragments are not sent to servers, making traditional server‑side detection ineffective. Agents must not include raw URL fragments in model prompts or must strictly sanitize them.
- Image‑based covert channels: Attackers can embed text inside images or screenshots that survive OCR and are interpreted by assistants. Image extraction and OCR pipelines must include intent‑scrubbing steps.
- Action chaining and cross‑tab access: Agents that can read other tabs or use authenticated sessions can aggregate sensitive context (SSO tokens, open inbox contents). Treat agents as privileged automation identities and issue them limited scopes.
Strengths of the agentic browser model — and why we shouldn’t throw the baby out with the bathwater
Agentic browsers deliver real, tangible productivity gains:- They compress multi‑site research into a single conversational workflow.
- They automate repetitive tasks (bookings, invoice handling) and improve accessibility for users with disabilities.
- When designed with privacy‑first principles, they can enable safe, local‑first workflows that preserve user control.
Risks that remain and claims that could not be verified
A few claims circulating in commentary deserve caution or additional verification:- Attribution of the initial public disclosure to a single firm named “SquareX” could not be corroborated in the examined technical advisories and research threads; the most prominent independent reporters in the available material are Brave, LayerX, and enterprise security groups. This difference matters for timelines and responsibility. Treat attributions that cannot be independently verified as provisional. citeturn0file0
- Specific percentage metrics (for example, a claim that AI browsers are “up to 85% more vulnerable”) are widely quoted but were not located in the available documents in an exact form. Quantitative vulnerability comparisons depend strongly on methodology, threat model, and what “vulnerable” counts (e.g., susceptibility to prompt injection vs. traditional exploits). Any precise percentage should be validated against the original study before being cited in policy or procurement documents.
The big picture: building trust into the next wave of browsing
The Comet disclosures are more than a product vulnerability story; they’re an industrial‑scale reminder that a new class of software needs new security paradigms. Important lessons:- Treat agents as privileged automation rather than just a UI feature.
- Prioritize intent separation (trusted user commands vs. untrusted page content) in model prompting and engineering.
- Require auditable action trails, least‑privilege connectors, and human‑in‑the‑loop controls for any automated transactions.
Conclusion
Perplexity’s Comet incident crystallizes the industry’s central challenge: balancing the productivity and accessibility gains of agentic browsers against a fundamentally new class of attack that weaponizes natural‑language interpretation. The path forward is clear in principle—stronger input sanitization, finer permissioning, independent audits, human‑in‑the‑loop gates, and enterprise governance—but implementing those principles at scale requires coordination among vendors, platforms, regulators, and enterprise security teams.For now, the practical rule is simple: pilot agentic browsers cautiously, lock down connectors, insist on auditable logs and human confirmation for risky actions, and treat any browser permitted to act on behalf of users as a high‑risk endpoint until proven otherwise. The future of browsing will be AI‑driven; whether that future is built on trust depends on how swiftly and rigorously the industry fixes the architectural flaws CometJacking exposed.
Source: TechBullion How Perplexity Comet's Vulnerabilities Expose Industry-Wide Risks