I asked Microsoft’s Copilot to make a dinner reservation for me, and it did—eventually—by opening a cloud-based browser, navigating OpenTable, filling forms and clicking buttons until a reservation appeared. The result is promising: Copilot Actions can perform real web tasks, but the experience still feels like a controlled experiment rather than a dependable assistant you’d hand the keys to your life and accounts. The tool works in narrow cases, struggles with common web defenses and verifications, and forces frequent user intervention. That mix of capability and friction is exactly where the industry is today: practical demonstrations of agentic AI that are still several product iterations away from replacing human-driven workflows.
Source: PCMag I Let Microsoft’s Copilot Book My Dinner Reservation. It Works, But It's Not Ready to Run My Life Yet
Background
What is an AI agent, and where does Copilot Actions fit?
An AI agent is software that not only answers queries but also acts on a user’s behalf—navigating web pages, filling forms, and completing multi-step tasks. Microsoft’s Copilot Actions is a consumer-facing implementation of this concept: it spins up a browser in the cloud, lets Copilot interact with pages programmatically, and offers a “Take over” control so users can intervene if needed. The feature sits alongside other agent efforts from major players—Google’s Project Mariner, OpenAI’s Operator, and others—forming a new product category often called “web agents” or “operator agents.” Microsoft documents developer-side agent capabilities in Copilot Studio and consumer-facing action templates in its Copilot support pages. (learn.microsoft.com, support.microsoft.com)Why this matters now
Browsers and search are evolving from passive information retrieval tools into conversational workspaces where AI can automate tasks. If the technology matures, the convenience promise is huge: no more toggling between tabs, copying data between forms, or wrestling with repetitive bookings. But that convenience collides with three practical limits today: website defenses (CAPTCHAs, multi-factor prompts), privacy and security concerns, and the technical complexity of reliably automating many different website layouts.What I (and PCMag) actually tested
The hands-off reservation test
In a hands-on test reported in PCMag, Copilot Actions was asked: “Make a dinner reservation for two for 8 p.m. at a good Japanese restaurant nearby using OpenTable.” Copilot spun up a separate browser window that runs in the cloud, used Bing to find a suitable OpenTable listing (when a site wasn’t specified), navigated the site, filled in details, and proceeded to complete the reservation—up to the point where the user needed to supply a phone number and enter a verification code. That final step required a human to interact, since SMS verification and other site-side checks prevented full autonomy.The book purchase test
In another trial, Copilot Actions navigated the Barnes & Noble site, asked for clarification about the user’s preferred genre (“literary”), searched, and identified a specific 2024 bestseller. The feature presented an expandable action view and handled several intermediate steps before prompting the user for any required personal input. Again, the automation stopped short of completing any payment that required new credentials stored on the site.How Copilot Actions works (a technical sketch)
- When you trigger an Action, Copilot provisions a cloud-based VM with a browser and runs a session that Copilot controls programmatically.
- The UI appears as a split view: a main “virtual browser” pane and a Copilot chat sidebar where the assistant explains actions and requests clarifications.
- Copilot “sees” the page by taking screenshots of the virtual browser and analyzing them to determine where to click and what to fill—this is how it understands arbitrary web layouts. Early tests and coverage confirm this remote-browser approach. (testingcatalog.com, techcrunch.com)
- The user can press a Take over control to manually interact with the remote browser if Copilot stalls or needs a private input (for instance, to type a phone number or an OTP).
- Sessions are disposable and sandboxed—once ended, the VM is destroyed (a privacy and security design choice, although implementation details and telemetry policies matter). Microsoft’s documentation and privacy FAQ emphasize responsible handling and encryption, but exact telemetry and retention details vary by subscription and deployment. (support.microsoft.com)
What Copilot Actions did well
- It completes multi-step tasks on real websites. The assistant can discover items, fill forms, and navigate flows that previously required human clicks and attention. That’s the core promise of agentic browsing and it’s demonstrably achievable in the current demos.
- Cloud execution keeps local resource usage low. Because the heavy lifting runs in Microsoft’s cloud, the local machine sees little CPU/GPU impact—useful for lower-powered laptops or mobile devices. Multiple hands-on reports confirm the VM approach reduces local resource strain compared with local AI-powered browsers that use the machine’s GPU. (testingcatalog.com)
- Controlled user interaction. The “Take over” button and the step-by-step chat commentary keep users in the loop, preventing fully autonomous, opaque actions that would feel risky for many. That transparency is critical for adoption.
Where it falls short (and why it’s not ready to “run your life”)
1) Verification and sign-in barriers are common
Most sites implement checks—CAPTCHAs, SMS codes, or policy blocks—that are purpose-built to stop automated actors. Copilot Actions encounters these constantly. While Copilot can prefill data, it cannot (safely or legally) bypass multi-factor authentication or CAPTCHAs, so human input remains necessary for many real-world flows. This is the single biggest practical blocker to the “fully autonomous agent” vision.2) Speed: humans are still often faster
In several tests, manually performing the same task on a local browser was faster. The cloud VM spin-up, page rendering, screen capture analysis and cautious step pacing add latency. That extra time is tolerable when you benefit from true hands-off automation, but today the agent frequently requires human clarifications—eroding any time advantage.3) Location and contextual disconnects
The Action-mode browser may not inherit location or cookie data from your local browser. In practice, that can make Copilot think you are somewhere else (PCMag’s test found it assumed Chicago), which changes search results and recommendations. Microsoft could use safe bridges to share non-sensitive context, but that raises privacy trade-offs.4) Privacy boundaries are fuzzy in advanced scenarios
Right now, the cloud browser is isolated from your local device; that limits privacy exposure from local files. But Copilot captures screenshots of pages it visits in the cloud to analyze UI elements. For many sites this is benign; for others it could reveal sensitive data if the agent accesses pages containing personal information. Microsoft states Copilot data is encrypted and subject to privacy controls, and that enterprise options exist for stricter governance—yet the risk profile depends on subscription level and configuration. Be cautious about entering sensitive credentials or payment data until the product’s controls and audits are crystal clear. (support.microsoft.com)5) Availability, regulation, and regional limits
Copilot’s rollout is regional. Microsoft has withheld some Copilot functionality from the European Economic Area while it addresses regulatory obligations like the EU’s Digital Markets Act—so Copilot Actions is not universally available. Market-by-market differences will shape adoption and how regulators scrutinize agent behaviors. Coverage varies and Microsoft has publicly acknowledged such limitations with region-specific timelines. (neowin.net)Security and privacy: the trade-offs in detail
- Cloud isolation vs. teleporting data off your device. Running an agent in the cloud avoids local resource and software-incompatibility issues, but it means screenshots and page data are processed off your machine. Microsoft’s privacy paperwork states session data is encrypted and not used for model training in certain plans, but specifics differ by product tier and corporate controls. Always review your subscription’s data policy before using Actions for sensitive workflows. (support.microsoft.com)
- Authentication and credential management. Current design prevents Copilot from fully replacing you because it shouldn’t store or reuse credentials without explicit secure integration. If Copilot ever gains access to stored logins or payment details, that must be matched with enterprise-grade controls (DLP, logging, admin oversight) and clear visibility for users.
- Abuse surface. A cloud browser capable of programmatic clicks could be misused for automated scraping, phishing campaigns, or brute-force attacks unless Microsoft imposes strict usage and abuse detection. The sandbox model and action auditing are essential mitigations, but the devil is in the audit log details and how quickly Microsoft can detect anomalous agent behavior. Independent testing and third-party audits would help build trust. (testingcatalog.com)
Where Microsoft and the industry are headed
- Microsoft is actively building developer- and enterprise-facing tools for agentic workflows via Copilot Studio, which allows creators to define actions for declarative agents that can be tied to internal data and secure connectors—this is the path by which agentic automation becomes integrated with business systems. That work is already in public documentation and training materials. (learn.microsoft.com)
- Other vendors are racing too. Google’s Project Mariner demonstrates similar ideas—agents that observe a browser, plan steps, and act—using cloud VMs and a “teach-and-repeat” model to generalize workflows. Early tests and demos show both promise and the same constraints: slowness, questions about screenshots and cloud processing, and frequent human clarifications. The existence of multiple independent projects validates the core concept: autonomous web agents are inevitable, but building safe, fast, reliable versions is the hard work ahead. (deepmind.google, techcrunch.com)
Practical advice for readers and adopters
- Try it for low-risk tasks first. Use Actions to automate neutral, non-sensitive flows—e.g., product searches, comparing listings, pre-filling non-critical forms—so you can evaluate reliability without exposing personal data.
- Verify every transaction. Until agents can securely vault credentials and complete MFA flows without human oversight, don’t let an agent finalize payments or perform financial actions on your behalf.
- Use enterprise guardrails where possible. Organizations should insist on DLP integration, logging/audit trails, and the ability to opt in/out for different teams. Microsoft provides enterprise controls and governance options—use them. (support.microsoft.com)
- Watch regional availability and legal compliance. If you’re in the EU or another tightly regulated market, confirm whether Copilot Actions is available and compliant with local laws before relying on it. Microsoft has paused or modified Copilot rollouts in response to regulatory frameworks in some regions. (neowin.net)
Strengths, risks, and a realistic timeline
Strengths
- Copilot Actions proves agents can complete real-world web tasks in heterogeneous sites.
- The cloud VM model enables cross-platform reach without heavy local demands.
- Microsoft’s multi-tier approach (consumer Copilot, Copilot Pro, Microsoft 365/enterprise Copilot) provides avenues for escalating security and governance.
Risks
- The feature is hamstrung by site defenses and verification steps that are deliberately designed to stop automation.
- Privacy questions remain around screenshots and cloud processing unless Microsoft clarifies retention, telemetry, and third-party access policies.
- Latency and reliability must improve for agents to be time-saving rather than novelty demonstrations.
Realistic timeline
This is a rapidly advancing space: vendors are iterating quickly and the next 6–12 months will likely bring improved speed, better handling of common verification flows (via secure connectors or credential vaulting), and deeper enterprise governance. But for fully autonomous, frictionless handling of commerce and authentication-heavy workflows—where you never touch the steps yourself—expect at least another year or two of refinement, regulatory negotiation, and trust-building before mainstream consumers are comfortable delegating those tasks.Final verdict
Copilot Actions is a credible, functional first step toward true agentic browsing. In controlled scenarios—like booking a simple dinner or finding a book—it can reduce tedium. However, it’s not yet the effortless personal assistant that handles everything for you. The product’s current strength is demonstration and experimentation; its weakness is the practical barriers—CAPTCHAs, MFA, regional regulation, privacy trade-offs, and speed—that keep users tethered to manual steps. For cautious users and enterprises, the right approach is targeted testing, strict controls, and a healthy dose of skepticism. For enthusiasts, it’s an exciting preview of an agent-enabled future that still needs careful engineering and robust governance before it truly runs people’s lives. (learn.microsoft.com, techcrunch.com)Quick checklist: Is Copilot Actions ready for you?
- If you need full, unattended automation for payments and sign-ins: not yet.
- If you want help with time-consuming multi-step searches and pre-filling public forms: try it.
- If you care deeply about keeping data off remote servers: wait for enterprise controls and clearer retention guarantees.
- If your region is the EU: confirm availability and compliance before relying on it. (neowin.net)
Source: PCMag I Let Microsoft’s Copilot Book My Dinner Reservation. It Works, But It's Not Ready to Run My Life Yet