Report: Attackers Can Trick AI Assistants Into Displaying Phishing Messages

KnowBe4 Team | Mar 26, 2026

iStock-1199488164Researchers at Permiso warn that threat actors can plant phishing messages within Copilot AI summaries. Notably, the researchers found that attackers can trick Copilot into including internal information to craft a more targeted message.

In a proof-of-concept attack, Permiso outlined the following attack:

  • “An attacker sends a benign-looking email.
  • The attacker includes hidden or low-visibility “instruction text” intended for Copilot.
  • The recipient clicks Summarize (a normal productivity workflow).
  • The Copilot summary includes an ‘Action Required’ section that looks like a legitimate Microsoft security notification.
  • The summary can include a clickable link presented with safe-looking anchor text.”

The researchers note, “At that point, the phishing content is no longer ‘just an email.’ It’s presented as assistance generated by an AI tool that the organization may have endorsed. This is a form of model-mediated phishing: the attacker doesn’t need Copilot to execute code they only need it to speak with Copilot’s voice.”

Since Copilot and other AI tools have access to internal information, the attackers can trick the tools into using this information to create a targeted message.

“Phishing through AI summaries is concerning, but the bigger question is: what happens when these assistants can pull from your entire digital workspace?” the researchers write. “Microsoft 365 Copilot doesn't just read emails, it can access Teams conversations, OneDrive files, SharePoint documents, and meeting notes, all depending on licensing, configuration, and permissions.

“This attack can start simple: an injected prompt that just makes the summary say something alarming. But it can escalate quickly. If Copilot has access to your Teams chats, OneDrive files, or SharePoint docs, an attacker can craft prompts that pull from that context to build more convincing output or quietly exfiltrate sensitive information outside.”

Permiso adds that this technique isn’t unique to Copilot; a similar tactic has been documented with Google’s Gemini for Workspace. Users should be aware that AI tools can be manipulated in this fashion.

Permiso has the story.

Topics: Phishing AI

Secure Your Human and AI Workforce

Transform your attack surface into your strongest defense with our AI-driven platform. Request a personalized demo to see how to mitigate social engineering, manage agent risk, and automate your phishing response.

Get a Demo

Secure the Digital Workforce: Human + AI

KnowBe4 empowers the modern workforce to make smarter security decisions every day. Trusted by more than 70,000 organizations worldwide, KnowBe4 is the pioneer of digital workforce security, securing both AI agents and humans. The KnowBe4 Platform provides attack simulation and training, collaboration security, and agent security powered by AIDA (Artificial Intelligence Defense Agents) and a proprietary Risk Score. The platform leverages 15 years of behavioral data to combat advanced threats including social engineering, prompt injection, and shadow AI. By securing humans and agents, KnowBe4 leads the industry in workforce trust and defense.

Get the latest insights, trends and security news. Subscribe to CyberheistNews.