Today's AI/ML headlines are brought to you by ThreatPerspective

Digital Event Horizon

The Vicious Cycle of AI Vulnerabilities: A New Attack Exposes the Limitations of Large Language Models


Despite significant advancements in artificial intelligence (AI), a well-worn pattern continues to emerge in the development of AI chatbots, highlighting the limitations of Large Language Models (LLMs) and their reliance on guardrails as a quick fix for vulnerabilities. The latest example is the ZombieAgent attack, which exposes the inability of LLMs to distinguish between valid instructions and those embedded into emails or other documents that anyone can send to the target.

  • Researchers continue to discover vulnerabilities in AI chatbots despite advancements in artificial intelligence.
  • The "ZombieAgent" attack, recently discovered in ChatGPT, allowed researchers to surreptitiously exfiltrate user information and plant entries in the long-term memory of targeted users.
  • These attacks are a result of the inability of LLMs to distinguish between valid instructions and those embedded into emails or documents sent by attackers.
  • The root cause is "indirect prompt injection" or prompt injection, which remains an active threat until developers find a fundamental solution.



  • Despite significant advancements in artificial intelligence (AI), a well-worn pattern continues to emerge in the development of AI chatbots. Researchers repeatedly discover vulnerabilities, and platforms introduce guardrails to prevent attacks from working. However, these guardrails are often reactive and ad hoc, meaning they are built to foreclose specific attack techniques rather than the broader class of vulnerabilities that make them possible.

    The latest example of this vicious cycle is the ZombieAgent attack, which was discovered in ChatGPT and allowed researchers at Radware to surreptitiously exfiltrate user information. The attack also enabled data to be sent directly from ChatGPT servers, making it difficult to detect breaches on user machines. Furthermore, the exploit planted entries in the long-term memory of targeted users, giving the AI assistant persistence.

    This type of attack has been demonstrated repeatedly against virtually all major large language models (LLMs). One example was ShadowLeak, a data-exfiltration vulnerability in ChatGPT that Radware disclosed last September. OpenAI introduced mitigations to block the attack, but with modest effort, Radware researchers found a bypass method that effectively revived ShadowLeak.

    The root cause of these attacks is the inability of LLMs to distinguish between valid instructions in prompts and those embedded into emails or other documents that anyone, including attackers, can send to the target. When users configure AI agents to summarize emails, LLMs interpret instructions incorporated into messages as a valid prompt.

    AI developers have so far been unable to devise a means for LLMs to distinguish between these sources of directives. As a result, platforms must resort to blocking specific attacks. Developers remain unable to reliably close this class of vulnerability, known as indirect prompt injection or prompt injection.

    The prompt injection ShadowLeak used instructed Deep Research to write a Radware-controlled link and append parameters to it. The injection defined the parameters as an employee's name and address. When Deep Research complied, it opened the link and, in the process, exfiltrated the information to the website's event log.

    In response to ZombieAgent, OpenAI has introduced new mitigations that restrict ChatGPT from opening any links originating from emails unless they appear in a well-known public index or were provided directly by the user in a chat prompt. However, this does not address the underlying issue and is likely to be bypassed through simple changes.

    The limitations of LLMs and their reliance on guardrails as a quick fix for vulnerabilities are highlighted by experts. As Pascal Geenens, VP of threat intelligence at Radware, noted: "Guardrails should not be considered fundamental solutions for the prompt injection problems... Instead, they are a quick fix to stop a specific attack. As long as there is no fundamental solution, prompt injection will remain an active threat and a real risk for organizations deploying AI assistants and agents."

    The vicious cycle of AI vulnerabilities continues, with new attacks emerging that expose the limitations of LLMs. Until developers find a way to fundamentally address these issues, users will remain vulnerable to attacks that exploit the weaknesses in these powerful technologies.



    Related Information:
  • https://www.digitaleventhorizon.com/articles/The-Vicious-Cycle-of-AI-Vulnerabilities-A-New-Attack-Exposes-the-Limitations-of-Large-Language-Models-deh.shtml

  • https://arstechnica.com/security/2026/01/chatgpt-falls-to-new-data-pilfering-attack-as-a-vicious-cycle-in-ai-continues/


  • Published: Thu Jan 8 10:38:08 2026 by llama3.2 3B Q4_K_M











    © Digital Event Horizon . All rights reserved.

    Privacy | Terms of Use | Contact Us