AI Malware:

Hype vs. Reality

Key Takeaways

  • Most “AI malware” observed so far falls into the AI malware Maturity Model (AIM3) Levels 1-3 (Experimenting through Optimizing), rather than fully automated campaigns.
  • AI is currently a force multiplier on existing attacker tradecraft, not a source of fundamentally new TTPs.
  • Many “first-ever AI malware” announcements are narrow research demos or PoCs with limited autonomy and unclear real-world impact.
  • Public reporting shows no confirmed examples of truly embedded, Bring-Your-Own-AI (BYOAI) malware running its own local model on victim hosts.
  • Defenders should prioritize monitoring abuse of legitimate AI services, hardening existing controls, and mapping threats to AIM3 levels rather than overreacting to sci-fi scenarios.

Introduction

Generative AI (GenAI) and large language models (LLMs) are being rapidly integrated into all aspects of our society, from communication to cybersecurity. Enterprises and vendors are already using GenAI and LLMs to augment their defenses. Attackers are also adopting LLMs, primarily as a force multiplier rather than the one-click super malware often implied in article headlines. From phishing lures to code generation and basic orchestration, GenAI is lowering the skill barrier and speeding up familiar workflows, not unleashing a brand new class of unstoppable, fully autonomous malware.

In practice, most AI malware activity today resides in the early stages of AI maturity, focused on AI-assisted coding & tradecraft, localized content, and experimental orchestration that still relies heavily on humans and traditional tools, even as vendors rush to brand narrow proof of concepts (PoCs) or niche incidents as the first-ever AI attacks. This post cuts through that hype by introducing a simple AI malware Maturity Model (AIM3) to define what truly counts as AI malware, map recent public claims to concrete maturity levels, and give defenders a realistic view of how AI is actually changing attacker economics today and how to prepare for more capable, orchestrated threats that are clearly on the way.

What counts as “AI malware”

We define AI malware as malicious or offensive software whose core development or runtime behavior is dependent on GenAI or LLMs. The malicious software can use LLMs to generate or select commands at runtime, inspect files or environment telemetry for action planning, or orchestrate parts of the attack chain without step-by-step human input. Five types of AI malware meet this description: LLM-Translated, LLM-Generated, LLM-Deployed, LLM-Driven, and LLM-Embedded.

Types of AI Malware

Measuring AI malware maturity with AIM3

Despite the numerous types of AI malware observed over the past few years, there is a lack of consensus regarding the current state of AI malware maturity. Maturity models exist for cybersecurity adoption, such as MITRE’s “AI Maturity Model”, but there's no attacker-centric model that directly measures the sophistication of AI-enabled malware itself and the operational risk it poses to organizations. To address this gap, Recorded Future proposes the five-level AIM3 that’s helping teams determine the sophistication level of AI threats and identify where to focus detection and governance efforts. AIM3 provides defenders with a means to distinguish genuine AI-driven threats from marketing noise.

Levels of Recorded Future’s AI Malware Maturity Model (AIM3)
Levels of Recorded Future’s AI Malware Maturity Model (AIM3)

Level 1 - Experimenting:

Attackers, Researchers, and Academia are creating prototypes, toy examples, and PoCs that leverage GenAI using rudimentary methods. At this stage, individuals are merely exploring the possibilities of GenAI and LLMs for malicious applications vs. operationalizing in a serious manner.

Level 2 - Adopting:

Threat actors incorporate GenAI into familiar workflows such as authoring phishing emails, researching targets, and developing code. While the core operational tasks remain conventional, there is an emphasis on automating and supporting low-order tasks without reinventing traditional tradecraft.

Level 3 - Optimizing:

Attackers are beginning to incorporate AI into their attack chains by leveraging GenAI on-host or via APIs to perform introspection, generate commands, and adapt code in near real-time. This is a shift in focus from bespoke GenAI use to treating GenAI as an integrated part of the attack chain.

Level 4 - Transforming:

AI-native offensive frameworks emerge at this level, combining multi-step planning and tool use with a human-in-the-loop (HITL) approach. These are the early, purposeful attempts at AI-first threat operations, utilizing agentic patterns rather than bolting GenAI onto legacy playbooks.

Level 5 - Scaling:

Threat actors are building agentic systems to manage campaigns end-to-end with no human oversight. Automated decision-making is implemented at scale for the planning, execution, and persistence stages of operations. This level of sophistication represents the upper bound of GenAI capabilities that current experimentation is moving toward.

With AIM3 defined, we can now examine what has actually been reported in the public domain and in research.

What we see in the wild

Timeline of AI Malware Reports Mapped to Maturity

2023 to 2024 (Early experimentation)

Early- to mid-2025 (AI-invoking malware appears)

Late-2025 (AI-driven implementations find success)

Hype vs. reality - what’s true, what’s exaggerated

With a few years of “AI malware” headlines behind us, some patterns are clear. Most public activity sits well below the fully autonomous, Hollywood-style threats often implied by marketing.

Where the activity really is (AIM3 Levels 1–3)

Mapped to AIM3, the picture is clear: the vast majority of examples sit at Levels 1–3 (Experimenting through Optimizing), with a single contested Level 4 case and no verified Level 5 activity. Families like PromptLock, PROMPTFLUX, and MalTerminal look more like PoC exercises rather than in-the-wild malware. The Anthropic case is the single and highly contested example of Level 4 (Transforming) maturity, but even this initial example was not fully autonomous.

Seen through that lens, a few things snap into focus: fully embedded models are still theoretical, many first-ever AI malware claims rest on thin or experimental evidence, and the real inflection point to watch is not brute autonomy but steadily improving AI-driven orchestration. In other words, the data tells us where the hype is running ahead of reality and where the next genuine shifts in attacker capability are likely to appear.

BYOAI and embedded models: still hypothetical in the wild

Most of the published implementations of AI malware call cloud or remote LLMs, not locally embedded models. Even in known malicious instances, such as Lamehug, families invoke remote services, like the HuggingFace API, for LLM calls at runtime. In some cases, the payload may use locally available LLM services that make these remote calls on its behalf; however, no observed sample currently features a Bring Your Own AI (BYOAI) capability. This is effectively our LLM-Embedded category: AI malware that ships its own model to run locally on the host.

The first-ever AI malware problem

Between July and November 2025, four “first-ever” reports emerged involving Malterminal, Lamehug, PromptLock, and the Anthropic cyberattack, which claimed different aspects of the AI malware landscape. After reporting, many of these claims were subject to scrutiny by the research community as academic, experimental, or underwhelming. Although these examples represent a clear incremental evolution of scripting and automation, vendors seem quick to make stronger claims than can be independently verified in the field.

Note to CISO’s

For CISO’s looking to ensure they can defend against AI malware, they should start by answering these three questions:

  • Do you have appropriate AI governance?
  • Can you find rogue usage of AI in your environment?
  • Are your defenses capable of defending against AI-powered tools?

Outlook

AI malware is still in its early stages of maturity, primarily situated at the experimental to optimization levels of the AIM3 framework. What we are seeing today is AI-assisted tradecraft that mirrors traditional TTPs; a natural result of increased AI adoption, which is being observed in several other industries as well.

That said, the direction AI adoption is taking is clear. The progression from straightforward AI-generated content to AI-invoking malware and red team orchestration frameworks is a sign that more capable and autonomous operations are on the horizon. The contested Anthropic disruption is one of the first examples of this and serves as an early warning sign for defenders.

All roads lead to AI orchestration

The trajectory of AI activity clearly leads toward AI orchestration, but not yet fully scaled AI operations. Frameworks like HexStrike-AI and CyberSpike’s Villager, although still niche today, point to attacker playbooks where orchestration and AI-driven tool use become the norm. The timeline of AI malware observations shows a clear progression: starting with simple AI-generated code, moving to on-host command generation and orchestration, and culminating in agentic, multi-step operations. If these frameworks continue to mature, they are likely to require less human intervention over time.

Recorded Future Malware Intelligence helps organizations understand and protect against malware

To see how Recorded Future can help your team track malware along the AI maturity scale, watch this 3-minute video on investigating LAMEHUG malware (AIM3).

Defenses