The AI Malware Surge: Behavior, Attribution, and Defensive Readiness

Arctic Wolf Labs analyzed over 22k AI‑assisted malware samples to measure how broadly threat actors have adopted AI. While AI is accelerating malware development, the core behaviors needed for execution, persistence, and C2 remain detectable with layered defenses.
6 min read

Executive Summary

Over the last year, AI-assisted malware development has evolved from an experimental practice into a common part of the attacker toolkit. In a rolling window from February 2025 to February 2026, Arctic Wolf Labs observed over 22K distinct files trigger AI-focused YARA rules across multiple malware repositories. These files included AI-generated code, Large Language Model (LLM)-style scaffolding, runtime AI API integration, and DeepSeek-derived artifacts.

This shift was structural, not because attackers suddenly became more sophisticated, but because AI lowered the barrier to producing functional malware. We found that AI made malware creation faster, broader, and more accessible to threat actors who previously lacked the skill to build functional tooling on their own. This pattern was observed across multiple categories of malware, including infostealers, remote-access tools (RATs), and ransomware engines.

Key Points

  • AI-assisted malware development has moved from isolated experimentation to a routine part of the attacker workflow, expanding the pool of threat actors who can produce operational malware.
  • DeepSeek R1, released in January 2025, emerged as a prevalent AI tool in this dataset; a substantial share of reviewed samples included a DeepSeek filename prefix.
  • 39% of analyzed samples had zero detections by signature-based antivirus (AV) solutions at the time of collection, indicating that a significant portion of malware being created is structurally new.
  • Only 1.4% of AI-assisted malware was linked to known targeted attacks, known threat actors, or financially motivated cybercriminal clusters; the overwhelming majority originated from unknown or lower-skill actors.
  • AI amplifies the speed, scale, and reach of malware, but its behaviors remain detectable to defenders with layered visibility.

The Impact of AI in the Threat Landscape

AI is reshaping the threat landscape primarily through scale: it broadens who can build malware and accelerates how quickly functional tools emerge. In our research, we observed threat actors using LLMs to produce infostealers, RATs, droppers, ransomware engines, or other malicious scripts.

We observed a consistent pattern of threat actors learning iteratively, using LLMs to scaffold code and fill in gaps, moving from broken proof-of-concept (POC) implementations into usable malware faster and more capably than they could on their own. In that sense, AI is not simply producing more code; it is narrowing the gap between developer skill and operational capability, and timeline.

These technological developments do not imply that every piece of AI-generated malware is necessarily dangerous on the first iteration of development. Some of these AI-assisted samples were mature and operational. Others were clumsy, incomplete, or outright hallucinatory. The broader effect is that AI compresses the path from initial idea to fully realized capability. Novice malware authors who previously couldn’t build functional malware without AI can now produce POC code with advanced features such as defense evasion and privilege escalation, and this changes the threat landscape structurally at scale.

Across 22,331 files we analyzed in this research, 39% were undetected by signature-based antivirus solutions at the time of our collection, suggesting active evasion or novel construction rather than code classified under existing known-bad patterns. Only 16% of files analyzed had 40 or more positive detections at the time of collection.

Another notable shift is that some malware is no longer limited to AI-assisted development; a growing subset now incorporates AI at runtime. Across the dataset, we observed roughly 8% of samples with runtime LLM API integration patterns, and approximately 7% of samples embedding hardcoded LLM API keys. Today, much of that observed usage is still relatively primitive, such as dynamic naming or message generation, but the architectural direction matters because it points toward more adaptive malware behaviors over time.

Behavior and Attribution

AI Use During Malware Development

AI models are being used by malicious actors to generate malware architecture, templates, and loader frameworks scaffolding them. Typical artifacts of LLM use included:

  • Verbose or tutorial-style comments
  • Numbered task breakdowns
  • Markdown-like section headers
  • Emoji characters
  • Cited web-search references accidentally carried into code

These distinct LLM-generation artifacts informed the custom YARA rules we used to detect AI-assisted files. Runtime-focused rules, by contrast, focused on malware samples that queried provider APIs at execution time, indicating the use of AI models for dynamic text generation, naming, or ad-hoc decision-making.

Our analysis revealed several patterns relevant to attribution: filename conventions, citation artifacts from AI-based web searches, recurring human and programming language clusters, and family-level behaviors that make clear AI is being absorbed into distinct malware development communities. A central finding here is that AI use was clustered in recognizable ways by language, workflow, and malware family. These fingerprints provided us with a rich view into threat actor workflows.

Several patterns point to distinct threat actor communities adopting AI. These include DeepSeek-derived filenames, web-search citation markers in code, and language-based sample clustering. At the same time, hallucinated, non-executable samples suggest an iterative learning process in which threat actors refine AI-generated output until it becomes operational. Together, these findings illustrate how AI is shaping threat actor tradecraft.

DeepSeek-Assisted Malware Generation

DeepSeek’s R1 model had an outsized influence in the malware collection that we analyzed. This AI model represented a significant uplift in capabilities compared to previous options, offering low-cost access, strong Chinese-language support, and integrated web search. These traits made DeepSeek R1 attractive to cost-sensitive novice threat actors focused on script generation.

The global threat actor community appears to have adopted DeepSeek’s new tooling within weeks of its January 2025 release, marking a dramatic step-change in malware output rather than gradual growth. The resulting increase in global use of DeepSeek R1 corresponded with a surge in malicious usage of the model. Malicious sample submission remained elevated throughout the remainder of the analyzed window.

Figure 1: Malware sample submissions noticeably increased in concert with corresponding global adoption of advanced AI agents such as DeepSeek R1.

In a subset of samples we selected for manual review, more than half of the samples carried deepseek_-prefixed filenames. Similar references were also observed persisting in module paths, log files, and GitHub-style self-update strings. This is consistent with a copy-paste workflow in which threat actors keep model-exported names and move quickly to packaging and deployment.

A notable example of malware likely generated via DeepSeek is a sample named deepseek_rootkit. This Python-based worm spreads via internet-wide scanning for unauthenticated or weak password instances of Redis, and also performs SSH brute forcing. This malware includes Monero cryptocurrency mining capabilities and a peer-to-peer command-and-control (C2) framework. The sample included numerous references to DeepSeek, including mentions in its own logging, installation paths, and self-update infrastructure. While this malware is indeed capable of self-propagation, it does not exhibit actual rootkit capabilities due to bugs in its implementation.

Figure 2: Initialization of the deepseek_rootkit script.

One of the clearest indicators of DeepSeek usage throughout our dataset was the appearance of inline [citation:N] markers inside executable code. Those markers are characteristic of DeepSeek’s web-search mode, where the model inserts source references into generated text. We found these artifacts embedded in malware comments and code blocks. This is one of the strongest signals in our dataset that threat actors are using AI to facilitate initial research into malicious activities, asking models to retrieve or synthesize technical details, then carrying that output straight into malware development.

Figure 3: Inline[citation:N] markers observed directly in the executable code of the samples we analyzed, generated when DeepSeek operates in web-search mode.

Notably, code generated with DeepSeek more frequently included all-caps Markdown-style headers, prompt-language comments, and the occasional presence of inline citation artifacts from web-search mode.

Developer Language Clusters

Across the dataset, language and stylistic markers in code comments, variable names, scaffolding structures, and embedded strings aligned into several recognizable developer-language clusters, each with distinct technical tendencies and AI‑usage patterns.

Russian-language samples were consistently the most mature, with tooling that included Flask-based Malware-as-a-Service (MaaS) panels, AES‑256 ransomware engines, and multi-capability Telegram RATs. These samples often contained Russian prompt-language comments and all-caps Markdown-style headers that matched DeepSeek output characteristics.

English-language samples represented the largest overall cohort and displayed the widest range of capability. Some were fully operational stealers, while others were strongly hallucinated, including repeated attempts to import nonexistent modules, which caused immediate failure. Many samples showed verbose instructional comments, numbered scaffolding, and inconsistent edits that reflected a dependence on LLM-generated structure rather than developer expertise.

Portuguese and Brazilian-linked samples frequently contained emoji in comments, often mixing Portuguese, English, and Spanish in social-engineering dialogs, particularly in malware families such as Alastor 2025.

Turkish, Indonesian, and Chinese-linked samples also formed distinct clusters. Turkish actors produced antivirus evasion and test tooling generated module by module. Indonesian samples, including the BunnyKit family, showed multi-module scaffolding for hacktools and Android or Termux utilities. Chinese-linked samples contained planning notes in Simplified Chinese and included DeepSeek web-search mode artifacts such as inline citation markers within Python code.

Language Cluster Capability Malware Types AI Usage Pattern Confidence
Russian High Flask MaaS C2, AES ransomware, Telegram RAT Architecture scaffold + Russian prompts High
English (Global) Low to High Stealers, BSOD bots, gaming theft, hallucinations Scaffold + wish-fulfillment generation Medium
Portuguese / Brazilian Low to Medium Fantastical (wish-fulfillment) malware, partial stealers Emoji-heavy, wish-fulfillment prompts High
Turkish Medium AV evasion testing tools Module-by-module generation in Turkish High
Indonesian Low to Medium Web hacktools, SQLi, defacement Multi-module framework scaffolding High
Chinese (Inferred by YARA rule data) Medium to High Cloud implants, LLM-researched tools Planning artifacts in CJK comments Medium

 

Mature Threat Clusters

While much of the AI-generated malware in our 22K-sample dataset originated from low- to mid-tier threat actors, we were able to attribute roughly 1.4% of the samples back to known advanced persistent threat (APT) and financially motivated groups tracked over time by Arctic Wolf.

Among the 92 deep-dive samples analyzed, NyxStealer emerged as one of the most operationally mature commercial MaaS families analyzed. These variants shared a common Node.js codebase, used a nyx-local working directory, performed DPAPI-backed browser credential extraction, and exfiltrated data through Discord webhooks.

NyxStealer sits at the boundary between mid-tier crimeware and disciplined tradecraft. The evolutionary trajectory across its six generations show deliberate iteration, from implementing browser credential theft to reverse-base64 webhook encoding. Its codebase has evidence of LLM-style scaffolding, but the operational choices signal a threat actor who understands deployment realities. The result is a MaaS offering that combines AI-assisted development speed with deliberate human refinement.

It would be a leap to suggest that mature threat actors are replacing their established development process entirely with AI. Instead, they are augmenting their existing processes with AI to reduce friction in their workflows, shortening the time to modify payloads for evasive purposes, eliminate common coding errors, and rapidly generate new variants without increasing detection risk.

The integration of AI into tooling development by advanced threat actors complicates attribution efforts. When malware samples are AI-generated, they exhibit consistent LLM artifacts—stylistic patterns, code structures, and implementation of choices characteristic of the underlying model. These artifacts appear regardless of whether the actor is a sophisticated state-sponsored group, or a commodity cybercriminal using the same AI tools. This convergence obscures traditional attribution signals that previously helped distinguish advanced persistent threats from lower-sophistication operators.

The broader malware ecosystem is also showing clear signs of commercialization, a trend that predates widespread AI adoption but is now accelerating. Shared licensing, multi-operator distribution, and product-style release patterns all point to AI-assisted malware becoming part of a market-driven supply chain rather than remaining a collection of one-off experiments. The use of LLMs may accelerate the initial build of a malware variant, but MaaS developers turn that output into a product by hardening it, releasing new versions, and supporting their illicit user base. Buyers then deploy the tooling at scale across compromised hosts.

AI-assisted malware development should be thought of as a process rather than a point of origin. Models accelerate early-stage scaffolding, but they do not determine how code is refined, deployed, or reused across campaigns. Those outcomes are shaped by human operators who apply judgment, infrastructure, and operational discipline. The operational threat here lies with the actors who adapt AI-assisted code into durable, monetized attack infrastructure.

Signals of Iterative Learning

Not every AI-generated malware sample ends up running as its authors intended. A small subset of samples in our dataset was found to be hallucinated or structurally impossible. Across the non-functional samples in this subset, one of the most common markers was the use of import mimikatz. Mimikatz is a Windows executable, not a Python library; Python’s import system raises the ModuleNotFoundError before a single line of malicious logic was able to execute. These samples failed immediately.

Though these samples did not achieve their intended purpose, these files are still analytically relevant. They capture the experiences of threat actors in the process of learning as they repeatedly iterate on their prompts, paste new outputs, and keep failing until they’re able to arrive at functional code. This activity suggests that AI is shrinking the gap between intent and capability and broadening the set of actors who can iteratively build impactful malware.

The Defender Advantage

AI may change how malware is assembled, but it does not change the fact that malicious behavior stands out on multiple levels.

Regardless of how it is created, malware of all types and complexity levels still needs to execute, persist, evade detection, move in memory, spawn scripts, create tasks, beacon out, or alter the system in other observable ways. When we detonated malware from this dataset in a controlled lab setting, these samples left observable traces across multiple points in the execution chain.

A layered detection approach is needed to capture and highlight these types of traces throughout different stages of malicious activity.

A Layered Approach to Detection

Layer 1: Signature and Indicator-Based Detection

The first layer remains signature and indicator-based detection.

The starting point for detection from a defender’s perspective is working from technical details already known about a threat, such as malicious hashes, domains, IP addresses, URLs, filesystem artifacts, and other relevant details.

In our own analysis for the purposes of this research study, this includes the kinds of recognizable artifacts that repeatedly appeared in AI-assisted malware, such as hardcoded infrastructure, inline citation remnants from AI web-search workflows, and model-linked filename conventions.

Protection at this layer is worth pursuing because a meaningful share of malware still reuses infrastructure, code fragments, or other recognizable scaffolding even when threat actors iterate on their generated malware. On the endpoint, that often points to file, process, and persistence artifacts. On the network, it means known-bad destinations, beaconing infrastructure, or recognizable protocol use. In cloud environments, it includes suspicious API endpoints, identity artifacts, or workload telemetry tied to established indicators.

Signature coverage will not catch everything, especially when AI lowers the cost of producing structurally novel samples, but it remains an essential first layer for speed, scale, and high-confidence detections.

Layer 2: Behavioral and Execution-Based Detection

The second layer is behavior and execution-based detection, where defenders begin to push back effectively against novelty in malicious activity.

AI-generated malware may differ from one sample to another, but many samples still converge around predictable runtime patterns that leave observable traces. These behaviors include unusual invocation of scripting runtimes, unexpected filesystem or network activity spawned by PowerShell, scheduled task creation, suspicious WMI queries, and process injection. These patterns are conspicuous when compared to baseline legitimate activity and yields a strong signal that cuts across numerous malware families.

At the behavioral level, network telemetry adds essential context that endpoint logs alone cannot provide. In the context of AI-generated malware, this includes C2 beaconing activity disguised as innocuous services, Discord webhooks, Telegram bot calls, and other unconventional channels utilized by threat actors.

Arctic Wolf’s own detection coverage reflects this reality, spanning major threat surfaces rather than treating them in isolation. Endpoint telemetry provides a fundamental understanding of what executed and what it changed. Network telemetry shows which devices malware communicated with and how. Cloud telemetry shows where identities, workloads, or APIs are abused. Even when the malware itself is new, these operational patterns are often not.

Layer 3: Machine Learning-Based Detection

The third layer is machine learning-based detection. This is not a replacement for the first two layers, but a means of complementing them when malicious activity falls through the cracks.

Machine learning (ML) becomes valuable when static indicators are sparse and individual behaviors are too weak or noisy to stand on their own. It can evaluate large volumes of telemetry for patterns that do not map cleanly to a single rule or analytic description. By ingesting telemetry from the artifact and behavioral layers, outliers in user or process behavior can be more readily identified at scale.

Arctic Wolf’s defensive strategy reflects this reality: machine learning is most effective when paired with broad visibility and correlation across endpoint, network, cloud, and identity telemetry, rather than treated as a standalone control. This strengthens defensive posture regardless of whether malware is AI-generated or not.

Static signatures and behavioral rules remain an essential part of the defensive toolkit, but they often struggle to keep up against the rapid iteration cycles made possible by AI-assisted development. By extending detection capabilities through evaluation of structural attributes and behavioral telemetry as part of a unified feature set, ML-based detections can recognize malware families and variants even when threat actors alter or disguise their malware.

As models are retrained on new samples and telemetry, they can improve resilience against emerging variants and help defenders keep pace with rapidly iterating malware families.

Arctic Wolf Detection Analysis

Methodology

As part of our analysis, malware samples were clustered and classified based on similarity and other features. A subset of these malware samples was manually detonated in a secure lab environment to evaluate Arctic Wolf’s detection capabilities. Telemetry was captured through a combination of Arctic Wolf® Aurora™ Endpoint Defense (including Arctic Wolf’s proprietary anti-ransomware technology) and Arctic Wolf® Managed Detection and Response (MDR) network and endpoint behavioral analytics.

Arctic Wolf Aurora Endpoint Defense

Aurora™ Endpoint Defense provides endpoint protection against modern threats. Rather than relying on signature- or reputation-based detection, the Aurora platform evaluates the structural and behavioral characteristics of files at machine-speed to identify malicious binaries and related threats. A core component of this technology stack is Arctic Wolf® Aurora™ Protect, an Endpoint Protection Platform (EPP) that detects and blocks ransomware and other malware pre-execution on Windows, macOS, and Linux devices.

During lab testing, Aurora™ Protect blocked the majority of samples pre-execution. Aurora Protect uses mature machine learning models to identify malicious binaries based on the behavioral and structural characteristics of the files, and blocks them before they can execute. In addition to blocking most of the original samples, our latest ML models also detected and blocked additional in-memory activity and files dropped at runtime.

Figure 4: During lab-based detonation of malicious AI-generated samples, Aurora Protect – the EPP component of Aurora Endpoint Defense – was able to identify and block activities from distinct stages of the cyber kill chain.

Aurora Protect also includes several controls that add preventive enforcement capabilities. Aurora Protect’s Script Control component allows organizations to enforce a zero-trust execution model for scripting engines. When configured in Block mode, this capability prevents execution of scripts such as VBS, JavaScript, PowerShell, Python, and other interpreter-based content.

Aurora Protect’s MemDef capability is designed to detect and block common memory-based attack techniques. Many of the analyzed AI-assisted samples attempted process injection or related in-memory manipulation. In our testing, MemDef generated detections for, and in relevant cases blocked, these behaviors.

Arctic Wolf® Aurora™ Focus, the Endpoint Detection and Response (EDR) capability of Arctic Wolf Endpoint Defense, collects and analyzes enriched events from devices, allowing defenders to identify and resolve threats before they impact users and data. In our research study, Aurora™ Focus generated detections across several behavioral stages of the malware execution chain.

Aurora Focus was particularly effective at identifying persistence and defense-evasion activity. Many of these behaviors were intentionally disconnected from the original malware process tree using techniques such as scheduled tasks, WMI execution, or other indirect execution mechanisms. Despite this separation, Aurora Focus successfully detected the resulting malicious behaviors, giving defenders strong visibility across post-exploitation activity.

Arctic Wolf Anti-Ransomware Protection

Arctic Wolf’s AI-powered Aurora™ Anti-Ransomware protection capability also generated alerts during our analysis. Aurora Anti-Ransomware was able to identify the loading of vulnerable kernel drivers, which is indicative of potential Bring Your Own Vulnerable Driver (BYOVD) activity. This suggests the malware attempted to introduce a driver with kernel-level privileges that could be leveraged to disable security controls, elevate privileges, or enable additional malicious actions. The agent also detected a file created by the malware in an apparent attempt to steal browser credentials. Both techniques were successfully blocked across the samples analyzed.

Arctic Wolf MDR Behavioral Detections

Arctic Wolf Managed Detection and Response behavioral detections identified several attacker techniques associated with malicious scripting activity across these samples. These included system discovery, reconnaissance, suspicious PowerShell execution, and scheduled task creation, providing behavioral coverage across common execution and post-exploitation stages.

Arctic Wolf’s network sensor provided useful visibility into command-and-control communication. Detections derived from network telemetry identified C2 beaconing patterns, miner check-ins, and suspicious DNS activity, adding context to the malware’s communication behavior during execution.

Conclusion

The key takeaway from our research is that AI is expanding who can build malware and how quickly they can do it. By reducing development effort and cost, AI is helping more actors translate intent into operational capability, especially in the low- to mid-tier segment of the threat landscape. This is less a story about breakthrough malware innovation than it is about broadened and accelerated access to malware development itself.

Our findings suggest that this change is already taking shape in practical ways. Across the dataset, we saw evidence of iterative development: some samples were clumsy or non-functional, while others had clearly evolved into usable stealers, RATs, ransomware components, and Linux implants.

Despite these changes in the threat landscape, AI does not remove the defender advantage. Regardless of whether malware is developed by hand or through AI-assisted coding, it still reveals itself through various observable behaviors. These points of visibility remain a useful signal to defenders, especially when layered together and methodically processed to elucidate different stages of the cyber kill chain. In our analysis, Arctic Wolf telemetry and next-generation detection and protection models provided effective coverage across numerous stages of malware deployment.

The practical implication is straightforward: the most durable response to AI-assisted malware is still a layered defense. Mature prevention, behavioral detection, and cross-domain visibility remain crucial aspects of an effective defensive strategy in this expanding threat category.

Appendix

Indicators of Compromise (IOCs) 

NOTE: This report contains sensitive technical indicators intended for defensive use. Do not use these indicators or techniques for offensive purposes.

Telegram Bot Tokens (Live at Collection Time)

Token
Associated Family
6542741914:AAHmV_V5ecICjXaQzwan3Gf6_kz4k2oI3nc UltimateDiscordStealer
7783894445:AAFa4sP1oV8_oVxU2R8rdFt7KhSrDM1WS3k Polymorphic engine RAT
8000470850:AAHyT_Gwj6685m2I5ozXvtOfKEetCzFHcgw French Telegram RAT
8560781579:AAEUDh85VzbLprw5-LhAjxmxqQU62awFbsE NyxStealer v1
8208206890:AAEtzuW4hmQFHTxTIOBugdICEciLB2s3uzE NyxStealer v2

Discord Webhooks

Webhook ID Associated Family
1395054734787743834 needhelp7 OBLITERATOR (live at collection)
1465066143516459277 INFERNAL GRABBER 9000 (live at collection)
1466914664373026817 TroyStealer (primary)
1466914033511694512 TroyStealer (secondary)
1448889380151365694 XOR-encrypted stealer
1474984647388565554 Roblox Logger

File System, Registry, and Persistence Artifacts

Artifact Value Family
Directory %LOCALAPPDATA%\nyx-local\ NyxStealer family
Directory %APPDATA%\Roaming\pika\ Pika Dropper
File %SystemRoot%\System32\drivers\BlueSkyInject.dat BlueSky Inject
File pattern %Drive%\System32\Cache\Volatile\sys_*.dat BlueSky Inject
File (Linux) /usr/local/bin/.deepseek_* (multiple modules) deepseek_rootkit
File (Linux) /tmp/deepseek.log deepseek_rootkit
File masquerade sihost32.exe in System32 sihost32 dropper
Scheduled Task D0MINAG0N D0MINAG0N
Scheduled Task HiddenOptimizer / WindowsUpdateManager Pika Dropper
Scheduled Task WinNetObject sihost32 dropper
Scheduled Task BlueSkyInject / BlueSkyInject Maintenance BlueSky Inject
IFEO Hijack utilman.exe → sihost32.exe (login screen trigger) sihost32 dropper
Winlogon Shell cmd.exe /c exit (replaces explorer.exe) BlueSky Inject
Registry Run SysHelper sihost32 dropper
Registry Run WinMaintenance / SysOptimizer / MasterOptimizer Pika family
Registry Run WindowsHelper / SystemService RAGE MODE

 

Dataset Overview

YARA Rule Suite and Matches

YARA Rule Name Signal Type Matches
AI_Gen_EmojiInCode_Batch Emoji characters in BAT/script code 30%
AI_Gen_PyInstallerLLMPayload_Python LLM-generated Python in PyInstaller packages 18%
AI_Gen_EmojiInCode_Generic Emoji in any executable code context 9%
AI_Gen_LLMApiAbuse_MultiPlatform Runtime LLM API integration patterns 8%
AI_Gen_HardcodedLLM_APIKeys Hardcoded AI provider credentials in code 7%
AI_Gen_ResidualTraces_Generic LLM residual markers and disclaimers 5%
AI_Gen_VerboseComments_Batch Comment density exceeding code density 5%
AI_Gen_LLMPowered_RAT_DotNet .NET RATs with LLM-generation signatures 4%
AI_Gen_SuspiciousCloudImplant_Generic Cloud/Linux LLM-generated implants 3%
AI_Gen_WormGPT_PowerShell_Scaffold WormGPT-style PowerShell scaffolding 1%

 

Of all the YARA matches we found, 92 script-based files were selected for manual analysis. Our selection criteria prioritized representation across primary scripting languages (Python, DOS batch, JavaScript/Node.js, VBScript, PowerShell), novel or unreported behavioral patterns, and samples with distinctive attribution markers.

Malware Family Clusters

Across fourteen primary clusters of malware families, the following overview table maps each family to its threat-severity, language, functional type, and AI role.

Family / Cluster Type Language Severity AI Role
NyxStealer / TroyStealer cluster Infostealer MaaS JS/Node.js CRITICAL Scaffold + iteration
D0MINAG0N Worm / Wiper / RAT PowerShell/BAT CRITICAL Structural scaffold
Russian C2 Suite (deepseek_) MaaS C2 / Ransomware / Telegram RAT Python/BAT CRITICAL Architecture + Russian prompts
deepseek_rootkit Linux Rootkit / Miner / Worm Python CRITICAL Full generation (self-named)
Alastor 2025 VBScript Multi-Language Dropper VBScript HIGH Iterative multi-version builds
Pika BAT Dropper chain Staged GitHub Dropper BAT HIGH Numbered-stage scaffolding
PHANTOM_REALM / SHADOW_REALM BYOVD Framework / RAT Python HIGH Research + citation generation
Somalifuscator Loader Obfuscated Stage-1 Dropper BAT HIGH Third-party obfuscation tooling
WannaCry Clone / NIGHTMARE File Locker / USB Worm BAT HIGH (symbolic) Template generation
BlueSky Inject Disk Exhaustion + Explorer Kill BAT HIGH Full generation
BunnyKit v3.2.0 Indonesian Hacktivist Multi-Tool Python MEDIUM Module scaffolding
Nuclear YARA Trigger EDR Probe / Red Team Tool PowerShell MEDIUM Full generation
WindowsAudioService RAT Full-Featured Python RAT Python HIGH Architecture + AMSI bypass
Fantastical Malware Cluster Non-functional LLM Output Python LOW Hallucinated / wish-fulfillment


Timeline Distribution

DeepSeek R1 was released on January 20, 2025, and the resulting increase in its global use corresponded with a corresponding surge in malicious usage/adoption of the model. Malicious sample submission remained elevated through the remainder of the February 2025–February 2026 review window.

Month Malicious Sample Submissions*
February 2025 155
March 2025 1441
April 2025 1632
May 2025 1881
June 2025 1793
July 2025 2003
August 2025 2454
September 2025 1751
October 2025 1453
November 2025 1823
December 2025 2150
January 2026 2049
February 2026 1321

*Provided counts do not fully capture all relevant sample submissions due to metadata limitations.

Referential Hashes

SHA-256
0e7802eeaca406ead3740d2eeacbb786b75e026212ec0c65e0f2f89561940d2b
7a9e20192d7391826adc96574ddb2778e67783ac317f07a01de717ab6f2955fe
8471257186db7db30d74816409fa09a09898ee099e7e0d1ad015546975e53a8f
b954ba7bca64b0f9bb98d61cd752859bd6edbcbf5052e75605a3644006ee9fd3
66a6ee009bf2de7703319a0e8523914822e28d88c2b755f30aa479a8d9c1a4ce
d9c7314568e03ff1f4c6e6ece56bdd46c9ea94ec37ba9fce56f707a24ebb1e93
4f94977a0d43789f66269578a6325f24a513aaef82c3334094448918cf9ad184

 

Legal disclaimer: Attribution reflects Arctic Wolf Labs’ assessment as of the report period and may evolve with new evidence. References to threat actor identity, nexus, and intent are analytical judgments, not statements of legal fact. This alert is provided for informational purposes only and does not constitute a guarantee of detection or prevention. Defensive effectiveness varies by environment, configuration, and available telemetry.

 

About Arctic Wolf Labs

Arctic Wolf Labs is a group of elite security researchers, data scientists, and security development engineers who explore security topics to deliver cutting-edge threat research on new and emerging adversaries, develop and refine advanced threat detection models with artificial intelligence and machine learning, and drive continuous improvement in the speed, scale, and detection efficacy of Arctic Wolf’s solution offerings.

Arctic Wolf Labs brings world-class security innovations to not only Arctic Wolf’s customer base, but the security community at large.

Share this post: