Blog

From packets to Patterns: How Foundation Models Detect Network Threats

|

In Part 1, we examined why traditional network detection fails: encryption blinds payload inspection, signatures only catch known threats, and anomaly systems drown teams in noise. Part 2 explored how Living off the Land (LOTL) attacks exploit these limitations by using legitimate tools that generate network traffic indistinguishable from normal operations.

The conclusion from both parts is unavoidable: traditional approaches to network threat detection cannot protect against modern attacks. Packet inspection, signature matching, and simple anomaly detection are fundamentally inadequate for identifying sophisticated threats designed to evade exactly these techniques.

So what actually works? The answer requires rethinking what we analyze, how we analyze it, and what we're looking for. It requires moving from packets to patterns, from signatures to understanding, and from reactive detection to predictive identification of malicious behavior.

The Data Source That Changes Everything

Before addressing how to detect network threats effectively, we need to examine what data we should analyze. Traditional network security focuses on packets: the raw data units transmitted across networks. Deep packet inspection analyzes packet contents, looking for malicious payloads or suspicious patterns.

But packet-based approaches face insurmountable challenges we've already discussed: encryption makes payloads opaque, infrastructure requirements for packet capture don't scale, and LOTL attacks use legitimate protocols that appear normal even when payloads are visible.

There's a different data source that avoids these limitations while providing richer behavioral context: NetFlow.

NetFlow records metadata about network communications without capturing actual packet contents. For each network conversation, NetFlow logs capture which systems communicated, when the communication occurred, how long it lasted, which protocols were used, how much data transferred, and other behavioral characteristics.

Think of it as the difference between recording every word of every phone conversation versus logging who called whom, when, for how long, and from which locations. NetFlow provides the behavioral summary without the content details.

This distinction matters enormously for modern network detection. NetFlow data remains useful even when traffic is encrypted because it doesn't rely on payload inspection. It scales efficiently because you're collecting metadata rather than full packets. And most importantly, it captures the behavioral patterns that reveal attack progression regardless of what legitimate tools attackers use.

When an attacker conducts LOTL reconnaissance by querying Active Directory from a compromised workstation, NetFlow captures: this workstation communicated with domain controllers, at these times, using LDAP protocol, transferring this volume of data, in these sequential patterns. The behavioral metadata reveals reconnaissance activity even though the individual connections use legitimate protocols.

Why Network Metadata Reveals Intent

The key insight enabling effective network threat detection is that attack behaviors create distinctive patterns in network metadata, even when individual actions appear completely legitimate.

Consider reconnaissance activity. An attacker systematically queries domain controllers to map network topology, enumerate user accounts, and identify valuable targets. Each individual query uses legitimate protocols and appears normal in isolation. But the pattern of queries reveals intent.

The behavioral signature shows: sequential connections to multiple domain controllers, consistent query patterns across systems, timing suggesting automated enumeration rather than interactive investigation, data volumes characteristic of bulk information gathering rather than specific lookups.

These patterns exist in the network metadata regardless of whether you can inspect packet payloads. The "who talked to whom, when, how much, and how" reveals reconnaissance behavior even when the "what was said" is encrypted or uses legitimate protocols.

Similarly, lateral movement creates distinctive metadata patterns. An attacker moving from an initial foothold to other systems generates network flows showing: sequential connections from one system to many others, use of administrative protocols, timing patterns indicating systematic progression rather than routine management, connection sequences that follow network topology in ways suggesting exploration.

Data exfiltration produces characteristic patterns in flow metadata: large sustained transfers to unusual destinations, data volumes exceeding normal baselines, timing patterns suggesting bulk copying rather than routine access, connections to cloud storage or external systems using legitimate protocols but in unusual patterns.

The critical difference between metadata-based detection and traditional approaches: you're not looking for specific malicious content or comparing traffic against known attack signatures. You're recognizing behavioral patterns that indicate malicious intent regardless of what tools or protocols are used.

Foundation Models: Learning the Language of Networks

Understanding how to identify these behavioral patterns at scale requires artificial intelligence, but not the superficial AI features many security vendors bolt onto existing products. It requires foundation models specifically designed to understand network behavior.

Foundation models represent a category of AI trained on massive datasets to develop generalizable understanding applicable across many scenarios. You're likely familiar with Large Language Models (LLMs) like ChatGPT or Claude that learned language patterns from enormous text datasets. These models understand language structure, context, and meaning in ways enabling them to handle novel situations they've never explicitly seen.

The same principle applies to network security, but the "language" being learned is network behavior rather than human language. Log Language Models (LogLMs) are foundation models trained on vast quantities of network flow data to understand patterns in how systems communicate.

Just as LLMs learned grammar, syntax, and semantic relationships from text, LogLMs learn communication patterns, behavioral sequences, and temporal relationships from network flows. They develop an understanding of what normal network behavior looks like across countless different environments and scenarios.

This understanding differs fundamentally from signature-based detection or simple anomaly detection. The model doesn't match against known bad patterns or flag statistical deviations. It comprehends normal behavior in sophisticated ways, then identifies sequences that don't make sense even when individual components appear legitimate.

The Architecture That Matters

The technical architecture underlying effective LogLMs differs significantly from general-purpose LLMs in ways that matter for network security.

Where LLMs use decoder architectures optimized for generating text, LogLMs typically employ transformer-based encoder models optimized for understanding patterns in sequential data. This architectural difference reflects the different tasks: LLMs generate human language, while LogLMs analyze temporal sequences of network events.

The temporal focus is critical for network security. Attacks unfold as sequences of events over time. Initial access, reconnaissance, lateral movement, privilege escalation, data exfiltration. Each phase leaves traces in network flows, but understanding these traces as a cohesive attack requires comprehending their temporal relationships.

LogLMs learn to recognize attack progression patterns by analyzing how network behaviors relate to each other over time. Not just "this system connected to that system" but "this system connected to multiple systems in this particular sequence, with this timing, following this pattern of exploration."

This temporal understanding enables detection of attack chains even when individual steps use legitimate tools and protocols. The model recognizes that the sequence of behaviors, taken together, indicates malicious activity regardless of whether each individual action appears normal.

Research has demonstrated that properly architected LogLMs can achieve detection accuracy exceeding 99 percent with false positive rates below one percent. At a major telco, LogLM deployments achieved F1 scores of 0.96 to 0.98 in zero-shot mode, detecting novel attacks without retraining, and 0.99 when fine-tuned to specific environments.

Zero-Shot Detection: Catching What You've Never Seen

One of the most powerful capabilities of foundation models is zero-shot learning: the ability to recognize patterns they've never been explicitly trained on.

For network security, this means detecting novel attacks that don't exist in the training data. Traditional signature-based systems can only detect threats someone has previously identified and created signatures for. LogLMs can identify malicious behavioral patterns they've never seen before because they understand what normal patterns look like at a fundamental level.

When an attacker develops a new LOTL technique using a previously unexploited combination of legitimate tools, signature-based systems have no way to detect it. LogLMs recognize that the behavioral pattern doesn't match learned normal patterns, even if the specific combination of actions is novel.

This capability addresses the fundamental problem with reactive detection: you're always defending against yesterday's threats. Zero-shot detection enables identification of new attack methods the moment they appear in your environment, before security researchers have analyzed them and created detection signatures.

The telco results demonstrated this capability in production environments. The LogLM detected novel attack patterns in zero-shot mode with F1 scores of 0.96 to 0.98, meaning it caught new threats it had never been trained on with exceptional accuracy and minimal false positives.

Continuous Adaptation Without Manual Tuning

Another critical advantage of foundation model approaches is continuous learning. Traditional network detection systems require constant manual maintenance: writing new rules, tuning thresholds, updating signature databases, adjusting anomaly baselines.

LogLMs adapt automatically as your environment evolves. When you deploy new applications, modify infrastructure, or shift user behavior patterns, the model's understanding of normal updates continuously without requiring manual intervention.

This addresses a persistent challenge in network security: environment drift. What counted as normal six months ago might look completely different today. Traditional systems either generate increasing false positives as they flag evolved normal behavior or miss attacks that blend with the new normal.

LogLMs handle this through active learning, continuously refining their understanding based on observed behavior. As your environment changes, the model's baseline understanding evolves automatically. This happens at machine speed, adapting in minutes or hours rather than the weeks or months required for manual rule updates.

Field deployments have demonstrated that LogLMs can adapt to new environments and evolving patterns within minutes, maintaining detection accuracy without the constant tuning traditional systems demand. This dramatically reduces the operational burden on detection engineering teams while improving detection effectiveness.

Contextual Understanding Over Isolated Alerts

Traditional network detection systems generate alerts about individual suspicious events. An IDS flags a connection matching a known attack signature. An NDR alerts on unusual traffic volume to a specific destination. A SIEM correlation rule triggers on multiple failed authentication attempts.

Each alert represents an isolated observation. Security analysts must manually correlate these fragments to understand whether they represent a coordinated attack or unrelated benign activities. This correlation process takes hours or days, during which the attack continues progressing.

LogLMs provide contextual understanding automatically. Because the model analyzes network behavior holistically, it doesn't just identify suspicious individual events. It shows how those events relate to other activity across your environment, constructing the complete attack narrative.

When the model identifies reconnaissance activity, it automatically correlates this with subsequent lateral movement attempts, privilege escalation indicators, and potential data exfiltration patterns. Rather than generating dozens of isolated alerts requiring manual correlation, it presents the attack story: here's what happened, in what sequence, affecting which systems.

This contextual understanding transforms incident response. Instead of starting with a single suspicious event and spending days building context, analysts begin with the complete picture already assembled. They understand what the attacker did, how far they progressed, which systems were affected, and what data might be at risk.

The Infrastructure Reality

Implementing foundation model-based network detection effectively requires appropriate infrastructure, but not the massive hardware investments packet-based inspection demands.

Modern security data lakes built on platforms like Snowflake provide the scalability needed to process network flow data at scale. These platforms handle petabytes of data while maintaining the query performance required for real-time analysis.

Agent-free architectures enable deployment without installing software on every endpoint or network device. The LogLM runs where your security data already exists, analyzing flows as they're collected without requiring data to leave your controlled environment. This addresses both security concerns about data handling and practical deployment challenges.

Integration with existing security infrastructure ensures LogLM insights drive actionable responses. Connections to SIEM systems enable alert routing through established workflows. Integration with investigation tools provides analysts additional context when needed. APIs enable automation of response actions based on detected threats.

The infrastructure approach mirrors broader industry trends toward security data lakes as central repositories for all security telemetry. Rather than scattering data across dozens of point solutions, organizations centralize security data then apply advanced analytics including LogLMs to extract insights. This architecture scales more efficiently than traditional approaches while providing better visibility.

Real-World Performance Metrics

The theoretical benefits of LogLM-based detection mean nothing without proven performance in production environments. Multiple deployments have now demonstrated consistent results across diverse organizations.

At one of the world's largest financial institutions, LogLM deployment achieved false positive rates below one percent while maintaining detection rates exceeding 99 percent. This combination of high accuracy and low noise represents a dramatic improvement over traditional network detection approaches that either miss sophisticated threats or overwhelm analysts with false positives.

The telco's deployment demonstrated another critical capability: rapid adaptation to new environments. The LogLM achieved F1 scores of 0.96 to 0.98 in zero-shot mode before any environment-specific tuning, then improved to 0.99 after fine-tuning. The system adapted to changing traffic patterns in minutes, maintaining detection accuracy as the network evolved.

These aren't laboratory results or proof-of-concept demonstrations. They represent production deployments in complex, high-stakes environments where detection failures could have massive consequences. The consistent performance across different organizations and network architectures demonstrates that the foundation model approach works reliably at scale.

What This Means for Security Teams

For security operations teams struggling with traditional network detection challenges, LogLM-based approaches provide practical relief on multiple fronts.

Alert volume decreases dramatically. Instead of triaging thousands of low-fidelity alerts daily, analysts focus on verified threats with sufficient context to respond effectively. The shift from noise to signal means security teams can actually investigate threats rather than just processing alerts.

Investigation time shrinks from days to minutes. The automatic narrative construction eliminates manual correlation work. Analysts understand what happened and what needs remediation immediately rather than spending days piecing together the attack story.

Detection coverage improves without adding tools. A single LogLM analyzing network flows provides visibility across on-premises infrastructure, cloud environments, and SaaS applications. This unified detection layer eliminates the gaps that arise from fragmented point solutions.

Maintenance burden drops substantially. Continuous learning eliminates the rule-writing treadmill. No more tuning thresholds, updating signatures, or adjusting baselines. The system adapts automatically, freeing detection engineers for strategic work rather than operational maintenance.

The Integration Reality

Organizations reasonably ask how LogLM-based detection fits into existing security infrastructure. Most enterprises have significant investments in current security tools and established workflows around them.

The practical answer: LogLMs augment rather than replace existing security stacks. They analyze network flow data that typically overwhelms SIEM systems, providing high-fidelity alerts that integrate with existing incident response workflows.

For many organizations, network flow data represents the highest-volume, lowest-value telemetry in their security data lakes. SIEMs struggle to process these volumes, leading organizations to either exclude flow data or severely sample it, creating detection blind spots.

LogLMs efficiently analyze full flow datasets without sampling, extracting meaningful threat indicators from data that would otherwise go unanalyzed. The resulting high-confidence alerts feed into existing SIEM platforms for workflow management and correlation with other telemetry.

Integration capabilities matter enormously here. Production LogLM implementations connect with tools like Cribl for data routing, CrowdStrike for endpoint context enrichment, and Snowflake for scalable data processing. These integrations enable LogLM insights to enhance existing security operations rather than requiring wholesale replacement of current infrastructure.

Looking Forward: Multi-Log Intelligence

Current LogLM deployments focus primarily on network flow data because this represents the most challenging detection gap for traditional approaches. But the foundation model architecture extends naturally to other log types.

The same techniques that enable LogLMs to understand network behavior patterns apply to cloud audit logs, SaaS application logs, identity and access management logs, and endpoint telemetry. Future implementations will likely unify detection across all these log types, providing truly holistic threat detection.

This expansion addresses a persistent challenge in security operations: correlation across different data sources. Today, network security teams analyze flow data, cloud security teams examine cloud logs, identity teams monitor authentication events, all using different tools with limited correlation.

A unified LogLM trained on multiple log types could automatically correlate suspicious network behavior with corresponding identity events and cloud API calls, providing complete attack visibility regardless of which environments attackers traverse.

Early research into multi-log LogLMs shows promising results, suggesting that the same foundation model architectures that excel at network flow analysis can effectively analyze diverse log types simultaneously. Production deployments of these capabilities will likely emerge as the technology matures.

The Shift from Reactive to Predictive

Perhaps the most significant implication of LogLM-based detection is the shift from reactive to predictive security.

Traditional detection is inherently reactive. Signature-based systems detect attacks after security researchers have identified and documented them. Anomaly systems flag suspicious activity after it occurs. Both approaches respond to threats rather than anticipating them.

LogLMs understand behavioral patterns in ways that enable earlier detection, often identifying attack indicators before the attack fully materializes. By recognizing reconnaissance patterns, they can alert on initial compromise stages before attackers achieve their primary objectives.

This predictive capability stems from understanding attack progression. Rather than waiting for the final stages of an attack to trigger alerts, the system identifies early-phase behaviors indicating likely attack progression. Security teams can intervene during reconnaissance or early lateral movement rather than discovering breaches after data exfiltration.

The timeline difference is substantial. Traditional detection often identifies breaches days, weeks, or months after initial compromise. LogLMs can alert within minutes of initial suspicious behavior, dramatically reducing attacker dwell time and limiting potential damage.

Beyond the Hype: What Actually Works

The cybersecurity industry has overpromised on AI capabilities for years. Every vendor claims to use machine learning. Security conferences overflow with AI buzzwords. Yet many organizations report disillusionment when AI features fail to deliver promised improvements.

This skepticism is warranted. Much of what's marketed as "AI-powered security" amounts to superficial features that don't fundamentally change detection capabilities.

But dismissing all AI and deep learning approaches because many implementations disappoint would be equally mistaken. Foundation models purpose-built for security logs represent a genuinely different technical approach that solves problems traditional methods cannot address.

The key differences: these models were trained specifically on security data, they learn behavioral patterns rather than matching signatures, they provide contextual understanding rather than isolated alerts, and they've proven effective in production environments at enterprise scale.

Organizations evaluating AI-powered network detection should look for these characteristics: foundation model architecture specifically designed for log analysis, proven performance metrics from production deployments, ability to detect novel attacks in zero-shot mode, continuous learning without manual maintenance, and integration capabilities with existing security infrastructure.

The Path Forward

Across this three-part series, we've examined why traditional network detection fails, how modern attacks exploit those failures, and what actually works for detecting sophisticated threats.

The conclusions are clear. Packet inspection fails against encrypted traffic and doesn't scale. Signatures only catch known threats. Anomalies generate overwhelming noise. LOTL attacks appear identical to legitimate traffic using traditional detection approaches.

Effective network threat detection requires analyzing behavioral metadata rather than packet contents, understanding patterns rather than matching signatures, and learning continuously rather than relying on manual rule maintenance.

Foundation models trained specifically on network flow data provide these capabilities. They detect novel attacks, adapt automatically, provide contextual understanding, and operate effectively at enterprise scale. Production deployments have demonstrated consistent performance across diverse environments.

For security leaders facing the reality that traditional network detection cannot protect against modern threats, the question isn't whether to change approaches but when and how. The technology exists. The performance is proven. The integration paths are clear.

Your network security stack will continue failing to detect sophisticated threats until you fundamentally change what you analyze and how you analyze it. The attackers already understand this. They've adapted their techniques to exploit the limitations of packet inspection, signatures, and simple anomalies.

The defenders who succeed will be those who recognize that effective network security requires understanding behavioral patterns in network metadata, enabled by foundation models specifically designed for this purpose.

The choice is straightforward: continue investing in detection approaches that sophisticated attackers have already learned to evade, or adopt the technology that actually works against modern threats. The longer you wait, the more time attackers have to operate undetected in your environment.

See the threats your tools can’t.

DeepTempo’s LogLM works with your existing stack to uncover evolving threats that traditional systems overlook — without adding complexity or replacing what already works.

Request a demo
Empowering SOC teams with real-time collective AI-defense and deep learning to stop breaches faster.
Built by engineers and operators who’ve lived the challenges of security operations, we deliver open, AI-native software that runs on any data lake—freeing teams from legacy constraints. Our LogLMs return control to defenders, enabling faster, smarter, and more collaborative responses to cyber threats.