top of page
Search

NetFlow: Something I Seriously Underestimated (Until I Didn’t)

  • 3 hours ago
  • 7 min read
I’ll be honest.

For a long time, I never really gave NetFlow the priority it deserves. PCAP was always the gold standard in my head. If you want to know what really happened on the network, you go straight to packet capture. End of story.


But after reading more, testing more, and actually thinking about scale, cost, and real-world SOC/DFIR constraints, my opinion changed.

Today, I want to talk about why NetFlow matters, when it actually makes your job easier, and why full PCAP is not always the right answer — especially in enterprise environments.


-------------------------------------------------------------------------------------------------------------

Why Full PCAP Sounds Great… But Breaks in Reality

Yes, full packet capture is still the holy grail of network analysis. But the moment you move into a large corporate environment, things start to fall apart.


Here’s why PCAP doesn’t scale well:

  • Privacy laws (As I am currently in EU I can see this problem): In many regions (especially parts of the EU), capturing full packet content can be legally problematic — even on corporate networks.

  • Modern network volume : Network speeds are insane compared to a few years ago, and modern operating systems generate massive amounts of traffic.

  • Duplicate packets everywhere : If you capture traffic at multiple points (internet gateway, internal segments, regional links), you often store the same packets multiple times. Storage costs explode fast.

  • Deep analysis is expensive : Parsing huge PCAP datasets requires powerful servers, fast storage, and enterprise-grade tooling — none of this is cheap.

  • Global environments are painful : If your capture points are spread across Europe, APAC, and North America, centralizing that data for analysis means serious bandwidth costs.

  • Budget reality : Some companies won’t blink at a million-dollar project. Most organizations simply can’t afford it.

  • Encrypted traffic changed the game : With TLS everywhere, perfect forward secrecy, and certificate pinning, full payload retention is often useless. You’re storing a lot of data with very little analytical value.


This is where NetFlow quietly becomes extremely powerful.

-------------------------------------------------------------------------------------------------------------

So What Exactly Is NetFlow?

A NetFlow record is basically a statistical summary of network traffic observed at a specific point in the network.

Instead of storing packet content, NetFlow groups packets into a flow based on shared attributes, such as:

  • source IP

  • destination IP

  • protocol

  • source port

  • destination port

These grouped packets become a flow record that tells you what talked to what, when, and how much data moved.

You don’t see the payload —but you see the story of the connection.



A Simple Example (No Wireshark Needed)

Let’s say a user opens a browser and connects to a website.

  • The client uses a high TCP port (above 1023)

  • The server listens on TCP 443

  • A TCP session is established


A NetFlow sensor watching this link will notice:

  • when the connection started

  • which IP talked to which IP

  • which ports were used

  • how much data moved in each direction

  • when the communication stopped


Important detail:👉 NetFlow is unidirectional

So one browser session usually becomes two flow records:

  • client → server

  • server → client


NetFlow doesn’t care about “client” or “server” roles — only sender and receiver.

-------------------------------------------------------------------------------------------------------------

NetFlow + Encrypted Traffic = Perfect Match

Encrypted traffic is where NetFlow really shines.

If TLS interception is not enabled:

  • full PCAP gives you encrypted blobs

  • payload analysis becomes almost useless

  • disk usage goes through the roof

NetFlow doesn’t care about encryption.

It records:

  • who connected

  • when

  • how often

  • how much data moved


That’s often exactly what you need for:

  • command-and-control detection

  • beaconing analysis

  • lateral movement

  • data exfiltration investigations


-------------------------------------------------------------------------------------------------------------

Why DFIR and Threat Hunting Love NetFlow

Another underrated benefit: retention.

Because NetFlow data is:

  • small

  • content-free

  • privacy-friendly

Organizations often retain it for long periods of time.


This means:

  • you can apply new threat intelligence to old data

  • you can hunt retroactively

  • you can validate attacker dwell time

This is what real threat hunting looks like.

-------------------------------------------------------------------------------------------------------------

A Quick Word on NetFlow History (Because It Matters)

NetFlow was originally developed by Cisco back in 1996, not for security, but to optimize routing performance.

Over time, it evolved into a powerful traffic visibility mechanism.


Key points:

  • NetFlow v5 is still widely used (IPv4 only, unidirectional, old design)

  • NetFlow v9 introduced flexibility and extensibility

  • The IETF standardized this as IPFIX (sometimes called NetFlow v10)


Today:

  • Cisco supports v5 and v9

  • Most vendors support v5, v9, and IPFIX

  • VMware ESX can export IPFIX

  • Cloud providers have their own equivalents


Examples:


  • AWS VPC Flow Logs

  • Azure NSG Flow Logs

  • Google VPC Flow Logs

  • Zeek conn.log (very similar concept)


-------------------------------------------------------------------------------------------------------------

One Important Caveat: Sampling Matters

Before using NetFlow for investigations, you must know:

  1. Where it’s enabled : If only certain interfaces export flows, blind spots exist.

  2. Whether it’s sampled or not

    • Standard NetFlow tracks every packet

    • Sampled NetFlow tracks every n packets


Sampled NetFlow:
  • under-represents data volume

  • is not suitable for forensic accuracy

  • still useful for trends and visibility


This distinction is critical.

-------------------------------------------------------------------------------------------------------------


Now you know why NetFlow matters,

let’s talk about how it actually works in practice 

To build a NetFlow monitoring setup, you really only need four core components.



The Four Building Blocks of NetFlow

Think of NetFlow like a pipeline.


  1. Exporter

This is the device that creates NetFlow records.

Usually this is:

  • a router

  • a firewall

  • a Layer 3 / Layer 4 switch


But it doesn’t have to be limited to that. Anything that can observe traffic and summarize it can act as an exporter.


  1. Collector

The collector is where all those flow records land.

It receives NetFlow data from:

  • one exporter

  • or dozens of exporters


This is where visibility starts to come together.


  1. Storage

Collectors don’t just receive data — they store it.

This storage needs to be:

  • indexed

  • searchable

  • optimized for time-based queries


NetFlow records are small, but over time they add up, and that’s actually a good thing — historical data is gold for investigations.


  1. Analysis Console

This is where you sit.

It could be:

  • a web UI

  • a thin client

  • a CLI tool


This is where you ask questions like:
  • “Who talked to this IP?”

  • “When did this start?”

  • “How much data moved?”

  • “Is this normal?”


-------------------------------------------------------------------------------------------------------------

One Box or Many? Both Are Valid

In labs, training environments, or testing setups:

  • exporter

  • collector

  • storage

  • analysis

can all live on one Linux system.

In real environments, it’s usually different.

A more common setup looks like this:

  • routers and firewalls export NetFlow

  • flows go to a centralized collector/storage system

  • analysts access data via browser or CLI


Simple. Scalable. Effective.

-------------------------------------------------------------------------------------------------------------

Plan NetFlow Early — You’ll Thank Yourself Later

One thing I’ve learned the hard way:

NetFlow is much easier to deploy before the network design is finalized.

Yes, NetFlow records are small —but like all behavioral data, their real value is time.

The longer you retain them, the more powerful they become.


Exporters Are Not Just Routers

We often picture a router when we think about NetFlow exporters — and that’s fair.

But exporters can also be:

  • firewalls

  • L3/L4 switches

  • dedicated probes

  • even endpoints (in some cases)

Exporting from every endpoint usually isn’t practical.

However, during an incident?

Running a probe on a tap or SPAN port can give tactical visibility exactly where you need it.


-------------------------------------------------------------------------------------------------------------

Why NetFlow Uses UDP (And Why That’s Not as Bad as It Sounds)

By default, NetFlow uses UDP to send flow data from exporter to collector.


At first glance, this sounds scary:

“UDP is unreliable!”

But there are solid reasons for this choice.


Why UDP works well here:

  • Very low protocol overhead

  • Minimal CPU and memory usage

  • Scales well on high-speed links

  • Easy to send data to multiple collectors at once

That last point is important.

Security teams and network teams can:

  • receive the same flows

  • from the same exporters

  • without duplication overhead



What About Reliability?

Some platforms support SCTP (Stream Control Transmission Protocol) instead of UDP.

With SCTP:

  • collectors confirm receipt

  • exporters can retransmit if needed

  • data loss risk is reduced


On very large or fast links, exporters may also send interim flow updates before a flow ends. That way, even if something is lost, you still have partial visibility.

-------------------------------------------------------------------------------------------------------------


Where You Place Exporters Matters (A Lot)

This is where NetFlow design becomes interesting.

You don’t need NetFlow everywhere. You need it in the right places.

For example:

  • capturing at the firewall gives you internet visibility

  • capturing one hop behind the firewall gives you internal east–west traffic

  • capturing at core switches shows workstation-to-workstation movement


That subtle difference can completely change what you can detect.



Attackers Love a Three things— So Should You

Attackers usually operate around these three things:

  1. Internet

  2. Administrator workstations

  3. Critical data systems


If you design NetFlow around these things, you gain massive visibility.

This is how you catch:

  • lateral movement

  • credential abuse

  • staging before exfiltration


-------------------------------------------------------------------------------------------------------------

Full Packet Capture Still Has a Role

In some segments — especially small, high-risk ones —full packet capture still makes sense.

For example:

  • R&D networks

  • sensitive server

  • limited user groups

But even then:

NetFlow should support PCAP, not replace it.

PCAP gives detail. NetFlow gives context and speed.


-------------------------------------------------------------------------------------------------------------

A Practical Way to Think About Design

If you’re involved in NetFlow planning, these steps help keep things realistic:


  1. Identify critical data

Not everything matters equally. Find what actually needs protection.


  1. Map the network

You’d be surprised how many organizations don’t have a complete network map.

If you don’t know:

  • where data lives

  • how it moves

  • who accesses it

You can’t protect it properly.


  1. Find choke points

Where do multiple networks join?

Where does traffic have to pass?

Those are perfect NetFlow locations.


  1. Identify data “centers of gravity”

Not just “domain controllers” or “file servers”.

Think:

  • source code repositories

  • admin systems

  • finance platforms

  • executive devices

  • DNS servers

These are attacker magnets.


  1. Combine NetFlow and PCAP smartly


  • NetFlow at high-volume locations

  • PCAP at critical choke points

  • Centralized analysis

  • Sensible retention policies

If storage is tight, even a basic Linux box with large disks can do the job when configured properly.

-------------------------------------------------------------------------------------------------------------


Final Thought (For This Section)

NetFlow isn’t flashy. It doesn’t show payloads. It doesn’t decode protocols.

But it gives you speed, scale, and historical visibility —three things every SOC and DFIR team desperately needs.

We will continue further in next article!
Next Article : Where NetFlow Either Shines or Struggles

-------------------------------------------Dean-------------------------------------------------------------




 
 
 

Comments


bottom of page