Skip to content

SNL-UCSB/netFound

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

28 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

netFound: Foundation Model for Network Security

This repository contains the source code for netFound, a foundation model for network telemetry developed by the Systems & Networking Lab (SNL) at UC Santa Barbara.

Description

netFound is designed to learn spatial-temporal relationships from raw network traffic, making it a powerful tool for network analysis, anomaly detection, and traffic prediction.

πŸ”‘ Key Features

  • Raw Packet Processing: Directly processes raw PCAP files as input, enabling full-scale network traffic analysis.
  • Pretraining on Unlabeled Data: Requires pretraining on large-scale, unlabeled network telemetry datasets, leveraging self-supervised learning.
  • Hierarchical Transformer Architecture: Captures both packet bursts and flow-level behavior, ensuring robust feature extraction.
  • Metadata-Aware Processing: Integrates burst-level metadata such as:
    • Inter-arrival time*
    • Number of bytes per burst
    • Packet-level timing and structure

πŸ“Œ Why Use netFound?

netFound is part of a larger effort to develop self-driving networksβ€”autonomous, adaptive network systems that require minimal human intervention. By leveraging network foundation models, we aim to improve the efficiency and scalability of AI-powered Network Operations (AIOps). Corresponding paper: https://arxiv.org/abs/2310.17025

Checkpoint

https://huggingface.co/snlucsb/netFound-640M-base The checkpoint is pretrained on ~450mln flows of the real-world network traffic of the University of California, Santa Barbara.
As the checkpoint is built on the Large version of the netFound, use --netfound_large True as a fine-tuning flag.
Pretrained model metrics:

  eval_loss                         =     1.8847
  eval_macro_mlm_f1                 =     0.4038
  eval_macro_mlm_prec               =     0.7205
  eval_macro_mlm_recall             =     0.3005
  eval_mlm_acc                      =     0.8514
  eval_swapped_macro_pred_f1        =     0.9605
  eval_swapped_macro_pred_prec      =      0.963
  eval_swapped_macro_pred_recall    =     0.9603
  eval_swapped_pred_acc             =     0.9605
  eval_swapped_weighted_pred_f1     =     0.9605
  eval_swapped_weighted_pred_prec   =     0.9628
  eval_swapped_weighted_pred_recall =     0.9605
  eval_weighted_mlm_f1              =     0.8451
  eval_weighted_mlm_prec            =     0.8816
  eval_weighted_mlm_recall          =     0.8514
  perplexity                        =     6.5842
  Total params:  643,825,672

πŸš€ Quick Start: Running netFound with Docker & Makefile

The easiest way to verify that the preprocessing code and model work correctly is to use the provided Dockerfile and Makefile. This setup ensures a reproducible environment with all dependencies installed and includes a small test dataset to validate the pipeline.

πŸ› οΈ Step 1: Build the Docker Container

Run the following command to build the container:

docker build -t netfound:test .

This will create a Docker image named netfound:test, including the source code and a test dataset located in data/test.

▢️ Step 2: Run the Container

Start an interactive session inside the container:

docker run -it netfound:test

This will launch a shell inside the container in the /workspace directory.

⚑ Step 3: Run the Full Pipeline

Inside the container, execute:

make all

This will sequentially run the following three steps on the test dataset:

  1. Preprocessing: Converts raw PCAP files into a format suitable for training.
  2. Pretraining: Runs self-supervised learning on preprocessed data.
  3. Finetuning: Adapts the model for downstream tasks using the preprocessed test dataset.

πŸ—οΈ Understanding the Makefile & Dockerfile

The Dockerfile and Makefile automate the pipeline and provide a structured workflow:

πŸ“Œ Dockerfile

  • Creates a containerized environment with all necessary dependencies installed.
  • Ensures consistent execution across different systems.

πŸ“Œ Test Dataset (data/test/)

  • Contains raw PCAP files formatted for preprocessing.
  • Used to verify the pipeline’s functionality.

πŸ“Œ Makefile Structure

  • make preprocess:
    • Filters, splits, and tokenizes the raw packet data.
  • make pretrain:
    • Runs self-supervised pretraining on the preprocessed dataset.
  • make finetune:
    • Trains the model on task-specific labeled data.

πŸš€ Bring Your Own Data (BYOD)

To train or fine-tune netFound on your own dataset, follow the steps below to preprocess and tokenize your PCAP files.

πŸ“Œ Preprocessing Your Dataset

The easiest way to preprocess your dataset is to use the scripts/preprocess_data.py script.

πŸ“‚ Folder Structure for Pretraining

Organize your dataset as follows:

folder_name/
 β”œβ”€β”€ raw/
 β”‚   β”œβ”€β”€ file1.pcap
 β”‚   β”œβ”€β”€ file2.pcap
 β”‚   β”œβ”€β”€ ...

Then, run the following command:

python3 scripts/preprocess_data.py --input_folder folder_name --action pretrain --tokenizer_config configs/TestPretrainingConfig.json --combined

πŸ”Ή What happens next?

  • The script will generate intermediate folders (extracted, split, etc.).
  • The resulting tokenized data will be stored in the "tokens" folder.
  • The --combined flag merges all tokenized files into a single Arrow file (useful for training).
  • If you remove --combined, multiple Arrow files (one per PCAP) will be createdβ€”this is beneficial for parallel processing across multiple nodes.
  • You can modify the tokenizer configuration (configs/TestPretrainingConfig.json) to control how internal and external IPs are handled.

πŸ“‚ Folder Structure for Fine-Tuning

To fine-tune netFound, structure your dataset into class-separated folders, where folder names should be integers (used as class labels).

raw/
 β”œβ”€β”€ 0/
 β”‚   β”œβ”€β”€ class1_sample1.pcap
 β”‚   β”œβ”€β”€ class1_sample2.pcap
 β”‚   β”œβ”€β”€ ...
 β”œβ”€β”€ 1/
 β”‚   β”œβ”€β”€ class2_sample1.pcap
 β”‚   β”œβ”€β”€ class2_sample2.pcap
 β”‚   β”œβ”€β”€ ...

Run the preprocessing script again, changing the --action to finetune:

python3 scripts/preprocess_data.py --input_folder folder_name --action finetune --tokenizer_config configs/TestPretrainingConfig.json --combined

πŸ”Ή Fine-Tuning Notes:

  • Class labels must be integers (e.g., 1, 2, 3, ...).
  • The resulting Arrow files will include a "labels" column.
  • You can manually edit the "labels" column for custom class adjustments (including regression tasks).
  • As default validation data split does not shuffle the data file before the split, if your data is not shuffled, please use scripts/shuffler.py to shuffle the train file to ensure that the resulting test file contains instances of different classes.

πŸ”§ Advanced Options

Handling TCP Options

  • To include TCPOptions in your preprocessed data, use the --tcp_options flag:
python3 scripts/preprocess_data.py --input_folder folder_name --action pretrain --tokenizer_config configs/TCPOptionsConfig.json --combined --tcp_options
  • Prerequisite: Your dataset must be preprocessed with an additional flag when using 3_extract_fields.py:
python3 scripts/3_extract_fields.py input.pcap output.pcap 1
  • Ensure you use a config file that includes TCPOptions processing (e.g., configs/TCPOptionsConfig.json).

How to cite

@misc{guthula2024netfoundfoundationmodelnetwork,
      title={netFound: Foundation Model for Network Security},
      author={Satyandra Guthula and Roman Beltiukov and Navya Battula and Wenbo Guo and Arpit Gupta and Inder Monga},
      year={2024},
      eprint={2310.17025},
      archivePrefix={arXiv},
      primaryClass={cs.NI},
      url={https://arxiv.org/abs/2310.17025},
}

Acknowledgements

NSF Awards CNS-2323229, OAC-2126327, and OAC2126281 supported this work. This research used resources at the National Energy Research Scientific Computing Center (NERSC), a DOE Office of Science User Facility supported by the Office of Science of the U.S. Department of Energy under Contract No. DE-AC02-05CH11231 using NERSC award NERSC DDR-ERCAP0029768. Additionally, we would like to thank Cisco Research for their support.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published