Papers
arxiv:2409.11276

Hackphyr: A Local Fine-Tuned LLM Agent for Network Security Environments

Published on Sep 17
· Submitted by marik0 on Sep 23
Authors:
,

Abstract

Large Language Models (LLMs) have shown remarkable potential across various domains, including cybersecurity. Using commercial cloud-based LLMs may be undesirable due to privacy concerns, costs, and network connectivity constraints. In this paper, we present Hackphyr, a locally fine-tuned LLM to be used as a red-team agent within network security environments. Our fine-tuned 7 billion parameter model can run on a single GPU card and achieves performance comparable with much larger and more powerful commercial models such as GPT-4. Hackphyr clearly outperforms other models, including GPT-3.5-turbo, and baselines, such as Q-learning agents in complex, previously unseen scenarios. To achieve this performance, we generated a new task-specific cybersecurity dataset to enhance the base model's capabilities. Finally, we conducted a comprehensive analysis of the agents' behaviors that provides insights into the planning abilities and potential shortcomings of such agents, contributing to the broader understanding of LLM-based agents in cybersecurity contexts

Community

Paper author Paper submitter

Hackphyr is a fine-tuned model that works as a red team agent in the NetSecGame environment. The paper describes the fine-tuning process, the experiments in different network setups, and the comparisons with LLM agents based on GPT-4 and other models. It also contains a behavioral analysis of the agents' actions that shows that the best models behave in a very reasonable manner that resembles human practices.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2409.11276 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2409.11276 in a Space README.md to link it from this page.

Collections including this paper 1