{"title":"Hackphyr:用于网络安全环境的本地微调 LLM 代理","authors":"Maria Rigaki, Carlos Catania, Sebastian Garcia","doi":"arxiv-2409.11276","DOIUrl":null,"url":null,"abstract":"Large Language Models (LLMs) have shown remarkable potential across various\ndomains, including cybersecurity. Using commercial cloud-based LLMs may be\nundesirable due to privacy concerns, costs, and network connectivity\nconstraints. In this paper, we present Hackphyr, a locally fine-tuned LLM to be\nused as a red-team agent within network security environments. Our fine-tuned 7\nbillion parameter model can run on a single GPU card and achieves performance\ncomparable with much larger and more powerful commercial models such as GPT-4.\nHackphyr clearly outperforms other models, including GPT-3.5-turbo, and\nbaselines, such as Q-learning agents in complex, previously unseen scenarios.\nTo achieve this performance, we generated a new task-specific cybersecurity\ndataset to enhance the base model's capabilities. Finally, we conducted a\ncomprehensive analysis of the agents' behaviors that provides insights into the\nplanning abilities and potential shortcomings of such agents, contributing to\nthe broader understanding of LLM-based agents in cybersecurity contexts","PeriodicalId":501332,"journal":{"name":"arXiv - CS - Cryptography and Security","volume":"89 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Hackphyr: A Local Fine-Tuned LLM Agent for Network Security Environments\",\"authors\":\"Maria Rigaki, Carlos Catania, Sebastian Garcia\",\"doi\":\"arxiv-2409.11276\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Large Language Models (LLMs) have shown remarkable potential across various\\ndomains, including cybersecurity. Using commercial cloud-based LLMs may be\\nundesirable due to privacy concerns, costs, and network connectivity\\nconstraints. In this paper, we present Hackphyr, a locally fine-tuned LLM to be\\nused as a red-team agent within network security environments. Our fine-tuned 7\\nbillion parameter model can run on a single GPU card and achieves performance\\ncomparable with much larger and more powerful commercial models such as GPT-4.\\nHackphyr clearly outperforms other models, including GPT-3.5-turbo, and\\nbaselines, such as Q-learning agents in complex, previously unseen scenarios.\\nTo achieve this performance, we generated a new task-specific cybersecurity\\ndataset to enhance the base model's capabilities. Finally, we conducted a\\ncomprehensive analysis of the agents' behaviors that provides insights into the\\nplanning abilities and potential shortcomings of such agents, contributing to\\nthe broader understanding of LLM-based agents in cybersecurity contexts\",\"PeriodicalId\":501332,\"journal\":{\"name\":\"arXiv - CS - Cryptography and Security\",\"volume\":\"89 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Cryptography and Security\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11276\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Cryptography and Security","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11276","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Hackphyr: A Local Fine-Tuned LLM Agent for Network Security Environments
Large Language Models (LLMs) have shown remarkable potential across various
domains, including cybersecurity. Using commercial cloud-based LLMs may be
undesirable due to privacy concerns, costs, and network connectivity
constraints. In this paper, we present Hackphyr, a locally fine-tuned LLM to be
used as a red-team agent within network security environments. Our fine-tuned 7
billion parameter model can run on a single GPU card and achieves performance
comparable with much larger and more powerful commercial models such as GPT-4.
Hackphyr clearly outperforms other models, including GPT-3.5-turbo, and
baselines, such as Q-learning agents in complex, previously unseen scenarios.
To achieve this performance, we generated a new task-specific cybersecurity
dataset to enhance the base model's capabilities. Finally, we conducted a
comprehensive analysis of the agents' behaviors that provides insights into the
planning abilities and potential shortcomings of such agents, contributing to
the broader understanding of LLM-based agents in cybersecurity contexts