{"title":"Models Are Codes: Towards Measuring Malicious Code Poisoning Attacks on Pre-trained Model Hubs","authors":"Jian Zhao, Shenao Wang, Yanjie Zhao, Xinyi Hou, Kailong Wang, Peiming Gao, Yuanchao Zhang, Chen Wei, Haoyu Wang","doi":"arxiv-2409.09368","DOIUrl":null,"url":null,"abstract":"The proliferation of pre-trained models (PTMs) and datasets has led to the\nemergence of centralized model hubs like Hugging Face, which facilitate\ncollaborative development and reuse. However, recent security reports have\nuncovered vulnerabilities and instances of malicious attacks within these\nplatforms, highlighting growing security concerns. This paper presents the\nfirst systematic study of malicious code poisoning attacks on pre-trained model\nhubs, focusing on the Hugging Face platform. We conduct a comprehensive threat\nanalysis, develop a taxonomy of model formats, and perform root cause analysis\nof vulnerable formats. While existing tools like Fickling and ModelScan offer\nsome protection, they face limitations in semantic-level analysis and\ncomprehensive threat detection. To address these challenges, we propose MalHug,\nan end-to-end pipeline tailored for Hugging Face that combines dataset loading\nscript extraction, model deserialization, in-depth taint analysis, and\nheuristic pattern matching to detect and classify malicious code poisoning\nattacks in datasets and models. In collaboration with Ant Group, a leading\nfinancial technology company, we have implemented and deployed MalHug on a\nmirrored Hugging Face instance within their infrastructure, where it has been\noperational for over three months. During this period, MalHug has monitored\nmore than 705K models and 176K datasets, uncovering 91 malicious models and 9\nmalicious dataset loading scripts. These findings reveal a range of security\nthreats, including reverse shell, browser credential theft, and system\nreconnaissance. This work not only bridges a critical gap in understanding the\nsecurity of the PTM supply chain but also provides a practical, industry-tested\nsolution for enhancing the security of pre-trained model hubs.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09368","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The proliferation of pre-trained models (PTMs) and datasets has led to the
emergence of centralized model hubs like Hugging Face, which facilitate
collaborative development and reuse. However, recent security reports have
uncovered vulnerabilities and instances of malicious attacks within these
platforms, highlighting growing security concerns. This paper presents the
first systematic study of malicious code poisoning attacks on pre-trained model
hubs, focusing on the Hugging Face platform. We conduct a comprehensive threat
analysis, develop a taxonomy of model formats, and perform root cause analysis
of vulnerable formats. While existing tools like Fickling and ModelScan offer
some protection, they face limitations in semantic-level analysis and
comprehensive threat detection. To address these challenges, we propose MalHug,
an end-to-end pipeline tailored for Hugging Face that combines dataset loading
script extraction, model deserialization, in-depth taint analysis, and
heuristic pattern matching to detect and classify malicious code poisoning
attacks in datasets and models. In collaboration with Ant Group, a leading
financial technology company, we have implemented and deployed MalHug on a
mirrored Hugging Face instance within their infrastructure, where it has been
operational for over three months. During this period, MalHug has monitored
more than 705K models and 176K datasets, uncovering 91 malicious models and 9
malicious dataset loading scripts. These findings reveal a range of security
threats, including reverse shell, browser credential theft, and system
reconnaissance. This work not only bridges a critical gap in understanding the
security of the PTM supply chain but also provides a practical, industry-tested
solution for enhancing the security of pre-trained model hubs.