A novel system for strengthening security in large language models against hallucination and injection attacks with effective strategies

IF 6.8 2区 工程技术 Q1 ENGINEERING, MULTIDISCIPLINARY alexandria engineering journal Pub Date : 2025-03-22 DOI:10.1016/j.aej.2025.03.030
Tunahan Gokcimen , Bihter Das
{"title":"A novel system for strengthening security in large language models against hallucination and injection attacks with effective strategies","authors":"Tunahan Gokcimen ,&nbsp;Bihter Das","doi":"10.1016/j.aej.2025.03.030","DOIUrl":null,"url":null,"abstract":"<div><div>To address the escalating demand for secure and trustworthy interactions with Large Language Models (LLMs), this study introduces a pioneering security framework that mitigates critical vulnerabilities, including injection attacks, hallucinations, and data privacy breaches. By incorporating advanced technologies such as VectorDB, Kernel, and Retrieval-Augmented Generation (RAG) within a cross-LLM architecture, the system delivers enhanced resilience and adaptability to adversarial scenarios. Comprehensive evaluations across leading models—including PaLM, Llama, GPT-3.5, GPT-4, Gemini, and GPT-4o—reveal the system’s exceptional performance, achieving a 98 % accuracy in eligibility scoring and outperforming conventional models in both reliability and security. This study underscores the significance of a multi-layered defense mechanism that not only detects and neutralizes threats but also ensures ethical, accurate, and contextually relevant responses. The novel cross-LLM strategy enhances system robustness by leveraging the strengths of multiple models, minimizing inconsistencies and reinforcing output integrity. With its adaptability to emerging linguistic manipulation techniques and compliance with strict ethical standards, the proposed framework establishes a secure, scalable ecosystem for LLM applications. The findings promise transformative impacts across domains such as cybersecurity, multilingual processing, and adaptive threat detection, paving the way for safer and more reliable language model deployments.</div></div>","PeriodicalId":7484,"journal":{"name":"alexandria engineering journal","volume":"123 ","pages":"Pages 71-90"},"PeriodicalIF":6.8000,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"alexandria engineering journal","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S111001682500328X","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

To address the escalating demand for secure and trustworthy interactions with Large Language Models (LLMs), this study introduces a pioneering security framework that mitigates critical vulnerabilities, including injection attacks, hallucinations, and data privacy breaches. By incorporating advanced technologies such as VectorDB, Kernel, and Retrieval-Augmented Generation (RAG) within a cross-LLM architecture, the system delivers enhanced resilience and adaptability to adversarial scenarios. Comprehensive evaluations across leading models—including PaLM, Llama, GPT-3.5, GPT-4, Gemini, and GPT-4o—reveal the system’s exceptional performance, achieving a 98 % accuracy in eligibility scoring and outperforming conventional models in both reliability and security. This study underscores the significance of a multi-layered defense mechanism that not only detects and neutralizes threats but also ensures ethical, accurate, and contextually relevant responses. The novel cross-LLM strategy enhances system robustness by leveraging the strengths of multiple models, minimizing inconsistencies and reinforcing output integrity. With its adaptability to emerging linguistic manipulation techniques and compliance with strict ethical standards, the proposed framework establishes a secure, scalable ecosystem for LLM applications. The findings promise transformative impacts across domains such as cybersecurity, multilingual processing, and adaptive threat detection, paving the way for safer and more reliable language model deployments.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
一种新的增强大型语言模型安全性的系统,具有有效的策略来抵御幻觉和注入攻击
为了解决对大型语言模型(llm)安全可靠交互的日益增长的需求,本研究引入了一个开创性的安全框架,以减轻关键漏洞,包括注入攻击、幻觉和数据隐私泄露。通过在跨llm架构中结合先进的技术,如VectorDB、Kernel和检索增强生成(RAG),该系统提供了增强的弹性和对抗性场景的适应性。对包括PaLM、Llama、GPT-3.5、GPT-4、Gemini和gpt - 40在内的领先模型的综合评估显示了该系统的卓越性能,在合格评分方面达到了98% %的准确率,在可靠性和安全性方面都优于传统模型。该研究强调了多层防御机制的重要性,该机制不仅可以检测和消除威胁,还可以确保道德、准确和与上下文相关的反应。新的跨llm策略通过利用多个模型的优势、最小化不一致性和增强输出完整性来增强系统的鲁棒性。由于其对新兴语言操作技术的适应性和严格的道德标准的遵守,所提出的框架为法学硕士应用程序建立了一个安全的、可扩展的生态系统。这些发现有望在网络安全、多语言处理和自适应威胁检测等领域产生变革性影响,为更安全、更可靠的语言模型部署铺平道路。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
alexandria engineering journal
alexandria engineering journal Engineering-General Engineering
CiteScore
11.20
自引率
4.40%
发文量
1015
审稿时长
43 days
期刊介绍: Alexandria Engineering Journal is an international journal devoted to publishing high quality papers in the field of engineering and applied science. Alexandria Engineering Journal is cited in the Engineering Information Services (EIS) and the Chemical Abstracts (CA). The papers published in Alexandria Engineering Journal are grouped into five sections, according to the following classification: • Mechanical, Production, Marine and Textile Engineering • Electrical Engineering, Computer Science and Nuclear Engineering • Civil and Architecture Engineering • Chemical Engineering and Applied Sciences • Environmental Engineering
期刊最新文献
ST-Former: A transformer-based temporal-scene fusion-driven auditory experience analysis model ET-YOLO:A study on a malaria pathogen detection model based on YOLO11 From words to proverbs: Evaluating LLMs’ linguistic and cultural competence in Saudi dialects with Absher Data-Driven approach for the unified influence of media function and information density on the transmission dynamics of SIRS epidemic model via: Disease informed neural networks Editorial Board
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1