{"title":"A novel system for strengthening security in large language models against hallucination and injection attacks with effective strategies","authors":"Tunahan Gokcimen , Bihter Das","doi":"10.1016/j.aej.2025.03.030","DOIUrl":null,"url":null,"abstract":"<div><div>To address the escalating demand for secure and trustworthy interactions with Large Language Models (LLMs), this study introduces a pioneering security framework that mitigates critical vulnerabilities, including injection attacks, hallucinations, and data privacy breaches. By incorporating advanced technologies such as VectorDB, Kernel, and Retrieval-Augmented Generation (RAG) within a cross-LLM architecture, the system delivers enhanced resilience and adaptability to adversarial scenarios. Comprehensive evaluations across leading models—including PaLM, Llama, GPT-3.5, GPT-4, Gemini, and GPT-4o—reveal the system’s exceptional performance, achieving a 98 % accuracy in eligibility scoring and outperforming conventional models in both reliability and security. This study underscores the significance of a multi-layered defense mechanism that not only detects and neutralizes threats but also ensures ethical, accurate, and contextually relevant responses. The novel cross-LLM strategy enhances system robustness by leveraging the strengths of multiple models, minimizing inconsistencies and reinforcing output integrity. With its adaptability to emerging linguistic manipulation techniques and compliance with strict ethical standards, the proposed framework establishes a secure, scalable ecosystem for LLM applications. The findings promise transformative impacts across domains such as cybersecurity, multilingual processing, and adaptive threat detection, paving the way for safer and more reliable language model deployments.</div></div>","PeriodicalId":7484,"journal":{"name":"alexandria engineering journal","volume":"123 ","pages":"Pages 71-90"},"PeriodicalIF":6.8000,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"alexandria engineering journal","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S111001682500328X","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
To address the escalating demand for secure and trustworthy interactions with Large Language Models (LLMs), this study introduces a pioneering security framework that mitigates critical vulnerabilities, including injection attacks, hallucinations, and data privacy breaches. By incorporating advanced technologies such as VectorDB, Kernel, and Retrieval-Augmented Generation (RAG) within a cross-LLM architecture, the system delivers enhanced resilience and adaptability to adversarial scenarios. Comprehensive evaluations across leading models—including PaLM, Llama, GPT-3.5, GPT-4, Gemini, and GPT-4o—reveal the system’s exceptional performance, achieving a 98 % accuracy in eligibility scoring and outperforming conventional models in both reliability and security. This study underscores the significance of a multi-layered defense mechanism that not only detects and neutralizes threats but also ensures ethical, accurate, and contextually relevant responses. The novel cross-LLM strategy enhances system robustness by leveraging the strengths of multiple models, minimizing inconsistencies and reinforcing output integrity. With its adaptability to emerging linguistic manipulation techniques and compliance with strict ethical standards, the proposed framework establishes a secure, scalable ecosystem for LLM applications. The findings promise transformative impacts across domains such as cybersecurity, multilingual processing, and adaptive threat detection, paving the way for safer and more reliable language model deployments.
期刊介绍:
Alexandria Engineering Journal is an international journal devoted to publishing high quality papers in the field of engineering and applied science. Alexandria Engineering Journal is cited in the Engineering Information Services (EIS) and the Chemical Abstracts (CA). The papers published in Alexandria Engineering Journal are grouped into five sections, according to the following classification:
• Mechanical, Production, Marine and Textile Engineering
• Electrical Engineering, Computer Science and Nuclear Engineering
• Civil and Architecture Engineering
• Chemical Engineering and Applied Sciences
• Environmental Engineering