Artificial intelligence (AI) is transforming industries worldwide, and the e-commerce sector is at the forefront of leveraging its capabilities to drive innovation and efficiency. The paper explores the integration of artificial intelligence in e-commerce, focusing on the ethical and regulatory implications introduced by the EU AI Act. This legislative framework aims to ensure the responsible deployment of AI by classifying AI systems into risk categories and imposing compliance requirements. It also underscores both the opportunities and challenges that AI presents to businesses, particularly in enhancing consumer experiences through automation and data-driven decision-making processes. The paper provides a comprehensive review of the AI landscape in Europe, analyzing the impact of the EU AI Act, particularly on small and medium-sized enterprises and startups. Through a mixed-methods approach, the study investigates how regulatory compliance may influence business innovation, market competitiveness, and consumer trust. The recommendations proposed aim to develop a trustworthy AI ecosystem that could stimulate long-term growth and enhance the global positioning of small European businesses.
{"title":"Artificial intelligence and the impact of the EU AI Act in business organizations","authors":"Marc Selgas Cors, Renata Thiébaut","doi":"10.1002/aaai.70039","DOIUrl":"https://doi.org/10.1002/aaai.70039","url":null,"abstract":"<p>Artificial intelligence (AI) is transforming industries worldwide, and the e-commerce sector is at the forefront of leveraging its capabilities to drive innovation and efficiency. The paper explores the integration of artificial intelligence in e-commerce, focusing on the ethical and regulatory implications introduced by the EU AI Act. This legislative framework aims to ensure the responsible deployment of AI by classifying AI systems into risk categories and imposing compliance requirements. It also underscores both the opportunities and challenges that AI presents to businesses, particularly in enhancing consumer experiences through automation and data-driven decision-making processes. The paper provides a comprehensive review of the AI landscape in Europe, analyzing the impact of the EU AI Act, particularly on small and medium-sized enterprises and startups. Through a mixed-methods approach, the study investigates how regulatory compliance may influence business innovation, market competitiveness, and consumer trust. The recommendations proposed aim to develop a trustworthy AI ecosystem that could stimulate long-term growth and enhance the global positioning of small European businesses.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70039","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145469537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Agentic AI shifts stacks from request-response to plan-execute. Systems no longer just answer; they act—planning tasks, calling tools, keeping memory, and changing external state. That shift moves privacy from policy docs into the runtime. This opinion piece argues that we do not need a new privacy theory for agents; we need enforceable, observable controls that render existing rights as product behavior. Anchoring on GDPR—with portable touchpoints to CPRA, LGPD, and PDPA, we propose a developer-first toolkit: optional, bounded, user-visible memory; a purpose-aware egress gate that enforces minimization and transfer rules; proportional safeguards that scale with stakes; and traces that tell a coherent story across components and suppliers. We show how the EU AI Act's risk management, logging, and oversight can scaffold these controls and enable evidence reuse. The result is an agentic runtime that keeps people in control and teams audit-ready by design.
{"title":"From rights to runtime: Privacy engineering for agentic AI","authors":"Keivan Navaie","doi":"10.1002/aaai.70036","DOIUrl":"https://doi.org/10.1002/aaai.70036","url":null,"abstract":"<p>Agentic AI shifts stacks from request-response to plan-execute. Systems no longer just answer; they act—planning tasks, calling tools, keeping memory, and changing external state. That shift moves privacy from policy docs into the runtime. This opinion piece argues that we do not need a new privacy theory for agents; we need enforceable, observable controls that render existing rights as product behavior. Anchoring on GDPR—with portable touchpoints to CPRA, LGPD, and PDPA, we propose a developer-first toolkit: optional, bounded, user-visible memory; a purpose-aware egress gate that enforces minimization and transfer rules; proportional safeguards that scale with stakes; and traces that tell a coherent story across components and suppliers. We show how the EU AI Act's risk management, logging, and oversight can scaffold these controls and enable evidence reuse. The result is an agentic runtime that keeps people in control and teams audit-ready by design.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70036","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145367007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Vinay K Chaudhri, Chaitan Baru, Brandon Bennett, Mehul Bhatt, Darion Cassel, Anthony G Cohn, Rina Dechter, Esra Erdem, Dave Ferrucci, Ken Forbus, Gregory Gelfond, Michael Genesereth, Andrew S. Gordon, Benjamin Grosof, Gopal Gupta, Jim Hendler, Sharat Israni, Tyler R. Josephson, Patrick Kyllonen, Yuliya Lierler, Vladimir Lifschitz, Clifton McFate, Hande Küçük McGinty, Leora Morgenstern, Alessandro Oltramari, Praveen Paritosh, Dan Roth, Blake Shepard, Cogan Shimizu, Denny Vrandečić, Mark Whiting, Michael Witbrock
The long-standing goal of creating a comprehensive, multi-purpose knowledge resource, reminiscent of the 1984 Cyc project, still persists in AI. Despite the success of knowledge resources like WordNet, ConceptNet, Wolfram|Alpha and other commercial knowledge graphs, verifiable, general-purpose, widely available sources of knowledge remain a critical deficiency in AI infrastructure. Large language models struggle due to knowledge gaps; robotic planning lacks necessary world knowledge; and the detection of factually false information relies heavily on human expertise. What kind of knowledge resource is most needed in AI today? How can modern technology shape its development and evaluation? A recent AAAI workshop gathered over 50 researchers to explore these questions. This paper synthesizes our findings and outlines a community-driven vision for a new knowledge infrastructure. In addition to leveraging contemporary advances in knowledge representation and reasoning, one promising idea is to build an open engineering framework to exploit knowledge modules effectively within the context of practical applications. Such a framework should include sets of conventions and social structures that are adopted by contributors.
{"title":"A community-driven vision for a new knowledge resource for AI","authors":"Vinay K Chaudhri, Chaitan Baru, Brandon Bennett, Mehul Bhatt, Darion Cassel, Anthony G Cohn, Rina Dechter, Esra Erdem, Dave Ferrucci, Ken Forbus, Gregory Gelfond, Michael Genesereth, Andrew S. Gordon, Benjamin Grosof, Gopal Gupta, Jim Hendler, Sharat Israni, Tyler R. Josephson, Patrick Kyllonen, Yuliya Lierler, Vladimir Lifschitz, Clifton McFate, Hande Küçük McGinty, Leora Morgenstern, Alessandro Oltramari, Praveen Paritosh, Dan Roth, Blake Shepard, Cogan Shimizu, Denny Vrandečić, Mark Whiting, Michael Witbrock","doi":"10.1002/aaai.70035","DOIUrl":"https://doi.org/10.1002/aaai.70035","url":null,"abstract":"<p>The long-standing goal of creating a comprehensive, multi-purpose knowledge resource, reminiscent of the 1984 Cyc project, still persists in AI. Despite the success of knowledge resources like WordNet, ConceptNet, Wolfram|Alpha and other commercial knowledge graphs, verifiable, general-purpose, widely available sources of knowledge remain a critical deficiency in AI infrastructure. Large language models struggle due to knowledge gaps; robotic planning lacks necessary world knowledge; and the detection of factually false information relies heavily on human expertise. What kind of knowledge resource is most needed in AI today? How can modern technology shape its development and evaluation? A recent AAAI workshop gathered over 50 researchers to explore these questions. This paper synthesizes our findings and outlines a community-driven vision for a new knowledge infrastructure. In addition to leveraging contemporary advances in knowledge representation and reasoning, one promising idea is to build an open engineering framework to exploit knowledge modules effectively within the context of practical applications. Such a framework should include sets of conventions and social structures that are adopted by contributors.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 4","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70035","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145367001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mitchell Kiely, Metin Ahiskali, Etienne Borde, Benjamin Bowman, David Bowman, Dirk Van Bruggen, K. C. Cowan, Prithviraj Dasgupta, Erich Devendorf, Ben Edwards, Alex Fitts, Sunny Fugate, Ryan Gabrys, Wayne Gould, H. Howie Huang, Jules Jacobs, Ryan Kerr, Isaiah J. King, Li Li, Luis Martinez, Christopher Moir, Craig Murphy, Olivia Naish, Claire Owens, Miranda Purchase, Ahmad Ridley, Adrian Taylor, Sara Farmer, William John Valentine, Yiyi Zhang
As cyber threats become increasingly automated and sophisticated, novel solutions must be introduced to improve defense of enterprise networks. Deep reinforcement learning (DRL) has demonstrated potential in mitigating these advanced threats. Single DRL agents have proven utility toward execution of autonomous cyber defense. Despite the success of employing single DRL agents, this approach presents significant limitations, especially regarding scalability within large enterprise networks. An attractive alternative to the single-agent approach is the use of multi-agent reinforcement learning (MARL). However, developing MARL agents is costly with few options for examining MARL cyber defense techniques against adversarial agents. This paper presents a MARL network security environment, the fourth iteration of the cyber autonomy gym for experimentation (CAGE) challenges. This challenge was specifically designed to test the efficacy of MARL algorithms in an enterprise network. Our work aims to evaluate the potential of MARL as a robust and scalable solution for autonomous network defense.
{"title":"CAGE challenge 4: A scalable multi-agent reinforcement learning gym for autonomous cyber defence","authors":"Mitchell Kiely, Metin Ahiskali, Etienne Borde, Benjamin Bowman, David Bowman, Dirk Van Bruggen, K. C. Cowan, Prithviraj Dasgupta, Erich Devendorf, Ben Edwards, Alex Fitts, Sunny Fugate, Ryan Gabrys, Wayne Gould, H. Howie Huang, Jules Jacobs, Ryan Kerr, Isaiah J. King, Li Li, Luis Martinez, Christopher Moir, Craig Murphy, Olivia Naish, Claire Owens, Miranda Purchase, Ahmad Ridley, Adrian Taylor, Sara Farmer, William John Valentine, Yiyi Zhang","doi":"10.1002/aaai.70021","DOIUrl":"https://doi.org/10.1002/aaai.70021","url":null,"abstract":"<p>As cyber threats become increasingly automated and sophisticated, novel solutions must be introduced to improve defense of enterprise networks. Deep reinforcement learning (DRL) has demonstrated potential in mitigating these advanced threats. Single DRL agents have proven utility toward execution of autonomous cyber defense. Despite the success of employing single DRL agents, this approach presents significant limitations, especially regarding scalability within large enterprise networks. An attractive alternative to the single-agent approach is the use of multi-agent reinforcement learning (MARL). However, developing MARL agents is costly with few options for examining MARL cyber defense techniques against adversarial agents. This paper presents a MARL network security environment, the fourth iteration of the cyber autonomy gym for experimentation (CAGE) challenges. This challenge was specifically designed to test the efficacy of MARL algorithms in an enterprise network. Our work aims to evaluate the potential of MARL as a robust and scalable solution for autonomous network defense.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70021","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145366457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Model-based planning and execution systems offer a principled approach to building flexible autonomous robots that can perform diverse tasks by automatically combining a host of basic skills. This idea is almost as old as modern robotics. Yet, while diverse general-purpose reasoning architectures have been proposed since, general-purpose software platforms that support the construction of planner-based controllers and their integration with modern robotic platforms have emerged only recently, starting with the influential ROSPlan system. Since then, a growing number of domain-independent model-based platforms for robot task-level control have emerged. In this paper, we consider the diverse design choices and issues existing platforms attempt to address, the different solutions proposed so far, and suggest avenues for future development. We also briefly discuss the elephant in the room: foundation models.
{"title":"Model-based AI planning and execution platforms for robotics","authors":"Or Wertheim, Ronen I. Brafman","doi":"10.1002/aaai.70034","DOIUrl":"https://doi.org/10.1002/aaai.70034","url":null,"abstract":"<p>Model-based planning and execution systems offer a principled approach to building flexible autonomous robots that can perform diverse tasks by automatically combining a host of basic skills. This idea is almost as old as modern robotics. Yet, while diverse general-purpose reasoning architectures have been proposed since, general-purpose software platforms that support the construction of planner-based controllers and their integration with modern robotic platforms have emerged only recently, starting with the influential ROSPlan system. Since then, a growing number of domain-independent model-based platforms for robot task-level control have emerged. In this paper, we consider the diverse design choices and issues existing platforms attempt to address, the different solutions proposed so far, and suggest avenues for future development. We also briefly discuss the elephant in the room: foundation models.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70034","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145271727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Francesco Leofante, André Artelt, Demetrios Eliades, Anna Korre, Francesca Toni, Tim Miller
The AAAI 2025 Bridge on “Explainable AI, Energy and Critical Infrastructure Systems” was held at the Pennsylvania Convention Centre, Philadelphia, Pennsylvania, USA, on February 25, 2025. The bridge gathered researchers and practitioners, bringing together innovation research across explainable AI, energy and critical infrastructure systems so they can enhance each other. The Bridge featured five keynote presentations by experts, one tutorial, poster presentations by authors who contributed their research findings, and three breakout sessions to discuss new challenges arising at the intersection of these exciting disciplines.
{"title":"Explainable AI, energy and critical infrastructure systems","authors":"Francesco Leofante, André Artelt, Demetrios Eliades, Anna Korre, Francesca Toni, Tim Miller","doi":"10.1002/aaai.70033","DOIUrl":"https://doi.org/10.1002/aaai.70033","url":null,"abstract":"<p>The AAAI 2025 Bridge on “Explainable AI, Energy and Critical Infrastructure Systems” was held at the Pennsylvania Convention Centre, Philadelphia, Pennsylvania, USA, on February 25, 2025. The bridge gathered researchers and practitioners, bringing together innovation research across explainable AI, energy and critical infrastructure systems so they can enhance each other. The Bridge featured five keynote presentations by experts, one tutorial, poster presentations by authors who contributed their research findings, and three breakout sessions to discuss new challenges arising at the intersection of these exciting disciplines.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70033","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145181548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We present a comprehensive enterprise AI strategy developed within the AI Center of Excellence at Fidelity Investments, emphasizing the strategic integration of open-source AI frameworks into scalable, modular, and reproducible enterprise-grade solutions. Our approach is structured around five key pillars: learning from offline data, learning from online feedback, intelligent decision-making, automated assistants, and responsible AI practices. Through a suite of 12 open-source libraries, we demonstrate how modular and interoperable tools can collectively enhance scalability, fairness, and explainability in real-world AI deployments. We further illustrate the impact of this strategy through three enterprise case studies. Finally, we distill a set of best deployment practices to guide organizations in implementing modular, open-source AI strategies at scale.
{"title":"Open-source AI at scale: Establishing an enterprise AI strategy through modular frameworks","authors":"Serdar Kadıoğlu","doi":"10.1002/aaai.70032","DOIUrl":"https://doi.org/10.1002/aaai.70032","url":null,"abstract":"<p>We present a comprehensive enterprise AI strategy developed within the AI Center of Excellence at Fidelity Investments, emphasizing the strategic integration of open-source AI frameworks into scalable, modular, and reproducible enterprise-grade solutions. Our approach is structured around five key pillars: learning from offline data, learning from online feedback, intelligent decision-making, automated assistants, and responsible AI practices. Through a suite of 12 open-source libraries, we demonstrate how modular and interoperable tools can collectively enhance scalability, fairness, and explainability in real-world AI deployments. We further illustrate the impact of this strategy through three enterprise case studies. Finally, we distill a set of best deployment practices to guide organizations in implementing modular, open-source AI strategies at scale.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70032","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145111174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper extends our previously published work on the virtual AI teacher (VATE) system, presented at IAAI-25. VATE is designed to autonomously analyze and correct student errors in mathematical problem-solving using advanced large language models (LLMs). By incorporating student draft images as a primary input for reasoning, the system provides fine-grained error cause analysis and supports real-time, multi-round AI—student dialogues. In this extended version, we introduce a new snap-to-solve module for handling low-reasoning tasks using edge-deployed LLMs, enabling faster and partially offline interaction. We also include expanded benchmarking experiments, including human expert evaluations and ablation studies, to assess model performance and learning outcomes. Deployed on the Squirrel AI platform, VATE demonstrates high accuracy (78.3%) in error analysis and improves student learning efficiency, with strong user satisfaction. These results suggest that VATE is a scalable, cost-effective solution with the potential to transform educational practices.
{"title":"Multimodal AI Teacher: Integrating Edge Computing and Reasoning Models for Enhanced Student Error Analysis","authors":"Tianlong Xu, Yi-Fan Zhang, Zhendong Chu, Qingsong Wen","doi":"10.1002/aaai.70030","DOIUrl":"https://doi.org/10.1002/aaai.70030","url":null,"abstract":"<p>This paper extends our previously published work on the virtual AI teacher (VATE) system, presented at IAAI-25. VATE is designed to autonomously analyze and correct student errors in mathematical problem-solving using advanced large language models (LLMs). By incorporating student draft images as a primary input for reasoning, the system provides fine-grained error cause analysis and supports real-time, multi-round AI—student dialogues. In this extended version, we introduce a new snap-to-solve module for handling low-reasoning tasks using edge-deployed LLMs, enabling faster and partially offline interaction. We also include expanded benchmarking experiments, including human expert evaluations and ablation studies, to assess model performance and learning outcomes. Deployed on the Squirrel AI platform, VATE demonstrates high accuracy (78.3%) in error analysis and improves student learning efficiency, with strong user satisfaction. These results suggest that VATE is a scalable, cost-effective solution with the potential to transform educational practices.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70030","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145102329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rikhiya Ghosh, Hans-Martin von Stockhausen, Martin Schmitt, George Marica Vasile, Sanjeev Kumar Karn, Oladimeji Farri
The National Vulnerability Database (NVD) publishes over a thousand new vulnerabilities monthly, with a projected 25 percent increase in 2024, highlighting the crucial need for rapid vulnerability identification to mitigate cybersecurity attacks and save costs and resources. In this work, we propose using large language models (LLMs) to learn vulnerability evaluation from historical assessments of medical device vulnerabilities in a single manufacturer's portfolio. We highlight the effectiveness and challenges of using LLMs for automatic vulnerability evaluation and introduce a method to enrich historical data with cybersecurity ontologies, enabling the system to understand new vulnerabilities without retraining the LLM. Our LLM system integrates with the in-house application—Cybersecurity Management System (CSMS)—to help Siemens Healthineers (SHS) product cybersecurity experts efficiently assess the vulnerabilities in our products. Also, we present a comprehensive set of experiments that helps showcase the properties of the LLM and dataset, the various guardrails we have implemented to safeguard the system in production, and the guidelines for efficient integration of LLMs into the cybersecurity tool.
{"title":"Automated vulnerability evaluation with large language models and vulnerability ontologies","authors":"Rikhiya Ghosh, Hans-Martin von Stockhausen, Martin Schmitt, George Marica Vasile, Sanjeev Kumar Karn, Oladimeji Farri","doi":"10.1002/aaai.70031","DOIUrl":"https://doi.org/10.1002/aaai.70031","url":null,"abstract":"<p>The National Vulnerability Database (NVD) publishes over a thousand new vulnerabilities monthly, with a projected 25 percent increase in 2024, highlighting the crucial need for rapid vulnerability identification to mitigate cybersecurity attacks and save costs and resources. In this work, we propose using large language models (LLMs) to learn vulnerability evaluation from historical assessments of medical device vulnerabilities in a single manufacturer's portfolio. We highlight the effectiveness and challenges of using LLMs for automatic vulnerability evaluation and introduce a method to enrich historical data with cybersecurity ontologies, enabling the system to understand new vulnerabilities without retraining the LLM. Our LLM system integrates with the in-house application—Cybersecurity Management System (CSMS)—to help Siemens Healthineers (SHS) product cybersecurity experts efficiently assess the vulnerabilities in our products. Also, we present a comprehensive set of experiments that helps showcase the properties of the LLM and dataset, the various guardrails we have implemented to safeguard the system in production, and the guidelines for efficient integration of LLMs into the cybersecurity tool.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70031","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145057736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Evana Gizzi, Connor Firth, Caleb Adams, James Berck, P. Timothy Chase Jr, Christian Cassamajor-Paul, Rachael Chertok, Lily Clough, Jonathan Davis, Melissa De La Cruz, Matthew Dosberg, Alan Gibson, Jonathan Hammer, Ibrahim Haroon, Michael A. Johnson, Brian Kempa, James Marshall, Patrick Maynard, Brett McKinney, Leyton McKinney, Michael Monaghan, Robin Onsay, Hayley Owens, Sam Pedrotty, Daniel Rogers, Mahmooda Sultana, Jivko Sinapov, Bethany Theiling, Aaron Woodard, Caroline Zouloumian, Connor Williams
Infusing artificial intelligence algorithms into production aerospace systems can be challenging due to costs, timelines, and a risk-averse industry. We introduce the Onboard Artificial Intelligence Research (OnAIR) platform, an open-source software pipeline and cognitive architecture tool that enables full life cycle AI research for on-board intelligent systems. We begin with a description and user walk-through of the OnAIR tool. Next, we describe four use cases of OnAIR for both research and deployed onboard applications, detailing their use of OnAIR and the benefits it provided to the development and function of each respective scenario. Lastly, we describe two upcoming planned deployments which will leverage OnAIR for crucial mission outcomes. We conclude with remarks on future work and goals for the forward progression of OnAIR as a tool to enable a larger AI and aerospace research community.
{"title":"OnAIR: Applications of the NASA on-board artificial intelligence research platform","authors":"Evana Gizzi, Connor Firth, Caleb Adams, James Berck, P. Timothy Chase Jr, Christian Cassamajor-Paul, Rachael Chertok, Lily Clough, Jonathan Davis, Melissa De La Cruz, Matthew Dosberg, Alan Gibson, Jonathan Hammer, Ibrahim Haroon, Michael A. Johnson, Brian Kempa, James Marshall, Patrick Maynard, Brett McKinney, Leyton McKinney, Michael Monaghan, Robin Onsay, Hayley Owens, Sam Pedrotty, Daniel Rogers, Mahmooda Sultana, Jivko Sinapov, Bethany Theiling, Aaron Woodard, Caroline Zouloumian, Connor Williams","doi":"10.1002/aaai.70020","DOIUrl":"https://doi.org/10.1002/aaai.70020","url":null,"abstract":"<p>Infusing artificial intelligence algorithms into production aerospace systems can be challenging due to costs, timelines, and a risk-averse industry. We introduce the Onboard Artificial Intelligence Research (OnAIR) platform, an open-source software pipeline and cognitive architecture tool that enables full life cycle AI research for on-board intelligent systems. We begin with a description and user walk-through of the OnAIR tool. Next, we describe four use cases of OnAIR for both research and deployed onboard applications, detailing their use of OnAIR and the benefits it provided to the development and function of each respective scenario. Lastly, we describe two upcoming planned deployments which will leverage OnAIR for crucial mission outcomes. We conclude with remarks on future work and goals for the forward progression of OnAIR as a tool to enable a larger AI and aerospace research community.</p>","PeriodicalId":7854,"journal":{"name":"Ai Magazine","volume":"46 3","pages":""},"PeriodicalIF":3.2,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/aaai.70020","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145057735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}