首页 > 最新文献

arXiv - CS - Emerging Technologies最新文献

英文 中文
Reinforcement Learning for an Efficient and Effective Malware Investigation during Cyber Incident Response 在网络事件响应期间进行强化学习以实现高效和有效的恶意软件调查
Pub Date : 2024-08-04 DOI: arxiv-2408.01999
Dipo Dunsin, Mohamed Chahine Ghanem, Karim Ouazzane, Vassil Vassilev
This research focused on enhancing post-incident malware forensicinvestigation using reinforcement learning RL. We proposed an advanced MDP postincident malware forensics investigation model and framework to expedite postincident forensics. We then implement our RL Malware Investigation Model basedon structured MDP within the proposed framework. To identify malware artefacts,the RL agent acquires and examines forensics evidence files, iterativelyimproving its capabilities using Q Table and temporal difference learning. TheQ learning algorithm significantly improved the agent ability to identifymalware. An epsilon greedy exploration strategy and Q learning updates enabledefficient learning and decision making. Our experimental testing revealed thatoptimal learning rates depend on the MDP environment complexity, with simplerenvironments benefiting from higher rates for quicker convergence and complexones requiring lower rates for stability. Our model performance in identifyingand classifying malware reduced malware analysis time compared to humanexperts, demonstrating robustness and adaptability. The study highlighted thesignificance of hyper parameter tuning and suggested adaptive strategies forcomplex environments. Our RL based approach produced promising results and isvalidated as an alternative to traditional methods notably by offeringcontinuous learning and adaptation to new and evolving malware threats whichultimately enhance the post incident forensics investigations.
本研究的重点是利用强化学习 RL 增强事故后恶意软件取证调查。我们提出了一种先进的 MDP 事件后恶意软件取证调查模型和框架,以加快事件后取证工作。然后,我们在该框架内实现了基于结构化 MDP 的 RL 恶意软件调查模型。为了识别恶意软件人工制品,RL 代理获取并检查取证证据文件,利用 Q 表和时差学习迭代改进其能力。Q 学习算法大大提高了代理识别恶意软件的能力。ε贪婪探索策略和 Q 学习更新实现了高效的学习和决策。我们的实验测试表明,最佳学习率取决于 MDP 环境的复杂程度,简单的环境需要较高的学习率以加快收敛速度,而复杂的环境则需要较低的学习率以保持稳定。与人类专家相比,我们的模型在识别和分类恶意软件方面的表现缩短了恶意软件分析时间,证明了模型的鲁棒性和适应性。研究强调了超参数调整的重要性,并提出了针对复杂环境的自适应策略。我们基于 RL 的方法取得了可喜的成果,并被证实是传统方法的替代方法,特别是通过提供持续学习和适应新的和不断演变的恶意软件威胁,最终提高了事件后取证调查的效率。
{"title":"Reinforcement Learning for an Efficient and Effective Malware Investigation during Cyber Incident Response","authors":"Dipo Dunsin, Mohamed Chahine Ghanem, Karim Ouazzane, Vassil Vassilev","doi":"arxiv-2408.01999","DOIUrl":"https://doi.org/arxiv-2408.01999","url":null,"abstract":"This research focused on enhancing post-incident malware forensic\u0000investigation using reinforcement learning RL. We proposed an advanced MDP post\u0000incident malware forensics investigation model and framework to expedite post\u0000incident forensics. We then implement our RL Malware Investigation Model based\u0000on structured MDP within the proposed framework. To identify malware artefacts,\u0000the RL agent acquires and examines forensics evidence files, iteratively\u0000improving its capabilities using Q Table and temporal difference learning. The\u0000Q learning algorithm significantly improved the agent ability to identify\u0000malware. An epsilon greedy exploration strategy and Q learning updates enabled\u0000efficient learning and decision making. Our experimental testing revealed that\u0000optimal learning rates depend on the MDP environment complexity, with simpler\u0000environments benefiting from higher rates for quicker convergence and complex\u0000ones requiring lower rates for stability. Our model performance in identifying\u0000and classifying malware reduced malware analysis time compared to human\u0000experts, demonstrating robustness and adaptability. The study highlighted the\u0000significance of hyper parameter tuning and suggested adaptive strategies for\u0000complex environments. Our RL based approach produced promising results and is\u0000validated as an alternative to traditional methods notably by offering\u0000continuous learning and adaptation to new and evolving malware threats which\u0000ultimately enhance the post incident forensics investigations.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141937230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extracting Object Heights From LiDAR & Aerial Imagery 从激光雷达和航空图像中提取物体高度
Pub Date : 2024-08-02 DOI: arxiv-2408.00967
Jesus Guerrero
This work shows a procedural method for extracting object heights from LiDARand aerial imagery. We discuss how to get heights and the future of LiDAR andimagery processing. SOTA object segmentation allows us to take get objectheights with no deep learning background. Engineers will be keeping track ofworld data across generations and reprocessing them. They will be using olderprocedural methods like this paper and newer ones discussed here. SOTA methodsare going beyond analysis and into generative AI. We cover both a proceduralmethodology and the newer ones performed with language models. These includepoint cloud, imagery and text encoding allowing for spatially aware AI.
本作品展示了一种从激光雷达和航空图像中提取物体高度的程序方法。我们讨论了如何获取高度以及激光雷达和图像处理的未来。SOTA 物体分割使我们能够在没有深度学习背景的情况下获取物体高度。工程师们将对世界数据进行跨代跟踪和重新处理。他们将使用像本文这样的旧程序方法和本文讨论的新方法。SOTA 方法正在超越分析,进入生成式人工智能领域。我们既包括程序方法,也包括使用语言模型的新方法。这些方法包括点云、图像和文本编码,可实现空间感知人工智能。
{"title":"Extracting Object Heights From LiDAR & Aerial Imagery","authors":"Jesus Guerrero","doi":"arxiv-2408.00967","DOIUrl":"https://doi.org/arxiv-2408.00967","url":null,"abstract":"This work shows a procedural method for extracting object heights from LiDAR\u0000and aerial imagery. We discuss how to get heights and the future of LiDAR and\u0000imagery processing. SOTA object segmentation allows us to take get object\u0000heights with no deep learning background. Engineers will be keeping track of\u0000world data across generations and reprocessing them. They will be using older\u0000procedural methods like this paper and newer ones discussed here. SOTA methods\u0000are going beyond analysis and into generative AI. We cover both a procedural\u0000methodology and the newer ones performed with language models. These include\u0000point cloud, imagery and text encoding allowing for spatially aware AI.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"22 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141937233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rubric-based Learner Modelling via Noisy Gates Bayesian Networks for Computational Thinking Skills Assessment 通过噪声门贝叶斯网络为计算思维能力评估建立基于评分标准的学习者模型
Pub Date : 2024-08-02 DOI: arxiv-2408.01221
Giorgia Adorni, Francesca Mangili, Alberto Piatti, Claudio Bonesana, Alessandro Antonucci
In modern and personalised education, there is a growing interest indeveloping learners' competencies and accurately assessing them. In a previouswork, we proposed a procedure for deriving a learner model for automatic skillassessment from a task-specific competence rubric, thus simplifying theimplementation of automated assessment tools. The previous approach, however,suffered two main limitations: (i) the ordering between competencies defined bythe assessment rubric was only indirectly modelled; (ii) supplementary skills,not under assessment but necessary for accomplishing the task, were notincluded in the model. In this work, we address issue (i) by introducing dummyobserved nodes, strictly enforcing the skills ordering without changing thenetwork's structure. In contrast, for point (ii), we design a network with twolayers of gates, one performing disjunctive operations by noisy-OR gates andthe other conjunctive operations through logical ANDs. Such changes improve themodel outcomes' coherence and the modelling tool's flexibility withoutcompromising the model's compact parametrisation, interpretability and simpleexperts' elicitation. We used this approach to develop a learner model forComputational Thinking (CT) skills assessment. The CT-cube skills assessmentframework and the Cross Array Task (CAT) are used to exemplify it anddemonstrate its feasibility.
在现代个性化教育中,人们越来越关注培养学习者的能力并对其进行准确评估。在之前的工作中,我们提出了一种从特定任务的能力标准中推导出学习者模型的方法,用于自动技能评估,从而简化了自动评估工具的实施。然而,以前的方法有两个主要局限:(i) 评估标准所定义的能力之间的排序只是间接建模;(ii) 模型中没有包括不在评估范围内但完成任务所必需的辅助技能。在这项工作中,我们通过引入虚拟观察节点来解决第(i)点问题,在不改变网络结构的情况下严格执行技能排序。相反,针对问题(ii),我们设计了一个具有两层门的网络,一层通过噪声-OR 门进行非连接操作,另一层通过逻辑 AND 进行连接操作。这种改变提高了模型结果的一致性和建模工具的灵活性,同时又不影响模型的紧凑参数化、可解释性和简单的专家诱导。我们采用这种方法开发了用于计算思维(CT)技能评估的学习者模型。CT-立方体技能评估框架和交叉阵列任务(CAT)被用来示范和证明其可行性。
{"title":"Rubric-based Learner Modelling via Noisy Gates Bayesian Networks for Computational Thinking Skills Assessment","authors":"Giorgia Adorni, Francesca Mangili, Alberto Piatti, Claudio Bonesana, Alessandro Antonucci","doi":"arxiv-2408.01221","DOIUrl":"https://doi.org/arxiv-2408.01221","url":null,"abstract":"In modern and personalised education, there is a growing interest in\u0000developing learners' competencies and accurately assessing them. In a previous\u0000work, we proposed a procedure for deriving a learner model for automatic skill\u0000assessment from a task-specific competence rubric, thus simplifying the\u0000implementation of automated assessment tools. The previous approach, however,\u0000suffered two main limitations: (i) the ordering between competencies defined by\u0000the assessment rubric was only indirectly modelled; (ii) supplementary skills,\u0000not under assessment but necessary for accomplishing the task, were not\u0000included in the model. In this work, we address issue (i) by introducing dummy\u0000observed nodes, strictly enforcing the skills ordering without changing the\u0000network's structure. In contrast, for point (ii), we design a network with two\u0000layers of gates, one performing disjunctive operations by noisy-OR gates and\u0000the other conjunctive operations through logical ANDs. Such changes improve the\u0000model outcomes' coherence and the modelling tool's flexibility without\u0000compromising the model's compact parametrisation, interpretability and simple\u0000experts' elicitation. We used this approach to develop a learner model for\u0000Computational Thinking (CT) skills assessment. The CT-cube skills assessment\u0000framework and the Cross Array Task (CAT) are used to exemplify it and\u0000demonstrate its feasibility.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"6 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141937236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analyzing Quantum Circuit Depth Reduction with Ancilla Qubits in MCX Gates 用 MCX 栅极中的 Ancilla Qubits 分析量子电路深度缩减情况
Pub Date : 2024-08-02 DOI: arxiv-2408.01304
Ahmad Bennakhi, Paul Franzon, Gregory T. Byrd
This paper aims to give readers a high-level overview of the different MCXdepth reduction techniques that utilize ancilla qubits. We also exhibit a briefanalysis of how they would perform under different quantum topologicalsettings. The techniques examined are recursion and v-chain, as they are themost commonly used techniques in the most popular quantum computing libraries,Qiskit. The target audience of this paper is people who do not have intricatemathematical or physics knowledge related to quantum computing.
本文旨在为读者提供一个利用 ancilla 量子比特的不同 MCX 深度缩减技术的高层次概览。我们还简要分析了这些技术在不同量子拓扑结构下的表现。我们研究的技术是递归和 v 链,因为它们是最流行的量子计算库 Qiskit 中最常用的技术。本文的目标读者是那些不具备与量子计算相关的复杂数学或物理知识的人。
{"title":"Analyzing Quantum Circuit Depth Reduction with Ancilla Qubits in MCX Gates","authors":"Ahmad Bennakhi, Paul Franzon, Gregory T. Byrd","doi":"arxiv-2408.01304","DOIUrl":"https://doi.org/arxiv-2408.01304","url":null,"abstract":"This paper aims to give readers a high-level overview of the different MCX\u0000depth reduction techniques that utilize ancilla qubits. We also exhibit a brief\u0000analysis of how they would perform under different quantum topological\u0000settings. The techniques examined are recursion and v-chain, as they are the\u0000most commonly used techniques in the most popular quantum computing libraries,\u0000Qiskit. The target audience of this paper is people who do not have intricate\u0000mathematical or physics knowledge related to quantum computing.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"29 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141937286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Energy Cost of Artificial Intelligence of Things Lifecycle 人工智能物联网生命周期的能源成本
Pub Date : 2024-08-01 DOI: arxiv-2408.00540
Shih-Kai Chou, Jernej Hribar, Mihael Mohorčič, Carolina Fortuna
Artificial intelligence (AI)coupled with existing Internet of Things (IoT)enables more streamlined and autonomous operations across various economicsectors. Consequently, the paradigm of Artificial Intelligence of Things (AIoT)having AI techniques at its core implies additional energy and carbon coststhat may become significant with more complex neural architectures. To betterunderstand the energy and Carbon Footprint (CF) of some AIoT components, veryrecent studies employ conventional metrics. However, these metrics are notdesigned to capture energy efficiency aspects of inference. In this paper, wepropose a new metric, the Energy Cost of AIoT Lifecycle (eCAL) to capture theoverall energy cost of inference over the lifecycle of an AIoT system. Wedevise a new methodology for determining eCAL of an AIoT system by analyzingthe complexity of data manipulation in individual components involved in theAIoT lifecycle and derive the overall and per bit energy consumption. With eCALwe show that the better a model is and the more it is used, the more energyefficient an inference is. For an example AIoT configuration, eCAL for making$100$ inferences is $1.43$ times higher than for $1000$ inferences. We alsoevaluate the CF of the AIoT system by calculating the equivalent CO$_{2}$emissions based on the energy consumption and the Carbon Intensity (CI) acrossdifferent countries. Using 2023 renewable data, our analysis reveals thatdeploying an AIoT system in Germany results in emitting $4.62$ times higherCO$_2$ than in Finland, due to latter using more low-CI energy sources.
人工智能(AI)与现有的物联网(IoT)相结合,使各经济部门的运作更加合理和自主。因此,以人工智能技术为核心的人工智能物联网(AIoT)模式意味着额外的能源和碳成本,而随着神经架构变得更加复杂,这些成本可能会变得非常高昂。为了更好地了解一些人工智能物联网组件的能源和碳足迹(CF),最近的研究采用了传统的指标。然而,这些指标并不是为了捕捉推理的能效方面而设计的。在本文中,我们提出了一个新指标--人工智能物联网生命周期能源成本(eCAL),以捕捉人工智能物联网系统生命周期内推理的总体能源成本。我们通过分析参与人工智能物联网生命周期的各个组件中数据操作的复杂性,提出了一种确定人工智能物联网系统 eCAL 的新方法,并推导出整体能耗和每比特能耗。eCAL 表明,模型越好、使用越多,推理的能效就越高。在一个 AIoT 配置示例中,进行 100 美元推理的 eCAL 是进行 1000 美元推理的 1.43 美元。我们还通过计算不同国家基于能源消耗和碳强度(CI)的等效 CO$_{2}$ 排放量来评估 AIoT 系统的 CF。利用 2023 年的可再生数据,我们的分析表明,在德国部署 AIoT 系统的 CO$_{2}$ 排放量是芬兰的 4.62$ 倍,原因是后者使用了更多低碳强度能源。
{"title":"The Energy Cost of Artificial Intelligence of Things Lifecycle","authors":"Shih-Kai Chou, Jernej Hribar, Mihael Mohorčič, Carolina Fortuna","doi":"arxiv-2408.00540","DOIUrl":"https://doi.org/arxiv-2408.00540","url":null,"abstract":"Artificial intelligence (AI)coupled with existing Internet of Things (IoT)\u0000enables more streamlined and autonomous operations across various economic\u0000sectors. Consequently, the paradigm of Artificial Intelligence of Things (AIoT)\u0000having AI techniques at its core implies additional energy and carbon costs\u0000that may become significant with more complex neural architectures. To better\u0000understand the energy and Carbon Footprint (CF) of some AIoT components, very\u0000recent studies employ conventional metrics. However, these metrics are not\u0000designed to capture energy efficiency aspects of inference. In this paper, we\u0000propose a new metric, the Energy Cost of AIoT Lifecycle (eCAL) to capture the\u0000overall energy cost of inference over the lifecycle of an AIoT system. We\u0000devise a new methodology for determining eCAL of an AIoT system by analyzing\u0000the complexity of data manipulation in individual components involved in the\u0000AIoT lifecycle and derive the overall and per bit energy consumption. With eCAL\u0000we show that the better a model is and the more it is used, the more energy\u0000efficient an inference is. For an example AIoT configuration, eCAL for making\u0000$100$ inferences is $1.43$ times higher than for $1000$ inferences. We also\u0000evaluate the CF of the AIoT system by calculating the equivalent CO$_{2}$\u0000emissions based on the energy consumption and the Carbon Intensity (CI) across\u0000different countries. Using 2023 renewable data, our analysis reveals that\u0000deploying an AIoT system in Germany results in emitting $4.62$ times higher\u0000CO$_2$ than in Finland, due to latter using more low-CI energy sources.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"30 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141881507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Generation of Behavioral Test Cases For Natural Language Processing Using Clustering and Prompting 利用聚类和提示自动生成自然语言处理行为测试用例
Pub Date : 2024-07-31 DOI: arxiv-2408.00161
Ying Li, Rahul Singh, Tarun Joshi, Agus Sudjianto
Recent work in behavioral testing for natural language processing (NLP)models, such as Checklist, is inspired by related paradigms in softwareengineering testing. They allow evaluation of general linguistic capabilitiesand domain understanding, hence can help evaluate conceptual soundness andidentify model weaknesses. However, a major challenge is the creation of testcases. The current packages rely on semi-automated approach using manualdevelopment which requires domain expertise and can be time consuming. Thispaper introduces an automated approach to develop test cases by exploiting thepower of large language models and statistical techniques. It clusters the textrepresentations to carefully construct meaningful groups and then applyprompting techniques to automatically generate Minimal Functionality Tests(MFT). The well-known Amazon Reviews corpus is used to demonstrate ourapproach. We analyze the behavioral test profiles across four differentclassification algorithms and discuss the limitations and strengths of thosemodels.
自然语言处理(NLP)模型的行为测试(如核对表)方面的最新研究受到了软件工程测试相关范例的启发。它们允许对一般语言能力和领域理解进行评估,因此有助于评估概念的合理性和识别模型的弱点。然而,创建测试用例是一项重大挑战。当前的软件包依赖于使用人工开发的半自动化方法,这需要领域专业知识,而且可能很耗时。本文介绍了一种利用大型语言模型和统计技术开发测试用例的自动化方法。它对文本表述进行聚类,仔细构建有意义的组,然后应用提示技术自动生成最小功能测试(MFT)。著名的亚马逊评论语料库被用来演示我们的方法。我们分析了四种不同分类算法的行为测试概况,并讨论了这些模型的局限性和优势。
{"title":"Automatic Generation of Behavioral Test Cases For Natural Language Processing Using Clustering and Prompting","authors":"Ying Li, Rahul Singh, Tarun Joshi, Agus Sudjianto","doi":"arxiv-2408.00161","DOIUrl":"https://doi.org/arxiv-2408.00161","url":null,"abstract":"Recent work in behavioral testing for natural language processing (NLP)\u0000models, such as Checklist, is inspired by related paradigms in software\u0000engineering testing. They allow evaluation of general linguistic capabilities\u0000and domain understanding, hence can help evaluate conceptual soundness and\u0000identify model weaknesses. However, a major challenge is the creation of test\u0000cases. The current packages rely on semi-automated approach using manual\u0000development which requires domain expertise and can be time consuming. This\u0000paper introduces an automated approach to develop test cases by exploiting the\u0000power of large language models and statistical techniques. It clusters the text\u0000representations to carefully construct meaningful groups and then apply\u0000prompting techniques to automatically generate Minimal Functionality Tests\u0000(MFT). The well-known Amazon Reviews corpus is used to demonstrate our\u0000approach. We analyze the behavioral test profiles across four different\u0000classification algorithms and discuss the limitations and strengths of those\u0000models.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"35 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141881505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Explainable Vision Transformer with Transfer Learning Combined with Support Vector Machine Based Efficient Drought Stress Identification 基于支持向量机的可解释视觉变换器与迁移学习相结合的高效干旱压力识别技术
Pub Date : 2024-07-31 DOI: arxiv-2407.21666
Aswini Kumar Patra, Ankit Varshney, Lingaraj Sahoo
Early detection of drought stress is critical for taking timely measures forreducing crop loss before the drought impact becomes irreversible. The subtlephenotypical and physiological changes in response to drought stress arecaptured by non-invasive imaging techniques and these imaging data serve asvaluable resource for machine learning methods to identify drought stress.While convolutional neural networks (CNNs) are in wide use, vision transformers(ViTs) present a promising alternative in capturing long-range dependencies andintricate spatial relationships, thereby enhancing the detection of subtleindicators of drought stress. We propose an explainable deep learning pipelinethat leverages the power of ViTs for drought stress detection in potato cropsusing aerial imagery. We applied two distinct approaches: a synergisticcombination of ViT and support vector machine (SVM), where ViT extractsintricate spatial features from aerial images, and SVM classifies the crops asstressed or healthy and an end-to-end approach using a dedicated classificationlayer within ViT to directly detect drought stress. Our key findings explainthe ViT model's decision-making process by visualizing attention maps. Thesemaps highlight the specific spatial features within the aerial images that theViT model focuses as the drought stress signature. Our findings demonstratethat the proposed methods not only achieve high accuracy in drought stressidentification but also shedding light on the diverse subtle plant featuresassociated with drought stress. This offers a robust and interpretable solutionfor drought stress monitoring for farmers to undertake informed decisions forimproved crop management.
早期检测干旱胁迫对于在干旱影响变得不可逆转之前及时采取措施减少作物损失至关重要。虽然卷积神经网络(CNNs)得到了广泛应用,但视觉变换器(ViTs)在捕捉长程依赖性和错综复杂的空间关系方面提供了一种有前途的替代方法,从而增强了对干旱胁迫微妙指标的检测。我们提出了一种可解释的深度学习流水线,利用视觉转换器的强大功能,利用航空图像检测马铃薯作物的干旱胁迫。我们采用了两种不同的方法:一种是 ViT 和支持向量机(SVM)的协同组合,其中 ViT 从航空图像中提取错综复杂的空间特征,SVM 将作物分类为受胁迫或健康;另一种是端到端方法,使用 ViT 中的专用分类层直接检测干旱胁迫。我们的主要发现通过可视化注意力地图解释了 ViT 模型的决策过程。这些地图突出显示了 ViT 模型作为干旱胁迫特征所关注的航空图像中的特定空间特征。我们的研究结果表明,所提出的方法不仅能实现高精度的干旱胁迫识别,还能揭示与干旱胁迫相关的各种细微植物特征。这为干旱胁迫监测提供了一种稳健且可解释的解决方案,农民可据此做出明智的决策,改善作物管理。
{"title":"An Explainable Vision Transformer with Transfer Learning Combined with Support Vector Machine Based Efficient Drought Stress Identification","authors":"Aswini Kumar Patra, Ankit Varshney, Lingaraj Sahoo","doi":"arxiv-2407.21666","DOIUrl":"https://doi.org/arxiv-2407.21666","url":null,"abstract":"Early detection of drought stress is critical for taking timely measures for\u0000reducing crop loss before the drought impact becomes irreversible. The subtle\u0000phenotypical and physiological changes in response to drought stress are\u0000captured by non-invasive imaging techniques and these imaging data serve as\u0000valuable resource for machine learning methods to identify drought stress.\u0000While convolutional neural networks (CNNs) are in wide use, vision transformers\u0000(ViTs) present a promising alternative in capturing long-range dependencies and\u0000intricate spatial relationships, thereby enhancing the detection of subtle\u0000indicators of drought stress. We propose an explainable deep learning pipeline\u0000that leverages the power of ViTs for drought stress detection in potato crops\u0000using aerial imagery. We applied two distinct approaches: a synergistic\u0000combination of ViT and support vector machine (SVM), where ViT extracts\u0000intricate spatial features from aerial images, and SVM classifies the crops as\u0000stressed or healthy and an end-to-end approach using a dedicated classification\u0000layer within ViT to directly detect drought stress. Our key findings explain\u0000the ViT model's decision-making process by visualizing attention maps. These\u0000maps highlight the specific spatial features within the aerial images that the\u0000ViT model focuses as the drought stress signature. Our findings demonstrate\u0000that the proposed methods not only achieve high accuracy in drought stress\u0000identification but also shedding light on the diverse subtle plant features\u0000associated with drought stress. This offers a robust and interpretable solution\u0000for drought stress monitoring for farmers to undertake informed decisions for\u0000improved crop management.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"46 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141866682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Sovereign Identity for Consented and Content-Based Access to Medical Records using Blockchain 使用区块链以自主身份同意和基于内容访问医疗记录
Pub Date : 2024-07-31 DOI: arxiv-2407.21559
Marie Tcholakian, Karolina Gorna, Maryline Laurent, Hella Kaffel Ben Ayed, Montassar Naghmouchi
Electronic Health Records (EHRs) and Medical Data are classified as personaldata in every privacy law, meaning that any related service that includesprocessing such data must come with full security, confidentiality, privacy andaccountability. Solutions for health data management, as in storing it, sharingand processing it, are emerging quickly and were significantly boosted by theCovid-19 pandemic that created a need to move things online. EHRs makes acrucial part of digital identity data, and the same digital identity trends --as in self sovereign identity powered by decentralized ledger technologies likeBlockchain, are being researched or implemented in contexts managing digitalinteractions between health facilities, patients and health professionals. Inthis paper, we propose a blockchain-based solution enabling secure exchange ofEHRs between different parties powered by a self-sovereign identity (SSI)wallet and decentralized identifiers. We also make use of a consortium IPFSnetwork for off-chain storage and attribute-based encryption (ABE) to ensuredata confidentiality and integrity. Through our solution, we grant users fullcontrol over their medical data, and enable them to securely share it in totalconfidentiality over secure communication channels between user wallets usingencryption. We also use DIDs for better user privacy and limit any possiblecorrelations or identification by using pairwise DIDs. Overall, combining thisset of technologies guarantees secure exchange of EHRs, secure storage andmanagement along with by-design features inherited from the technologicalstack.
电子健康记录 (EHR) 和医疗数据在所有隐私法中都被归类为个人数据,这意味着任何包括处理此类数据的相关服务都必须具备全面的安全性、保密性、隐私性和责任性。医疗数据管理解决方案,如存储、共享和处理数据,正在迅速兴起,并因第 19 号科维德病毒大流行而得到极大推动,该病毒产生了将数据转移到网上的需求。电子病历是数字身份数据的重要组成部分,而同样的数字身份趋势--即由区块链等分散式分类账技术驱动的自我主权身份--正在医疗机构、患者和医疗专业人员之间的数字互动管理中得到研究或实施。在本文中,我们提出了一种基于区块链的解决方案,通过自我主权身份(SSI)钱包和去中心化标识符,实现各方之间安全交换电子健康记录。我们还利用联盟 IPFS 网络进行链外存储,并使用基于属性的加密(ABE)来确保数据的机密性和完整性。通过我们的解决方案,用户可以完全控制自己的医疗数据,并通过用户钱包之间的安全通信渠道,使用加密技术安全地共享完全保密的数据。我们还使用 DID 来改善用户隐私,并通过使用成对 DID 来限制任何可能的关联或识别。总之,将这一系列技术结合起来,可以保证电子病历的安全交换、安全存储和管理,以及从技术栈中继承的设计功能。
{"title":"Self-Sovereign Identity for Consented and Content-Based Access to Medical Records using Blockchain","authors":"Marie Tcholakian, Karolina Gorna, Maryline Laurent, Hella Kaffel Ben Ayed, Montassar Naghmouchi","doi":"arxiv-2407.21559","DOIUrl":"https://doi.org/arxiv-2407.21559","url":null,"abstract":"Electronic Health Records (EHRs) and Medical Data are classified as personal\u0000data in every privacy law, meaning that any related service that includes\u0000processing such data must come with full security, confidentiality, privacy and\u0000accountability. Solutions for health data management, as in storing it, sharing\u0000and processing it, are emerging quickly and were significantly boosted by the\u0000Covid-19 pandemic that created a need to move things online. EHRs makes a\u0000crucial part of digital identity data, and the same digital identity trends --\u0000as in self sovereign identity powered by decentralized ledger technologies like\u0000Blockchain, are being researched or implemented in contexts managing digital\u0000interactions between health facilities, patients and health professionals. In\u0000this paper, we propose a blockchain-based solution enabling secure exchange of\u0000EHRs between different parties powered by a self-sovereign identity (SSI)\u0000wallet and decentralized identifiers. We also make use of a consortium IPFS\u0000network for off-chain storage and attribute-based encryption (ABE) to ensure\u0000data confidentiality and integrity. Through our solution, we grant users full\u0000control over their medical data, and enable them to securely share it in total\u0000confidentiality over secure communication channels between user wallets using\u0000encryption. We also use DIDs for better user privacy and limit any possible\u0000correlations or identification by using pairwise DIDs. Overall, combining this\u0000set of technologies guarantees secure exchange of EHRs, secure storage and\u0000management along with by-design features inherited from the technological\u0000stack.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"49 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141866683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CultureVo: The Serious Game of Utilizing Gen AI for Enhancing Cultural Intelligence CultureVo:利用创人工智能提升文化智能的严肃游戏
Pub Date : 2024-07-30 DOI: arxiv-2407.20685
Ajita Agarwala, Anupam Purwar, Viswanadhasai Rao
CultureVo, Inc. has developed the Integrated Culture Learning Suite (ICLS) todeliver foundational knowledge of world cultures through a combination ofinteractive lessons and gamified experiences. This paper explores howGenerative AI powered by open source Large Langauge Models are utilized withinthe ICLS to enhance cultural intelligence. The suite employs Generative AItechniques to automate the assessment of learner knowledge, analyze behavioralpatterns, and manage interactions with non-player characters using real timelearner assessment. Additionally, ICLS provides contextual hint and recommendcourse content by assessing learner proficiency, while Generative AIfacilitates the automated creation and validation of educational content.
CultureVo 公司开发了 "综合文化学习套件"(ICLS),通过互动课程和游戏化体验相结合的方式传授有关世界文化的基础知识。本文探讨了如何在 ICLS 中利用由开源大型语言模型(Large Langauge Models)驱动的生成式人工智能(Generative AI)来提高文化智能。该套件采用了生成式人工智能技术来自动评估学习者的知识、分析行为模式,并利用实时学习者评估来管理与非玩家角色的互动。此外,ICLS 还可通过评估学习者的熟练程度来提供情境提示和推荐课程内容,而生成式人工智能则有助于自动创建和验证教育内容。
{"title":"CultureVo: The Serious Game of Utilizing Gen AI for Enhancing Cultural Intelligence","authors":"Ajita Agarwala, Anupam Purwar, Viswanadhasai Rao","doi":"arxiv-2407.20685","DOIUrl":"https://doi.org/arxiv-2407.20685","url":null,"abstract":"CultureVo, Inc. has developed the Integrated Culture Learning Suite (ICLS) to\u0000deliver foundational knowledge of world cultures through a combination of\u0000interactive lessons and gamified experiences. This paper explores how\u0000Generative AI powered by open source Large Langauge Models are utilized within\u0000the ICLS to enhance cultural intelligence. The suite employs Generative AI\u0000techniques to automate the assessment of learner knowledge, analyze behavioral\u0000patterns, and manage interactions with non-player characters using real time\u0000learner assessment. Additionally, ICLS provides contextual hint and recommend\u0000course content by assessing learner proficiency, while Generative AI\u0000facilitates the automated creation and validation of educational content.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"42 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141866684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Introducing a new hyper-parameter for RAG: Context Window Utilization 为 RAG 引入一个新的超参数:上下文窗口利用率
Pub Date : 2024-07-29 DOI: arxiv-2407.19794
Kush Juvekar, Anupam Purwar
This paper introduces a new hyper-parameter for Retrieval-AugmentedGeneration (RAG) systems called Context Window Utilization. RAG systems enhancegenerative models by incorporating relevant information retrieved from externalknowledge bases, improving the factual accuracy and contextual relevance ofgenerated responses. The size of the text chunks retrieved and processed is acritical factor influencing RAG performance. This study aims to identify theoptimal chunk size that maximizes answer generation quality. Through systematicexperimentation, we analyze the effects of varying chunk sizes on theefficiency and effectiveness of RAG frameworks. Our findings reveal that anoptimal chunk size balances the trade-off between providing sufficient contextand minimizing irrelevant information. These insights are crucial for enhancingthe design and implementation of RAG systems, underscoring the importance ofselecting an appropriate chunk size to achieve superior performance.
本文为检索增强生成(RAG)系统引入了一个新的超参数,称为 "上下文窗口利用率"。RAG 系统通过纳入从外部知识库检索到的相关信息来增强生成模型,从而提高生成的回复的事实准确性和上下文相关性。检索和处理文本块的大小是影响 RAG 性能的关键因素。本研究旨在确定能最大限度提高答案生成质量的最佳块大小。通过系统实验,我们分析了不同块大小对 RAG 框架效率和效果的影响。我们的研究结果表明,最佳的块大小可以在提供足够的上下文和尽量减少无关信息之间取得平衡。这些见解对于改进 RAG 系统的设计和实施至关重要,强调了选择适当的块大小以实现卓越性能的重要性。
{"title":"Introducing a new hyper-parameter for RAG: Context Window Utilization","authors":"Kush Juvekar, Anupam Purwar","doi":"arxiv-2407.19794","DOIUrl":"https://doi.org/arxiv-2407.19794","url":null,"abstract":"This paper introduces a new hyper-parameter for Retrieval-Augmented\u0000Generation (RAG) systems called Context Window Utilization. RAG systems enhance\u0000generative models by incorporating relevant information retrieved from external\u0000knowledge bases, improving the factual accuracy and contextual relevance of\u0000generated responses. The size of the text chunks retrieved and processed is a\u0000critical factor influencing RAG performance. This study aims to identify the\u0000optimal chunk size that maximizes answer generation quality. Through systematic\u0000experimentation, we analyze the effects of varying chunk sizes on the\u0000efficiency and effectiveness of RAG frameworks. Our findings reveal that an\u0000optimal chunk size balances the trade-off between providing sufficient context\u0000and minimizing irrelevant information. These insights are crucial for enhancing\u0000the design and implementation of RAG systems, underscoring the importance of\u0000selecting an appropriate chunk size to achieve superior performance.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"205 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141866687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Emerging Technologies
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1