首页 > 最新文献

arXiv - CS - Machine Learning最新文献

英文 中文
The Impact of Element Ordering on LM Agent Performance 元素排序对 LM Agent 性能的影响
Pub Date : 2024-09-18 DOI: arxiv-2409.12089
Wayne Chi, Ameet Talwalkar, Chris Donahue
There has been a surge of interest in language model agents that can navigatevirtual environments such as the web or desktop. To navigate such environments,agents benefit from information on the various elements (e.g., buttons, text,or images) present. It remains unclear which element attributes have thegreatest impact on agent performance, especially in environments that onlyprovide a graphical representation (i.e., pixels). Here we find that theordering in which elements are presented to the language model is surprisinglyimpactful--randomizing element ordering in a webpage degrades agent performancecomparably to removing all visible text from an agent's state representation.While a webpage provides a hierarchical ordering of elements, there is no suchordering when parsing elements directly from pixels. Moreover, as tasks becomemore challenging and models more sophisticated, our experiments suggest thatthe impact of ordering increases. Finding an effective ordering is non-trivial.We investigate the impact of various element ordering methods in web anddesktop environments. We find that dimensionality reduction provides a viableordering for pixel-only environments. We train a UI element detection model toderive elements from pixels and apply our findings to an agentbenchmark--OmniACT--where we only have access to pixels. Our method completesmore than two times as many tasks on average relative to the previousstate-of-the-art.
人们对能够浏览虚拟环境(如网络或桌面)的语言模型代理兴趣浓厚。要浏览这些环境,代理需要了解各种元素(如按钮、文本或图像)的信息。目前还不清楚哪些元素属性对代理性能的影响最大,尤其是在只提供图形表示(即像素)的环境中。在这里,我们发现元素呈现给语言模型的排序具有令人惊讶的影响--在网页中对元素进行随机排序会降低代理的性能,其程度相当于从代理的状态表示中移除所有可见文本。此外,随着任务越来越具有挑战性,模型越来越复杂,我们的实验表明,排序的影响也在增加。我们研究了网络和桌面环境中各种元素排序方法的影响。我们发现,降维为纯像素环境提供了一种可行的排序方法。我们训练了一个用户界面元素检测模型,以便从像素中提取元素,并将我们的研究结果应用于一个代理基准--OmniACT--其中我们只能访问像素。我们的方法平均完成的任务量是之前最先进方法的两倍多。
{"title":"The Impact of Element Ordering on LM Agent Performance","authors":"Wayne Chi, Ameet Talwalkar, Chris Donahue","doi":"arxiv-2409.12089","DOIUrl":"https://doi.org/arxiv-2409.12089","url":null,"abstract":"There has been a surge of interest in language model agents that can navigate\u0000virtual environments such as the web or desktop. To navigate such environments,\u0000agents benefit from information on the various elements (e.g., buttons, text,\u0000or images) present. It remains unclear which element attributes have the\u0000greatest impact on agent performance, especially in environments that only\u0000provide a graphical representation (i.e., pixels). Here we find that the\u0000ordering in which elements are presented to the language model is surprisingly\u0000impactful--randomizing element ordering in a webpage degrades agent performance\u0000comparably to removing all visible text from an agent's state representation.\u0000While a webpage provides a hierarchical ordering of elements, there is no such\u0000ordering when parsing elements directly from pixels. Moreover, as tasks become\u0000more challenging and models more sophisticated, our experiments suggest that\u0000the impact of ordering increases. Finding an effective ordering is non-trivial.\u0000We investigate the impact of various element ordering methods in web and\u0000desktop environments. We find that dimensionality reduction provides a viable\u0000ordering for pixel-only environments. We train a UI element detection model to\u0000derive elements from pixels and apply our findings to an agent\u0000benchmark--OmniACT--where we only have access to pixels. Our method completes\u0000more than two times as many tasks on average relative to the previous\u0000state-of-the-art.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Handling Long-Term Safety and Uncertainty in Safe Reinforcement Learning 在安全强化学习中处理长期安全性和不确定性
Pub Date : 2024-09-18 DOI: arxiv-2409.12045
Jonas Günster, Puze Liu, Jan Peters, Davide Tateo
Safety is one of the key issues preventing the deployment of reinforcementlearning techniques in real-world robots. While most approaches in the SafeReinforcement Learning area do not require prior knowledge of constraints androbot kinematics and rely solely on data, it is often difficult to deploy themin complex real-world settings. Instead, model-based approaches thatincorporate prior knowledge of the constraints and dynamics into the learningframework have proven capable of deploying the learning algorithm directly onthe real robot. Unfortunately, while an approximated model of the robotdynamics is often available, the safety constraints are task-specific and hardto obtain: they may be too complicated to encode analytically, too expensive tocompute, or it may be difficult to envision a priori the long-term safetyrequirements. In this paper, we bridge this gap by extending the safeexploration method, ATACOM, with learnable constraints, with a particular focuson ensuring long-term safety and handling of uncertainty. Our approach iscompetitive or superior to state-of-the-art methods in final performance whilemaintaining safer behavior during training.
安全是阻碍在真实世界机器人中应用强化学习技术的关键问题之一。虽然安全强化学习领域的大多数方法都不需要事先了解约束条件和机器人运动学知识,而完全依赖于数据,但在复杂的真实世界环境中部署这些方法往往很困难。相反,基于模型的方法将约束条件和动力学的先验知识纳入学习框架,已被证明能够直接在真实机器人上部署学习算法。遗憾的是,虽然机器人动力学的近似模型通常可用,但安全约束条件是特定任务且难以获得的:它们可能过于复杂,难以分析编码,计算成本过高,或者难以预先设想长期安全要求。在本文中,我们利用可学习的约束条件扩展了安全探索方法 ATACOM,从而弥补了这一差距,特别是在确保长期安全和处理不确定性方面。我们的方法在最终性能上可与最先进的方法媲美或更胜一筹,同时在训练过程中保持更安全的行为。
{"title":"Handling Long-Term Safety and Uncertainty in Safe Reinforcement Learning","authors":"Jonas Günster, Puze Liu, Jan Peters, Davide Tateo","doi":"arxiv-2409.12045","DOIUrl":"https://doi.org/arxiv-2409.12045","url":null,"abstract":"Safety is one of the key issues preventing the deployment of reinforcement\u0000learning techniques in real-world robots. While most approaches in the Safe\u0000Reinforcement Learning area do not require prior knowledge of constraints and\u0000robot kinematics and rely solely on data, it is often difficult to deploy them\u0000in complex real-world settings. Instead, model-based approaches that\u0000incorporate prior knowledge of the constraints and dynamics into the learning\u0000framework have proven capable of deploying the learning algorithm directly on\u0000the real robot. Unfortunately, while an approximated model of the robot\u0000dynamics is often available, the safety constraints are task-specific and hard\u0000to obtain: they may be too complicated to encode analytically, too expensive to\u0000compute, or it may be difficult to envision a priori the long-term safety\u0000requirements. In this paper, we bridge this gap by extending the safe\u0000exploration method, ATACOM, with learnable constraints, with a particular focus\u0000on ensuring long-term safety and handling of uncertainty. Our approach is\u0000competitive or superior to state-of-the-art methods in final performance while\u0000maintaining safer behavior during training.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing PM2.5 Data Imputation and Prediction in Air Quality Monitoring Networks Using a KNN-SINDy Hybrid Model 利用 KNN-SINDy 混合模型加强空气质量监测网络中的 PM2.5 数据推算和预测
Pub Date : 2024-09-18 DOI: arxiv-2409.11640
Yohan Choi, Boaz Choi, Jachin Choi
Air pollution, particularly particulate matter (PM2.5), poses significantrisks to public health and the environment, necessitating accurate predictionand continuous monitoring for effective air quality management. However, airquality monitoring (AQM) data often suffer from missing records due to varioustechnical difficulties. This study explores the application of SparseIdentification of Nonlinear Dynamics (SINDy) for imputing missing PM2.5 data bypredicting, using training data from 2016, and comparing its performance withthe established Soft Impute (SI) and K-Nearest Neighbors (KNN) methods.
空气污染,尤其是颗粒物(PM2.5),对公众健康和环境构成了重大风险,需要准确预测和持续监测,以进行有效的空气质量管理。然而,由于各种技术上的困难,空气质量监测(AQM)数据往往存在记录缺失的问题。本研究利用 2016 年的训练数据,探索应用非线性动力学稀疏识别(SparseIdentification of Nonlinear Dynamics,SINDy)对缺失的 PM2.5 数据进行预测归因,并将其性能与已有的软归因(Soft Impute,SI)和 K-近邻(K-Nearest Neighbors,KNN)方法进行比较。
{"title":"Enhancing PM2.5 Data Imputation and Prediction in Air Quality Monitoring Networks Using a KNN-SINDy Hybrid Model","authors":"Yohan Choi, Boaz Choi, Jachin Choi","doi":"arxiv-2409.11640","DOIUrl":"https://doi.org/arxiv-2409.11640","url":null,"abstract":"Air pollution, particularly particulate matter (PM2.5), poses significant\u0000risks to public health and the environment, necessitating accurate prediction\u0000and continuous monitoring for effective air quality management. However, air\u0000quality monitoring (AQM) data often suffer from missing records due to various\u0000technical difficulties. This study explores the application of Sparse\u0000Identification of Nonlinear Dynamics (SINDy) for imputing missing PM2.5 data by\u0000predicting, using training data from 2016, and comparing its performance with\u0000the established Soft Impute (SI) and K-Nearest Neighbors (KNN) methods.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Unified Framework for Neural Computation and Learning Over Time 神经计算和随时间学习的统一框架
Pub Date : 2024-09-18 DOI: arxiv-2409.12038
Stefano Melacci, Alessandro Betti, Michele Casoni, Tommaso Guidi, Matteo Tiezzi, Marco Gori
This paper proposes Hamiltonian Learning, a novel unified framework forlearning with neural networks "over time", i.e., from a possibly infinitestream of data, in an online manner, without having access to futureinformation. Existing works focus on the simplified setting in which the streamhas a known finite length or is segmented into smaller sequences, leveragingwell-established learning strategies from statistical machine learning. In thispaper, the problem of learning over time is rethought from scratch, leveragingtools from optimal control theory, which yield a unifying view of the temporaldynamics of neural computations and learning. Hamiltonian Learning is based ondifferential equations that: (i) can be integrated without the need of externalsoftware solvers; (ii) generalize the well-established notion of gradient-basedlearning in feed-forward and recurrent networks; (iii) open to novelperspectives. The proposed framework is showcased by experimentally proving howit can recover gradient-based learning, comparing it to out-of-the boxoptimizers, and describing how it is flexible enough to switch from fully-localto partially/non-local computational schemes, possibly distributed overmultiple devices, and BackPropagation without storing activations. HamiltonianLearning is easy to implement and can help researches approach in a principledand innovative manner the problem of learning over time.
本文提出了 "哈密顿学习"(Hamiltonian Learning)这一新颖的统一框架,用于神经网络的 "随时间 "学习,即以在线方式从可能无穷大的数据流中学习,而无需获取未来信息。现有的工作主要集中在数据流具有已知有限长度或被分割成较小序列的简化设置上,利用的是统计机器学习中成熟的学习策略。本文利用最优控制理论中的工具,从头开始重新思考随时间学习的问题,从而统一了神经计算和学习的时间动力学观点。汉密尔顿学习法基于以下微分方程(i) 无需外部软件求解器即可集成;(ii) 在前馈和递归网络中概括基于梯度学习的成熟概念;(iii) 向新观点开放。通过实验证明如何恢复基于梯度的学习,将其与开箱即用的优化器进行比较,以及描述如何灵活地从完全本地计算方案切换到部分/非本地计算方案(可能分布在多个设备上)和不存储激活的反向传播,展示了所提出的框架。汉密尔顿学习法易于实现,可以帮助研究人员以有原则和创新的方式解决随时间学习的问题。
{"title":"A Unified Framework for Neural Computation and Learning Over Time","authors":"Stefano Melacci, Alessandro Betti, Michele Casoni, Tommaso Guidi, Matteo Tiezzi, Marco Gori","doi":"arxiv-2409.12038","DOIUrl":"https://doi.org/arxiv-2409.12038","url":null,"abstract":"This paper proposes Hamiltonian Learning, a novel unified framework for\u0000learning with neural networks \"over time\", i.e., from a possibly infinite\u0000stream of data, in an online manner, without having access to future\u0000information. Existing works focus on the simplified setting in which the stream\u0000has a known finite length or is segmented into smaller sequences, leveraging\u0000well-established learning strategies from statistical machine learning. In this\u0000paper, the problem of learning over time is rethought from scratch, leveraging\u0000tools from optimal control theory, which yield a unifying view of the temporal\u0000dynamics of neural computations and learning. Hamiltonian Learning is based on\u0000differential equations that: (i) can be integrated without the need of external\u0000software solvers; (ii) generalize the well-established notion of gradient-based\u0000learning in feed-forward and recurrent networks; (iii) open to novel\u0000perspectives. The proposed framework is showcased by experimentally proving how\u0000it can recover gradient-based learning, comparing it to out-of-the box\u0000optimizers, and describing how it is flexible enough to switch from fully-local\u0000to partially/non-local computational schemes, possibly distributed over\u0000multiple devices, and BackPropagation without storing activations. Hamiltonian\u0000Learning is easy to implement and can help researches approach in a principled\u0000and innovative manner the problem of learning over time.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stronger Baseline Models -- A Key Requirement for Aligning Machine Learning Research with Clinical Utility 更强大的基线模型 -- 将机器学习研究与临床实用性相结合的关键要求
Pub Date : 2024-09-18 DOI: arxiv-2409.12116
Nathan Wolfrath, Joel Wolfrath, Hengrui Hu, Anjishnu Banerjee, Anai N. Kothari
Machine Learning (ML) research has increased substantially in recent years,due to the success of predictive modeling across diverse application domains.However, well-known barriers exist when attempting to deploy ML models inhigh-stakes, clinical settings, including lack of model transparency (or theinability to audit the inference process), large training data requirementswith siloed data sources, and complicated metrics for measuring model utility.In this work, we show empirically that including stronger baseline models inhealthcare ML evaluations has important downstream effects that aidpractitioners in addressing these challenges. Through a series of case studies,we find that the common practice of omitting baselines or comparing against aweak baseline model (e.g. a linear model with no optimization) obscures thevalue of ML methods proposed in the research literature. Using these insights,we propose some best practices that will enable practitioners to moreeffectively study and deploy ML models in clinical settings.
近年来,由于预测建模在不同应用领域取得了成功,机器学习(ML)研究大幅增加。然而,在高风险的临床环境中尝试部署 ML 模型时存在众所周知的障碍,包括缺乏模型透明度(或无法审计推理过程)、孤立数据源的大量训练数据要求以及衡量模型效用的复杂指标。在这项工作中,我们通过实证研究表明,在医疗保健 ML 评估中纳入更强的基线模型具有重要的下游效应,有助于实践者应对这些挑战。通过一系列案例研究,我们发现省略基线模型或与弱基线模型(如未优化的线性模型)进行比较的常见做法掩盖了研究文献中提出的 ML 方法的价值。利用这些洞察力,我们提出了一些最佳实践,使从业人员能够在临床环境中更有效地研究和部署 ML 模型。
{"title":"Stronger Baseline Models -- A Key Requirement for Aligning Machine Learning Research with Clinical Utility","authors":"Nathan Wolfrath, Joel Wolfrath, Hengrui Hu, Anjishnu Banerjee, Anai N. Kothari","doi":"arxiv-2409.12116","DOIUrl":"https://doi.org/arxiv-2409.12116","url":null,"abstract":"Machine Learning (ML) research has increased substantially in recent years,\u0000due to the success of predictive modeling across diverse application domains.\u0000However, well-known barriers exist when attempting to deploy ML models in\u0000high-stakes, clinical settings, including lack of model transparency (or the\u0000inability to audit the inference process), large training data requirements\u0000with siloed data sources, and complicated metrics for measuring model utility.\u0000In this work, we show empirically that including stronger baseline models in\u0000healthcare ML evaluations has important downstream effects that aid\u0000practitioners in addressing these challenges. Through a series of case studies,\u0000we find that the common practice of omitting baselines or comparing against a\u0000weak baseline model (e.g. a linear model with no optimization) obscures the\u0000value of ML methods proposed in the research literature. Using these insights,\u0000we propose some best practices that will enable practitioners to more\u0000effectively study and deploy ML models in clinical settings.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Consistent Estimation of a Class of Distances Between Covariance Matrices 协方差矩阵间一类距离的一致性估计
Pub Date : 2024-09-18 DOI: arxiv-2409.11761
Roberto Pereira, Xavier Mestre, Davig Gregoratti
This work considers the problem of estimating the distance between twocovariance matrices directly from the data. Particularly, we are interested inthe family of distances that can be expressed as sums of traces of functionsthat are separately applied to each covariance matrix. This family of distancesis particularly useful as it takes into consideration the fact that covariancematrices lie in the Riemannian manifold of positive definite matrices, therebyincluding a variety of commonly used metrics, such as the Euclidean distance,Jeffreys' divergence, and the log-Euclidean distance. Moreover, a statisticalanalysis of the asymptotic behavior of this class of distance estimators hasalso been conducted. Specifically, we present a central limit theorem thatestablishes the asymptotic Gaussianity of these estimators and provides closedform expressions for the corresponding means and variances. Empiricalevaluations demonstrate the superiority of our proposed consistent estimatorover conventional plug-in estimators in multivariate analytical contexts.Additionally, the central limit theorem derived in this study provides a robuststatistical framework to assess of accuracy of these estimators.
本研究考虑的问题是直接从数据中估计两个协方差矩阵之间的距离。特别是,我们对可以表示为分别应用于每个协方差矩阵的函数迹之和的距离族感兴趣。这个距离族特别有用,因为它考虑到了协方差矩阵位于正定矩阵的黎曼流形中这一事实,从而包括了各种常用度量,如欧氏距离、杰弗里斯发散和对数欧氏距离。此外,我们还对这类距离估计器的渐近行为进行了统计分析。具体来说,我们提出了一个中心极限定理,该定理证明了这些估计值的渐近高斯性,并提供了相应均值和方差的闭式表达式。经验评估表明,我们提出的一致估计器在多元分析背景下优于传统的插入式估计器。此外,本研究中得出的中心极限定理为评估这些估计器的准确性提供了一个稳健的统计框架。
{"title":"Consistent Estimation of a Class of Distances Between Covariance Matrices","authors":"Roberto Pereira, Xavier Mestre, Davig Gregoratti","doi":"arxiv-2409.11761","DOIUrl":"https://doi.org/arxiv-2409.11761","url":null,"abstract":"This work considers the problem of estimating the distance between two\u0000covariance matrices directly from the data. Particularly, we are interested in\u0000the family of distances that can be expressed as sums of traces of functions\u0000that are separately applied to each covariance matrix. This family of distances\u0000is particularly useful as it takes into consideration the fact that covariance\u0000matrices lie in the Riemannian manifold of positive definite matrices, thereby\u0000including a variety of commonly used metrics, such as the Euclidean distance,\u0000Jeffreys' divergence, and the log-Euclidean distance. Moreover, a statistical\u0000analysis of the asymptotic behavior of this class of distance estimators has\u0000also been conducted. Specifically, we present a central limit theorem that\u0000establishes the asymptotic Gaussianity of these estimators and provides closed\u0000form expressions for the corresponding means and variances. Empirical\u0000evaluations demonstrate the superiority of our proposed consistent estimator\u0000over conventional plug-in estimators in multivariate analytical contexts.\u0000Additionally, the central limit theorem derived in this study provides a robust\u0000statistical framework to assess of accuracy of these estimators.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NPAT Null-Space Projected Adversarial Training Towards Zero Deterioration NPAT 零空间预测对抗训练,实现零恶化
Pub Date : 2024-09-18 DOI: arxiv-2409.11754
Hanyi Hu, Qiao Han, Kui Chen, Yao Yang
To mitigate the susceptibility of neural networks to adversarial attacks,adversarial training has emerged as a prevalent and effective defense strategy.Intrinsically, this countermeasure incurs a trade-off, as it sacrifices themodel's accuracy in processing normal samples. To reconcile the trade-off, wepioneer the incorporation of null-space projection into adversarial trainingand propose two innovative Null-space Projection based AdversarialTraining(NPAT) algorithms tackling sample generation and gradient optimization,named Null-space Projected Data Augmentation (NPDA) and Null-space ProjectedGradient Descent (NPGD), to search for an overarching optimal solutions, whichenhance robustness with almost zero deterioration in generalizationperformance. Adversarial samples and perturbations are constrained within thenull-space of the decision boundary utilizing a closed-form null-spaceprojector, effectively mitigating threat of attack stemming from unreliablefeatures. Subsequently, we conducted experiments on the CIFAR10 and SVHNdatasets and reveal that our methodology can seamlessly combine withadversarial training methods and obtain comparable robustness while keepinggeneralization close to a high-accuracy model.
为了降低神经网络对对抗性攻击的敏感性,对抗性训练已成为一种普遍而有效的防御策略。从本质上讲,这种对策需要权衡利弊,因为它牺牲了模型处理正常样本的准确性。为了调和这种权衡,我们率先在对抗训练中加入了空空间投影,并提出了两种创新的基于空空间投影的对抗训练(NPAT)算法,即空空间投影数据增强算法(NPDA)和空空间投影梯度下降算法(NPGD),这两种算法解决了样本生成和梯度优化的问题,以寻找总体最优解,从而在几乎不降低泛化性能的情况下提高鲁棒性。利用闭式空空间投影器将对抗样本和扰动限制在决策边界的空空间内,从而有效降低了来自不可靠特征的攻击威胁。随后,我们在 CIFAR10 和 SVHN 数据集上进行了实验,结果表明我们的方法可以与对抗训练方法无缝结合,并获得相当的鲁棒性,同时使泛化接近高精度模型。
{"title":"NPAT Null-Space Projected Adversarial Training Towards Zero Deterioration","authors":"Hanyi Hu, Qiao Han, Kui Chen, Yao Yang","doi":"arxiv-2409.11754","DOIUrl":"https://doi.org/arxiv-2409.11754","url":null,"abstract":"To mitigate the susceptibility of neural networks to adversarial attacks,\u0000adversarial training has emerged as a prevalent and effective defense strategy.\u0000Intrinsically, this countermeasure incurs a trade-off, as it sacrifices the\u0000model's accuracy in processing normal samples. To reconcile the trade-off, we\u0000pioneer the incorporation of null-space projection into adversarial training\u0000and propose two innovative Null-space Projection based Adversarial\u0000Training(NPAT) algorithms tackling sample generation and gradient optimization,\u0000named Null-space Projected Data Augmentation (NPDA) and Null-space Projected\u0000Gradient Descent (NPGD), to search for an overarching optimal solutions, which\u0000enhance robustness with almost zero deterioration in generalization\u0000performance. Adversarial samples and perturbations are constrained within the\u0000null-space of the decision boundary utilizing a closed-form null-space\u0000projector, effectively mitigating threat of attack stemming from unreliable\u0000features. Subsequently, we conducted experiments on the CIFAR10 and SVHN\u0000datasets and reveal that our methodology can seamlessly combine with\u0000adversarial training methods and obtain comparable robustness while keeping\u0000generalization close to a high-accuracy model.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-Shot Class-Incremental Learning with Non-IID Decentralized Data 利用非 IID 分散数据进行少镜头分类增量学习
Pub Date : 2024-09-18 DOI: arxiv-2409.11657
Cuiwei Liu, Siang Xu, Huaijun Qiu, Jing Zhang, Zhi Liu, Liang Zhao
Few-shot class-incremental learning is crucial for developing scalable andadaptive intelligent systems, as it enables models to acquire new classes withminimal annotated data while safeguarding the previously accumulated knowledge.Nonetheless, existing methods deal with continuous data streams in acentralized manner, limiting their applicability in scenarios that prioritizedata privacy and security. To this end, this paper introduces federatedfew-shot class-incremental learning, a decentralized machine learning paradigmtailored to progressively learn new classes from scarce data distributed acrossmultiple clients. In this learning paradigm, clients locally update theirmodels with new classes while preserving data privacy, and then transmit themodel updates to a central server where they are aggregated globally. However,this paradigm faces several issues, such as difficulties in few-shot learning,catastrophic forgetting, and data heterogeneity. To address these challenges,we present a synthetic data-driven framework that leverages replay buffer datato maintain existing knowledge and facilitate the acquisition of new knowledge.Within this framework, a noise-aware generative replay module is developed tofine-tune local models with a balance of new and replay data, while generatingsynthetic data of new classes to further expand the replay buffer for futuretasks. Furthermore, a class-specific weighted aggregation strategy is designedto tackle data heterogeneity by adaptively aggregating class-specificparameters based on local models performance on synthetic data. This enableseffective global model optimization without direct access to client data.Comprehensive experiments across three widely-used datasets underscore theeffectiveness and preeminence of the introduced framework.
少量类递增学习对于开发可扩展和自适应的智能系统至关重要,因为它使模型能够利用最少的注释数据获取新类,同时保护先前积累的知识。然而,现有方法以集中方式处理连续数据流,限制了它们在优先考虑数据隐私和安全的场景中的适用性。为此,本文介绍了一种去中心化的机器学习范式--联合少量类递增学习(federatedfew-shot class-incremental learning),旨在从分布在多个客户端的稀缺数据中逐步学习新的类别。在这种学习范式中,客户端在保护数据隐私的前提下用新类别更新本地模型,然后将模型更新传输到中央服务器,由服务器进行全球汇总。然而,这种范例面临着一些问题,如少量学习困难、灾难性遗忘和数据异构等。为了应对这些挑战,我们提出了一个合成数据驱动框架,利用重放缓冲区数据来维护现有知识并促进新知识的获取。在这个框架内,我们开发了一个噪声感知生成重放模块,利用新数据和重放数据的平衡来精细调整本地模型,同时生成新类别的合成数据,为未来任务进一步扩展重放缓冲区。此外,还设计了一种特定类别的加权聚合策略,根据本地模型在合成数据上的表现,自适应地聚合特定类别的参数,从而解决数据异质性问题。在三个广泛使用的数据集上进行的综合实验证明了所引入框架的有效性和优越性。
{"title":"Few-Shot Class-Incremental Learning with Non-IID Decentralized Data","authors":"Cuiwei Liu, Siang Xu, Huaijun Qiu, Jing Zhang, Zhi Liu, Liang Zhao","doi":"arxiv-2409.11657","DOIUrl":"https://doi.org/arxiv-2409.11657","url":null,"abstract":"Few-shot class-incremental learning is crucial for developing scalable and\u0000adaptive intelligent systems, as it enables models to acquire new classes with\u0000minimal annotated data while safeguarding the previously accumulated knowledge.\u0000Nonetheless, existing methods deal with continuous data streams in a\u0000centralized manner, limiting their applicability in scenarios that prioritize\u0000data privacy and security. To this end, this paper introduces federated\u0000few-shot class-incremental learning, a decentralized machine learning paradigm\u0000tailored to progressively learn new classes from scarce data distributed across\u0000multiple clients. In this learning paradigm, clients locally update their\u0000models with new classes while preserving data privacy, and then transmit the\u0000model updates to a central server where they are aggregated globally. However,\u0000this paradigm faces several issues, such as difficulties in few-shot learning,\u0000catastrophic forgetting, and data heterogeneity. To address these challenges,\u0000we present a synthetic data-driven framework that leverages replay buffer data\u0000to maintain existing knowledge and facilitate the acquisition of new knowledge.\u0000Within this framework, a noise-aware generative replay module is developed to\u0000fine-tune local models with a balance of new and replay data, while generating\u0000synthetic data of new classes to further expand the replay buffer for future\u0000tasks. Furthermore, a class-specific weighted aggregation strategy is designed\u0000to tackle data heterogeneity by adaptively aggregating class-specific\u0000parameters based on local models performance on synthetic data. This enables\u0000effective global model optimization without direct access to client data.\u0000Comprehensive experiments across three widely-used datasets underscore the\u0000effectiveness and preeminence of the introduced framework.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unraveling the Hessian: A Key to Smooth Convergence in Loss Function Landscapes 揭开赫塞斯的面纱:损失函数景观平滑收敛的关键
Pub Date : 2024-09-18 DOI: arxiv-2409.11995
Nikita Kiselev, Andrey Grabovoy
The loss landscape of neural networks is a critical aspect of their training,and understanding its properties is essential for improving their performance.In this paper, we investigate how the loss surface changes when the sample sizeincreases, a previously unexplored issue. We theoretically analyze theconvergence of the loss landscape in a fully connected neural network andderive upper bounds for the difference in loss function values when adding anew object to the sample. Our empirical study confirms these results on variousdatasets, demonstrating the convergence of the loss function surface for imageclassification tasks. Our findings provide insights into the local geometry ofneural loss landscapes and have implications for the development of sample sizedetermination techniques.
神经网络的损失面是其训练的一个关键方面,了解其特性对于提高神经网络的性能至关重要。在本文中,我们研究了当样本量增加时损失面如何变化,这是一个以前从未探讨过的问题。我们从理论上分析了全连接神经网络中损失面的收敛性,并得出了在样本中添加新对象时损失函数值差异的上限。我们的实证研究在各种数据集上证实了这些结果,证明了图像分类任务中损失函数面的收敛性。我们的研究结果为神经损失景观的局部几何提供了见解,并对样本大小确定技术的发展产生了影响。
{"title":"Unraveling the Hessian: A Key to Smooth Convergence in Loss Function Landscapes","authors":"Nikita Kiselev, Andrey Grabovoy","doi":"arxiv-2409.11995","DOIUrl":"https://doi.org/arxiv-2409.11995","url":null,"abstract":"The loss landscape of neural networks is a critical aspect of their training,\u0000and understanding its properties is essential for improving their performance.\u0000In this paper, we investigate how the loss surface changes when the sample size\u0000increases, a previously unexplored issue. We theoretically analyze the\u0000convergence of the loss landscape in a fully connected neural network and\u0000derive upper bounds for the difference in loss function values when adding a\u0000new object to the sample. Our empirical study confirms these results on various\u0000datasets, demonstrating the convergence of the loss function surface for image\u0000classification tasks. Our findings provide insights into the local geometry of\u0000neural loss landscapes and have implications for the development of sample size\u0000determination techniques.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Symmetry-Enriched Learning: A Category-Theoretic Framework for Robust Machine Learning Models 对称丰富学习:稳健机器学习模型的类别理论框架
Pub Date : 2024-09-18 DOI: arxiv-2409.12100
Ronald Katende
This manuscript presents a novel framework that integrates higher-ordersymmetries and category theory into machine learning. We introduce newmathematical constructs, including hyper-symmetry categories and functorialrepresentations, to model complex transformations within learning algorithms.Our contributions include the design of symmetry-enriched learning models, thedevelopment of advanced optimization techniques leveraging categoricalsymmetries, and the theoretical analysis of their implications for modelrobustness, generalization, and convergence. Through rigorous proofs andpractical applications, we demonstrate that incorporating higher-dimensionalcategorical structures enhances both the theoretical foundations and practicalcapabilities of modern machine learning algorithms, opening new directions forresearch and innovation.
本手稿提出了一个新颖的框架,将高阶对称和范畴理论整合到机器学习中。我们的贡献包括设计对称性丰富的学习模型、开发利用分类对称性的高级优化技术,以及从理论上分析它们对模型稳健性、泛化和收敛性的影响。通过严格的证明和实际应用,我们证明了结合高维分类结构可以增强现代机器学习算法的理论基础和实际能力,为研究和创新开辟了新的方向。
{"title":"Symmetry-Enriched Learning: A Category-Theoretic Framework for Robust Machine Learning Models","authors":"Ronald Katende","doi":"arxiv-2409.12100","DOIUrl":"https://doi.org/arxiv-2409.12100","url":null,"abstract":"This manuscript presents a novel framework that integrates higher-order\u0000symmetries and category theory into machine learning. We introduce new\u0000mathematical constructs, including hyper-symmetry categories and functorial\u0000representations, to model complex transformations within learning algorithms.\u0000Our contributions include the design of symmetry-enriched learning models, the\u0000development of advanced optimization techniques leveraging categorical\u0000symmetries, and the theoretical analysis of their implications for model\u0000robustness, generalization, and convergence. Through rigorous proofs and\u0000practical applications, we demonstrate that incorporating higher-dimensional\u0000categorical structures enhances both the theoretical foundations and practical\u0000capabilities of modern machine learning algorithms, opening new directions for\u0000research and innovation.","PeriodicalId":501301,"journal":{"name":"arXiv - CS - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261826","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Machine Learning
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1