首页 > 最新文献

Annals of Mathematics and Artificial Intelligence最新文献

英文 中文
A study of universal morphological analysis using morpheme-based, holistic, and neural approaches under various data size conditions 在不同数据规模条件下使用基于语素、整体和神经方法进行通用形态分析的研究
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-05-11 DOI: 10.1007/s10472-024-09944-8
Rashel Fam, Yves Lepage

We perform a study on the universal morphological analysis task: given a word form, generate the lemma (lemmatisation) and its corresponding morphosyntactic descriptions (MSD analysis). Experiments are carried out on the SIGMORPHON 2018 Shared Task: Morphological Reinflection Task dataset which consists of more than 100 different languages with various morphological richness under three different data size conditions: low, medium and high. We consider three main approaches: morpheme-based (eager learning), holistic (lazy learning), and neural (eager learning). Performance is evaluated on the two subtasks of lemmatisation and MSD analysis. For the lemmatisation subtask, under all three data sizes, experimental results show that the holistic approach predicted more accurate lemmata, while the morpheme-based approach produced lemmata closer to the answers when it produces the wrong answers. For the MSD analysis subtask, under all three data sizes, the holistic approach achieves higher recall, while the morpheme-based approach is more precise. However, the trade-off between precision and recall of the two systems leads to a very similar overall F1 score. On the whole, neural approaches suffer under low resource conditions, but they achieve the best performance in comparison to the other approaches when the size of the training data increases.

我们对通用形态分析任务进行了研究:给定词形,生成词母(词母化)及其相应的形态句法描述(MSD 分析)。实验在 SIGMORPHON 2018 共享任务上进行:该数据集由 100 多种不同语言组成,在低、中、高三种不同的数据规模条件下具有不同的形态丰富度。我们考虑了三种主要方法:基于词素(急于学习)、整体(懒于学习)和神经(急于学习)。我们对词素化和 MSD 分析这两项子任务的性能进行了评估。对于词素化子任务,在所有三种数据规模下,实验结果表明整体方法预测的词素更准确,而基于词素的方法在产生错误答案时产生的词素更接近答案。在 MSD 分析子任务中,在所有三种数据规模下,整体方法的召回率更高,而基于语素的方法更精确。不过,这两种系统在精确度和召回率之间的权衡导致了非常相似的总体 F1 分数。总的来说,神经方法在资源较少的情况下会受到影响,但当训练数据规模增大时,神经方法的性能会比其他方法更好。
{"title":"A study of universal morphological analysis using morpheme-based, holistic, and neural approaches under various data size conditions","authors":"Rashel Fam, Yves Lepage","doi":"10.1007/s10472-024-09944-8","DOIUrl":"https://doi.org/10.1007/s10472-024-09944-8","url":null,"abstract":"<p>We perform a study on the universal morphological analysis task: given a word form, generate the lemma (lemmatisation) and its corresponding morphosyntactic descriptions (MSD analysis). Experiments are carried out on the SIGMORPHON 2018 Shared Task: Morphological Reinflection Task dataset which consists of more than 100 different languages with various morphological richness under three different data size conditions: low, medium and high. We consider three main approaches: morpheme-based (eager learning), holistic (lazy learning), and neural (eager learning). Performance is evaluated on the two subtasks of lemmatisation and MSD analysis. For the lemmatisation subtask, under all three data sizes, experimental results show that the holistic approach predicted more accurate lemmata, while the morpheme-based approach produced lemmata closer to the answers when it produces the wrong answers. For the MSD analysis subtask, under all three data sizes, the holistic approach achieves higher recall, while the morpheme-based approach is more precise. However, the trade-off between precision and recall of the two systems leads to a very similar overall F1 score. On the whole, neural approaches suffer under low resource conditions, but they achieve the best performance in comparison to the other approaches when the size of the training data increases.</p>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"42 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140930676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep data density estimation through Donsker-Varadhan representation 通过 Donsker-Varadhan 表示进行深度数据密度估算
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-26 DOI: 10.1007/s10472-024-09943-9
Seonho Park, Panos M. Pardalos

Estimating the data density is one of the challenging problem topics in the deep learning society. In this paper, we present a simple yet effective methodology for estimating the data density using the Donsker-Varadhan variational lower bound on the KL divergence and the modeling based on the deep neural network. We demonstrate that the optimal critic function associated with the Donsker-Varadhan representation on the KL divergence between the data and the uniform distribution can estimate the data density. Also, we present the deep neural network-based modeling and its stochastic learning procedure. The experimental results and possible applications of the proposed method demonstrate that it is competitive with the previous methods for data density estimation and has a lot of possibilities for various applications.

估算数据密度是深度学习领域具有挑战性的问题之一。在本文中,我们提出了一种简单而有效的方法,利用 KL 发散的 Donsker-Varadhan 变分下界和基于深度神经网络的建模来估计数据密度。我们证明,与数据和均匀分布之间 KL 发散的 Donsker-Varadhan 表示相关的最佳批判函数可以估计数据密度。此外,我们还介绍了基于深度神经网络的建模及其随机学习过程。实验结果和拟议方法的可能应用表明,该方法与之前的数据密度估计方法相比具有竞争力,并且在各种应用中具有很大的可能性。
{"title":"Deep data density estimation through Donsker-Varadhan representation","authors":"Seonho Park, Panos M. Pardalos","doi":"10.1007/s10472-024-09943-9","DOIUrl":"https://doi.org/10.1007/s10472-024-09943-9","url":null,"abstract":"<p>Estimating the data density is one of the challenging problem topics in the deep learning society. In this paper, we present a simple yet effective methodology for estimating the data density using the Donsker-Varadhan variational lower bound on the KL divergence and the modeling based on the deep neural network. We demonstrate that the optimal critic function associated with the Donsker-Varadhan representation on the KL divergence between the data and the uniform distribution can estimate the data density. Also, we present the deep neural network-based modeling and its stochastic learning procedure. The experimental results and possible applications of the proposed method demonstrate that it is competitive with the previous methods for data density estimation and has a lot of possibilities for various applications.</p>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"19 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140806658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Signifiers for conveying and exploiting affordances: from human-computer interaction to multi-agent systems 传递和利用负担能力的符号:从人机交互到多代理系统
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-17 DOI: 10.1007/s10472-024-09938-6
Jérémy Lemée, Danai Vachtsevanou, Simon Mayer, Andrei Ciortea

The ecological psychologist James J. Gibson defined the notion of affordances to refer to what action possibilities environments offer to animals. In this paper, we show how (artificial) agents can discover and exploit affordances in a Multi-Agent System (MAS) environment to achieve their goals. To indicate to agents what affordances are present in their environment and whether it is likely that these may help the agents to achieve their objectives, the environment may expose signifiers while taking into account the current situation of the environment and of the agent. On this basis, we define a Signifier Exposure Mechanism that is used by the environment to compute which signifiers should be exposed to agents in order to permit agents to only perceive information about affordances that are likely to be relevant to them, and thereby increase their interaction efficiency. If this is successful, agents can interact with partially observable environments more efficiently because the signifiers indicate the affordances they can exploit towards given purposes. Signifiers thereby facilitate the exploration and the exploitation of MAS environments. Implementations of signifiers and of the Signifier Exposure Mechanism are presented within the context of a Hypermedia Multi-Agent System, and the utility of this approach is presented through the development of a scenario.

生态心理学家詹姆斯-吉布森(James J. Gibson)定义了 "可负担性"(affordances)这一概念,指的是环境为动物提供的行动可能性。在本文中,我们将展示(人工)代理如何在多代理系统(MAS)环境中发现并利用可负担性来实现其目标。为了向代理指明其所处环境中存在哪些可负担性,以及这些可负担性是否有可能帮助代理实现其目标,环境可以在考虑到环境和代理当前情况的情况下暴露出标志物。在此基础上,我们定义了一种标识符暴露机制,由环境来计算哪些标识符应暴露给代理,以便让代理只感知可能与其相关的负担能力信息,从而提高其交互效率。如果这样做成功的话,代理就能更有效地与部分可观测环境进行交互,因为标识符指明了他们可以利用的能力,以达到特定目的。因此,标识符有助于探索和利用 MAS 环境。在超媒体多代理系统的背景下,介绍了标识符和标识符暴露机制的实施,并通过一个场景的开发介绍了这种方法的实用性。
{"title":"Signifiers for conveying and exploiting affordances: from human-computer interaction to multi-agent systems","authors":"Jérémy Lemée,&nbsp;Danai Vachtsevanou,&nbsp;Simon Mayer,&nbsp;Andrei Ciortea","doi":"10.1007/s10472-024-09938-6","DOIUrl":"10.1007/s10472-024-09938-6","url":null,"abstract":"<div><p>The ecological psychologist James J. Gibson defined the notion of affordances to refer to what action possibilities environments offer to animals. In this paper, we show how (artificial) agents can discover and exploit affordances in a Multi-Agent System (MAS) environment to achieve their goals. To indicate to agents what affordances are present in their environment and whether it is likely that these may help the agents to achieve their objectives, the environment may expose signifiers while taking into account the current situation of the environment and of the agent. On this basis, we define a Signifier Exposure Mechanism that is used by the environment to compute which signifiers should be exposed to agents in order to permit agents to only perceive information about affordances that are likely to be relevant to them, and thereby increase their interaction efficiency. If this is successful, agents can interact with partially observable environments more efficiently because the signifiers indicate the affordances they can exploit towards given purposes. Signifiers thereby facilitate the exploration and the exploitation of MAS environments. Implementations of signifiers and of the Signifier Exposure Mechanism are presented within the context of a Hypermedia Multi-Agent System, and the utility of this approach is presented through the development of a scenario.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 4","pages":"815 - 835"},"PeriodicalIF":1.2,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09938-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140616802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Near-term advances in quantum natural language processing 量子自然语言处理的近期进展
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-11 DOI: 10.1007/s10472-024-09940-y
Dominic Widdows, Aaranya Alexander, Daiwei Zhu, Chase Zimmerman, Arunava Majumder

This paper describes experiments showing that some tasks in natural language processing (NLP) can already be performed using quantum computers, though so far only with small datasets. We demonstrate various approaches to topic classification. The first uses an explicit word-based approach, in which word-topic weights are implemented as fractional rotations of individual qubits, and a phrase is classified based on the accumulation of these weights onto a scoring qubit, using entangling quantum gates. This is compared with more scalable quantum encodings of word embedding vectors, which are used to compute kernel values in a quantum support vector machine: this approach achieved an average of 62% accuracy on classification tasks involving over 10000 words, which is the largest such quantum computing experiment to date. We describe a quantum probability approach to bigram modeling that can be applied to understand sequences of words and formal concepts, investigate a generative approximation to these distributions using a quantum circuit Born machine, and introduce an approach to ambiguity resolution in verb-noun composition using single-qubit rotations for simple nouns and 2-qubit entangling gates for simple verbs. The smaller systems presented have been run successfully on physical quantum computers, and the larger ones have been simulated. We show that statistically meaningful results can be obtained, but the quality of individual results varies much more using real datasets than using artificial language examples from previous quantum NLP research. Related NLP research is compared, partly with respect to contemporary challenges including informal language, fluency, and truthfulness.

本文描述的实验表明,自然语言处理(NLP)中的某些任务已经可以使用量子计算机来完成,尽管迄今为止只能使用小型数据集。我们展示了各种主题分类方法。第一种方法使用基于单词的显式方法,其中单词-主题权重是作为单个量子比特的分数旋转来实现的,而短语的分类则是基于这些权重在一个评分量子比特上的累积,使用纠缠量子门。我们将这种方法与单词嵌入向量的更可扩展量子编码进行了比较,后者用于计算量子支持向量机中的内核值:这种方法在涉及 10000 多个单词的分类任务中平均达到了 62% 的准确率,这是迄今为止最大规模的此类量子计算实验。我们描述了一种可用于理解单词序列和形式概念的大词建模量子概率方法,研究了使用量子电路伯恩机对这些分布进行生成近似的方法,并介绍了一种使用单量子比特旋转简单名词和双量子比特纠缠门解决动名词构成中歧义的方法。所介绍的较小系统已在物理量子计算机上成功运行,较大系统也已模拟运行。我们的研究表明,可以获得有统计意义的结果,但使用真实数据集比使用以前量子 NLP 研究中的人工语言示例,单个结果的质量差异要大得多。我们对相关的 NLP 研究进行了比较,其中部分研究涉及当代的挑战,包括非正式语言、流畅性和真实性。
{"title":"Near-term advances in quantum natural language processing","authors":"Dominic Widdows,&nbsp;Aaranya Alexander,&nbsp;Daiwei Zhu,&nbsp;Chase Zimmerman,&nbsp;Arunava Majumder","doi":"10.1007/s10472-024-09940-y","DOIUrl":"10.1007/s10472-024-09940-y","url":null,"abstract":"<div><p>This paper describes experiments showing that some tasks in natural language processing (NLP) can already be performed using quantum computers, though so far only with small datasets. We demonstrate various approaches to topic classification. The first uses an explicit word-based approach, in which word-topic weights are implemented as fractional rotations of individual qubits, and a phrase is classified based on the accumulation of these weights onto a scoring qubit, using entangling quantum gates. This is compared with more scalable quantum encodings of word embedding vectors, which are used to compute kernel values in a quantum support vector machine: this approach achieved an average of 62% accuracy on classification tasks involving over 10000 words, which is the largest such quantum computing experiment to date. We describe a quantum probability approach to bigram modeling that can be applied to understand sequences of words and formal concepts, investigate a generative approximation to these distributions using a quantum circuit Born machine, and introduce an approach to ambiguity resolution in verb-noun composition using single-qubit rotations for simple nouns and 2-qubit entangling gates for simple verbs. The smaller systems presented have been run successfully on physical quantum computers, and the larger ones have been simulated. We show that statistically meaningful results can be obtained, but the quality of individual results varies much more using real datasets than using artificial language examples from previous quantum NLP research. Related NLP research is compared, partly with respect to contemporary challenges including informal language, fluency, and truthfulness.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1249 - 1272"},"PeriodicalIF":1.2,"publicationDate":"2024-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating optimized item selection with active learning for continuous exploration in recommender systems 将优化项目选择与主动学习相结合,在推荐系统中进行持续探索
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-04-05 DOI: 10.1007/s10472-024-09941-x
Serdar Kadıoğlu, Bernard Kleynhans, Xin Wang

Recommender Systems have become the backbone of personalized services that provide tailored experiences to individual users, yet designing new recommendation applications with limited or no available training data remains a challenge. To address this issue, we focus on selecting the universe of items for experimentation in recommender systems by leveraging a recently introduced combinatorial problem. On the one hand, selecting a large set of items is desirable to increase the diversity of items. On the other hand, a smaller set of items enables rapid experimentation and minimizes the time and the amount of data required to train machine learning models. We first present how to optimize for such conflicting criteria using a multi-level optimization framework. Then, we shift our focus to the operational setting of a recommender system. In practice, to work effectively in a dynamic environment where new items are introduced to the system, we need to explore users’ behaviors and interests continuously. To that end, we show how to integrate the item selection approach with active learning to guide randomized exploration in an ongoing fashion. Our hybrid approach combines techniques from discrete optimization, unsupervised clustering, and latent text embeddings. Experimental results on well-known movie and book recommendation benchmarks demonstrate the benefits of optimized item selection and efficient exploration.

推荐系统已成为为个人用户提供量身定制体验的个性化服务的支柱,然而,在训练数据有限或没有训练数据的情况下设计新的推荐应用仍然是一个挑战。为了解决这个问题,我们利用最近提出的一个组合问题,重点研究了在推荐系统中选择实验项目的问题。一方面,选择一个大的项目集可以增加项目的多样性。另一方面,较小的项目集可以实现快速实验,并最大限度地减少训练机器学习模型所需的时间和数据量。我们首先介绍了如何利用多层次优化框架对这些相互冲突的标准进行优化。然后,我们将重点转向推荐系统的运行环境。在实践中,为了在系统不断引入新项目的动态环境中有效工作,我们需要不断探索用户的行为和兴趣。为此,我们展示了如何将项目选择方法与主动学习相结合,以持续的方式指导随机探索。我们的混合方法结合了离散优化、无监督聚类和潜在文本嵌入等技术。在著名的电影和图书推荐基准上的实验结果证明了优化项目选择和高效探索的好处。
{"title":"Integrating optimized item selection with active learning for continuous exploration in recommender systems","authors":"Serdar Kadıoğlu, Bernard Kleynhans, Xin Wang","doi":"10.1007/s10472-024-09941-x","DOIUrl":"https://doi.org/10.1007/s10472-024-09941-x","url":null,"abstract":"<p>Recommender Systems have become the backbone of personalized services that provide tailored experiences to individual users, yet designing new recommendation applications with limited or no available training data remains a challenge. To address this issue, we focus on selecting the universe of items for experimentation in recommender systems by leveraging a recently introduced combinatorial problem. On the one hand, selecting a large set of items is desirable to increase the diversity of items. On the other hand, a smaller set of items enables rapid experimentation and minimizes the time and the amount of data required to train machine learning models. We first present how to optimize for such conflicting criteria using a multi-level optimization framework. Then, we shift our focus to the operational setting of a recommender system. In practice, to work effectively in a dynamic environment where new items are introduced to the system, we need to explore users’ behaviors and interests continuously. To that end, we show how to integrate the item selection approach with active learning to guide randomized exploration in an ongoing fashion. Our hybrid approach combines techniques from discrete optimization, unsupervised clustering, and latent text embeddings. Experimental results on well-known movie and book recommendation benchmarks demonstrate the benefits of optimized item selection and efficient exploration.</p>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"105 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140582298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-resolution continuous normalizing flows 多分辨率连续归一化流程
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-03-21 DOI: 10.1007/s10472-024-09939-5
Vikram Voleti, Chris Finlay, Adam Oberman, Christopher Pal

Recent work has shown that Neural Ordinary Differential Equations (ODEs) can serve as generative models of images using the perspective of Continuous Normalizing Flows (CNFs). Such models offer exact likelihood calculation, and invertible generation/density estimation. In this work we introduce a Multi-Resolution variant of such models (MRCNF), by characterizing the conditional distribution over the additional information required to generate a fine image that is consistent with the coarse image. We introduce a transformation between resolutions that allows for no change in the log likelihood. We show that this approach yields comparable likelihood values for various image datasets, with improved performance at higher resolutions, with fewer parameters, using only one GPU. Further, we examine the out-of-distribution properties of MRCNFs, and find that they are similar to those of other likelihood-based generative models.

最近的研究表明,神经常微分方程(ODE)可以从连续归一化流(CNF)的角度作为图像的生成模型。这种模型提供精确的似然计算和可反演的生成/密度估算。在这项工作中,我们通过描述生成与粗略图像一致的精细图像所需的附加信息的条件分布,引入了此类模型的多分辨率变体(MRCNF)。我们引入了分辨率之间的转换,这种转换不会改变对数似然。我们的研究表明,这种方法可为各种图像数据集生成可比的似然值,在分辨率更高、参数更少、仅使用一个 GPU 的情况下性能更佳。此外,我们还检查了 MRCNFs 的分布外特性,发现它们与其他基于似然法的生成模型类似。
{"title":"Multi-resolution continuous normalizing flows","authors":"Vikram Voleti,&nbsp;Chris Finlay,&nbsp;Adam Oberman,&nbsp;Christopher Pal","doi":"10.1007/s10472-024-09939-5","DOIUrl":"10.1007/s10472-024-09939-5","url":null,"abstract":"<div><p>Recent work has shown that Neural Ordinary Differential Equations (ODEs) can serve as generative models of images using the perspective of Continuous Normalizing Flows (CNFs). Such models offer exact likelihood calculation, and invertible generation/density estimation. In this work we introduce a Multi-Resolution variant of such models (MRCNF), by characterizing the conditional distribution over the additional information required to generate a fine image that is consistent with the coarse image. We introduce a transformation between resolutions that allows for no change in the log likelihood. We show that this approach yields comparable likelihood values for various image datasets, with improved performance at higher resolutions, with fewer parameters, using only one GPU. Further, we examine the out-of-distribution properties of MRCNFs, and find that they are similar to those of other likelihood-based generative models.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1295 - 1317"},"PeriodicalIF":1.2,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140203649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Clustering, coding, and the concept of similarity 聚类、编码和相似性概念
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-03-19 DOI: 10.1007/s10472-024-09929-7
L. Thorne McCarty

This paper develops a theory of clustering and coding that combines a geometric model with a probabilistic model in a principled way. The geometric model is a Riemannian manifold with a Riemannian metric, ({g}_{ij}(textbf{x})), which we interpret as a measure of dissimilarity. The probabilistic model consists of a stochastic process with an invariant probability measure that matches the density of the sample input data. The link between the two models is a potential function, (U(textbf{x})), and its gradient, (nabla U(textbf{x})). We use the gradient to define the dissimilarity metric, which guarantees that our measure of dissimilarity will depend on the probability measure. Finally, we use the dissimilarity metric to define a coordinate system on the embedded Riemannian manifold, which gives us a low-dimensional encoding of our original data.

本文提出了一种聚类和编码理论,它以一种原则性的方式将几何模型与概率模型相结合。几何模型是一个具有黎曼度量的黎曼流形,我们将其解释为异质性度量。概率模型包括一个随机过程,其不变概率度量与样本输入数据的密度相匹配。这两个模型之间的联系是一个势函数(U(textbf{x}))及其梯度(U(textbf{x}))。我们使用梯度来定义相似度量,这保证了我们的相似度量将取决于概率度量。最后,我们利用异质性度量定义嵌入黎曼流形上的坐标系,从而得到原始数据的低维编码。
{"title":"Clustering, coding, and the concept of similarity","authors":"L. Thorne McCarty","doi":"10.1007/s10472-024-09929-7","DOIUrl":"10.1007/s10472-024-09929-7","url":null,"abstract":"<div><p>This paper develops a theory of <i>clustering</i> and <i>coding</i> that combines a geometric model with a probabilistic model in a principled way. The geometric model is a Riemannian manifold with a Riemannian metric, <span>({g}_{ij}(textbf{x}))</span>, which we interpret as a measure of <i>dissimilarity</i>. The probabilistic model consists of a stochastic process with an invariant probability measure that matches the density of the sample input data. The link between the two models is a potential function, <span>(U(textbf{x}))</span>, and its gradient, <span>(nabla U(textbf{x}))</span>. We use the gradient to define the dissimilarity metric, which guarantees that our measure of dissimilarity will depend on the probability measure. Finally, we use the dissimilarity metric to define a coordinate system on the embedded Riemannian manifold, which gives us a low-dimensional encoding of our original data.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 5","pages":"1197 - 1248"},"PeriodicalIF":1.2,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140172695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coalition formation games – preface 联盟组建游戏--序言
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-03-07 DOI: 10.1007/s10472-024-09937-7
Judy Goldsmith, Jörg Rothe
{"title":"Coalition formation games – preface","authors":"Judy Goldsmith,&nbsp;Jörg Rothe","doi":"10.1007/s10472-024-09937-7","DOIUrl":"10.1007/s10472-024-09937-7","url":null,"abstract":"","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 3","pages":"535 - 537"},"PeriodicalIF":1.2,"publicationDate":"2024-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09937-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140076852","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A new definition for feature selection stability analysis 特征选择稳定性分析的新定义
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-03-01 DOI: 10.1007/s10472-024-09936-8
Teddy Lazebnik, Avi Rosenfeld

Feature selection (FS) stability is an important topic of recent interest. Finding stable features is important for creating reliable, non-overfitted feature sets, which in turn can be used to generate machine learning models with better accuracy and explanations and are less prone to adversarial attacks. There are currently several definitions of FS stability that are widely used. In this paper, we demonstrate that existing stability metrics fail to quantify certain key elements of many datasets such as resilience to data drift or non-uniformly distributed missing values. To address this shortcoming, we propose a new definition for FS stability inspired by Lyapunov stability in dynamic systems. We show the proposed definition is statistically different from the classical record-stability on ((n=90)) datasets. We present the advantages and disadvantages of using Lyapunov and other stability definitions and demonstrate three scenarios in which each one of the three proposed stability metrics is best suited.

特征选择(FS)的稳定性是近期备受关注的一个重要话题。找到稳定的特征对于创建可靠、非过度拟合的特征集非常重要,而这些特征集又可用于生成具有更高精度和解释力的机器学习模型,并且不易受到对抗性攻击。目前有几种关于 FS 稳定性的定义被广泛使用。在本文中,我们证明了现有的稳定性指标无法量化许多数据集的某些关键因素,例如对数据漂移或非均匀分布的缺失值的适应能力。针对这一缺陷,我们从动态系统的 Lyapunov 稳定性中汲取灵感,提出了 FS 稳定性的新定义。我们证明了所提出的定义在统计上不同于((n=90))数据集上的经典记录稳定性。我们介绍了使用李亚普诺夫稳定性定义和其他稳定性定义的优缺点,并展示了三种场景,在这些场景中,所提出的三种稳定性度量中的每一种都是最合适的。
{"title":"A new definition for feature selection stability analysis","authors":"Teddy Lazebnik,&nbsp;Avi Rosenfeld","doi":"10.1007/s10472-024-09936-8","DOIUrl":"10.1007/s10472-024-09936-8","url":null,"abstract":"<div><p>Feature selection (FS) stability is an important topic of recent interest. Finding stable features is important for creating reliable, non-overfitted feature sets, which in turn can be used to generate machine learning models with better accuracy and explanations and are less prone to adversarial attacks. There are currently several definitions of FS stability that are widely used. In this paper, we demonstrate that existing stability metrics fail to quantify certain key elements of many datasets such as resilience to data drift or non-uniformly distributed missing values. To address this shortcoming, we propose a new definition for FS stability inspired by Lyapunov stability in dynamic systems. We show the proposed definition is statistically different from the classical <i>record-stability</i> on (<span>(n=90)</span>) datasets. We present the advantages and disadvantages of using Lyapunov and other stability definitions and demonstrate three scenarios in which each one of the three proposed stability metrics is best suited.</p></div>","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 3","pages":"753 - 770"},"PeriodicalIF":1.2,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09936-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140017398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction: Personalized choice prediction with less user information 更正:利用较少的用户信息进行个性化选择预测
IF 1.2 4区 计算机科学 Q4 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-02-20 DOI: 10.1007/s10472-024-09934-w
Francine Chen, Yanxia Zhang, Minh Nguyen, Matt Klenk, Charlene Wu
{"title":"Correction: Personalized choice prediction with less user information","authors":"Francine Chen,&nbsp;Yanxia Zhang,&nbsp;Minh Nguyen,&nbsp;Matt Klenk,&nbsp;Charlene Wu","doi":"10.1007/s10472-024-09934-w","DOIUrl":"10.1007/s10472-024-09934-w","url":null,"abstract":"","PeriodicalId":7971,"journal":{"name":"Annals of Mathematics and Artificial Intelligence","volume":"92 3","pages":"771 - 771"},"PeriodicalIF":1.2,"publicationDate":"2024-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10472-024-09934-w.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142412597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Annals of Mathematics and Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1