首页 > 最新文献

Doklady Mathematics最新文献

英文 中文
SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation SwiftDepth++:用于精确深度估计的高效轻量级模型
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602038
Y. Dayoub, I. Makarov

Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.

深度估计是各个领域的关键任务,但收集标记深度数据的高成本导致人们对自监督单目深度估计方法的兴趣越来越大。在本文中,我们介绍了SwiftDepth++,这是一个轻量级的深度估计模型,在保持低计算预算的同时提供有竞争力的结果。SwiftDepth++的核心创新在于其新颖的深度解码器,在保留重要信息的同时,通过快速压缩特征来提高效率。此外,我们结合了一个师生知识蒸馏框架,指导学生模型改进其预测。我们在KITTI和NYU数据集上对swiftdeep++进行了评估,在没有微调的情况下,它在KITTI数据集上的绝对相对误差(Abs_rel)为10.2%,在NYU数据集上的绝对相对误差(Abs_rel)为22%,所有参数都有大约600万个参数。这些结果表明,SwiftDepth++不仅满足现代深度估计任务的要求,而且显著降低了计算复杂度,使其成为现实应用的实用选择。
{"title":"SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation","authors":"Y. Dayoub,&nbsp;I. Makarov","doi":"10.1134/S1064562424602038","DOIUrl":"10.1134/S1064562424602038","url":null,"abstract":"<p>Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S162 - S171"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602038.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rethinking Graph Classification Problem in Presence of Isomorphism 存在同构的图分类问题的再思考
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602385
S. Ivanov, S. Sviridov, E. Burnaev

There is an increasing interest in developing new models for graph classification problem that serves as a common benchmark for evaluation and comparison of GNNs and graph kernels. To ensure a fair comparison of the models several commonly used datasets exist and current assessments and conclusions rely on the validity of these datasets. However, as we show in this paper majority of these datasets contain isomorphic copies of the data points, which can lead to misleading conclusions. For example, the relative ranking of the graph models can change substantially if we remove isomorphic graphs in the test set.

To mitigate this we present several results. We show that explicitly incorporating the knowledge of isomorphism in the datasets can significantly boost the performance of any graph model. Finally, we re-evaluate commonly used graph models on refined graph datasets and provide recommendations for designing new datasets and metrics for graph classification problem.

人们对开发图分类问题的新模型越来越感兴趣,这些模型可以作为评估和比较gnn和图核的通用基准。为了确保模型之间的公平比较,存在几种常用的数据集,目前的评估和结论依赖于这些数据集的有效性。然而,正如我们在本文中所展示的,这些数据集中的大多数包含数据点的同构副本,这可能导致误导性的结论。例如,如果我们在测试集中删除同构图,图模型的相对排名会发生很大的变化。为了减轻这种情况,我们提出了几个结果。我们表明,在数据集中显式地结合同构知识可以显着提高任何图模型的性能。最后,我们在改进的图数据集上重新评估了常用的图模型,并为图分类问题设计新的数据集和度量提供了建议。
{"title":"Rethinking Graph Classification Problem in Presence of Isomorphism","authors":"S. Ivanov,&nbsp;S. Sviridov,&nbsp;E. Burnaev","doi":"10.1134/S1064562424602385","DOIUrl":"10.1134/S1064562424602385","url":null,"abstract":"<p>There is an increasing interest in developing new models for graph classification problem that serves as a common benchmark for evaluation and comparison of GNNs and graph kernels. To ensure a fair comparison of the models several commonly used datasets exist and current assessments and conclusions rely on the validity of these datasets. However, as we show in this paper majority of these datasets contain isomorphic copies of the data points, which can lead to misleading conclusions. For example, the relative ranking of the graph models can change substantially if we remove isomorphic graphs in the test set.</p><p>To mitigate this we present several results. We show that explicitly incorporating the knowledge of isomorphism in the datasets can significantly boost the performance of any graph model. Finally, we re-evaluate commonly used graph models on refined graph datasets and provide recommendations for designing new datasets and metrics for graph classification problem.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S312 - S331"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602385.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RuSciBench: Open Benchmark for Russian and English Scientific Document Representations RuSciBench:俄语和英语科学文献表示的开放基准
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602191
A. Vatolin, N. Gerasimenko, A. Ianina, K. Vorontsov

Sharing scientific knowledge in the community is an important endeavor. However, most papers are written in English, which makes dissemination of knowledge in countries where English is not spoken by the majority of people harder. Nowadays, machine translation and language models may help to solve this problem, but it is still complicated to train and evaluate models in languages other than English with no or little data in the required language. To address this, we propose the first benchmark for evaluating models on scientific texts in Russian. It consists of papers from Russian electronic library of scientific publications. We also present a set of tasks which can be used to fine-tune various models on our data and provide a detailed comparison between state-of-the-art models on our benchmark.

在社区中分享科学知识是一项重要的努力。然而,大多数论文都是用英语写的,这使得在大多数人不讲英语的国家传播知识变得更加困难。如今,机器翻译和语言模型可能有助于解决这个问题,但在没有或很少有所需语言数据的情况下,用英语以外的语言训练和评估模型仍然很复杂。为了解决这个问题,我们提出了评估俄语科学文本模型的第一个基准。它由来自俄罗斯科学出版物电子图书馆的论文组成。我们还提出了一组任务,可用于对数据上的各种模型进行微调,并在基准上提供最先进模型之间的详细比较。
{"title":"RuSciBench: Open Benchmark for Russian and English Scientific Document Representations","authors":"A. Vatolin,&nbsp;N. Gerasimenko,&nbsp;A. Ianina,&nbsp;K. Vorontsov","doi":"10.1134/S1064562424602191","DOIUrl":"10.1134/S1064562424602191","url":null,"abstract":"<p>Sharing scientific knowledge in the community is an important endeavor. However, most papers are written in English, which makes dissemination of knowledge in countries where English is not spoken by the majority of people harder. Nowadays, machine translation and language models may help to solve this problem, but it is still complicated to train and evaluate models in languages other than English with no or little data in the required language. To address this, we propose the first benchmark for evaluating models on scientific texts in Russian. It consists of papers from Russian electronic library of scientific publications. We also present a set of tasks which can be used to fine-tune various models on our data and provide a detailed comparison between state-of-the-art models on our benchmark.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S251 - S260"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602191.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hiding Backdoors within Event Sequence Data via Poisoning Attacks 通过中毒攻击在事件序列数据中隐藏后门
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602221
A. Ermilova, E. Kovtun, D. Berestnev, A. Zaytsev

Deep learning’s emerging role in the financial sector’s decision-making introduces risks of adversarial attacks. A specific threat is a poisoning attack that modifies the training sample to develop a backdoor that persists during model usage. However, data cleaning procedures and routine model checks are easy-to-implement actions that prevent the usage of poisoning attacks. The problem is even more challenging for event sequence models, for which it is hard to design an attack due to the discrete nature of the data. We start with a general investigation of the possibility of poisoning for event sequence models. Then, we propose a concealed poisoning attack that can bypass natural banks’ defences. The empirical investigation shows that the developed poisoned model trained on contaminated data passes the check procedure, being similar to a clean model, and simultaneously contains a simple to-implement backdoor.

深度学习在金融部门决策中的新兴作用带来了对抗性攻击的风险。一个特定的威胁是中毒攻击,它修改训练样本以开发一个在模型使用期间持续存在的后门。但是,数据清理过程和常规模型检查是易于实现的操作,可以防止使用中毒攻击。对于事件序列模型来说,这个问题更具挑战性,由于数据的离散性,很难设计攻击。我们从事件序列模型中毒可能性的一般调查开始。然后,我们提出了一种隐蔽的投毒攻击,可以绕过天然银行的防御。实证研究表明,在污染数据上训练的开发的有毒模型通过了检查程序,与干净模型相似,同时包含一个易于实现的后门。
{"title":"Hiding Backdoors within Event Sequence Data via Poisoning Attacks","authors":"A. Ermilova,&nbsp;E. Kovtun,&nbsp;D. Berestnev,&nbsp;A. Zaytsev","doi":"10.1134/S1064562424602221","DOIUrl":"10.1134/S1064562424602221","url":null,"abstract":"<p>Deep learning’s emerging role in the financial sector’s decision-making introduces risks of adversarial attacks. A specific threat is a poisoning attack that modifies the training sample to develop a backdoor that persists during model usage. However, data cleaning procedures and routine model checks are easy-to-implement actions that prevent the usage of poisoning attacks. The problem is even more challenging for event sequence models, for which it is hard to design an attack due to the discrete nature of the data. We start with a general investigation of the possibility of poisoning for event sequence models. Then, we propose a concealed poisoning attack that can bypass natural banks’ defences. The empirical investigation shows that the developed poisoned model trained on contaminated data passes the check procedure, being similar to a clean model, and simultaneously contains a simple to-implement backdoor.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S288 - S298"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602221.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
About Modifications of the Loss Function for the Causal Training of Physics-Informed Neural Networks 关于物理信息神经网络因果训练中损失函数的修正
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S106456242460194X
V. A. Es’kin, D. V. Davydov, E. D. Egorova, A. O. Malkhanov, M. A. Akhukov, M. E. Smorkalov

A method is presented that allows to reduce a problem described by differential equations with initial and boundary conditions to a problem described only by differential equations which encapsulate initial and boundary conditions. It becomes possible to represent the loss function for physics-informed neural networks (PINNs) methodology in the form of a single term associated with modified differential equations. Thus eliminating the need to tune the scaling coefficients for the terms of loss function related to boundary and initial conditions. The weighted loss functions respecting causality were modified and new weighted loss functions, based on generalized functions, are derived. Numerical experiments have been carried out for a number of problems, demonstrating the accuracy of the proposed approaches. The neural network architecture was proposed for the Korteweg–De Vries equation, which is more relevant for this problem under consideration, and it demonstrates superior extrapolation of the solution in the space-time domain where training was not performed.

提出了一种将具有初始和边界条件的微分方程所描述的问题简化为仅包含初始和边界条件的微分方程所描述的问题的方法。用与修正微分方程相关的单个项的形式来表示物理信息神经网络(pinn)方法的损失函数成为可能。从而消除了调整与边界和初始条件相关的损失函数项的缩放系数的需要。对考虑因果关系的加权损失函数进行了修正,在广义函数的基础上导出了新的加权损失函数。对一些问题进行了数值实验,证明了所提出方法的准确性。提出了Korteweg-De Vries方程的神经网络体系结构,该体系结构与所考虑的问题更为相关,并且在未进行训练的时空域中具有较好的解外推性。
{"title":"About Modifications of the Loss Function for the Causal Training of Physics-Informed Neural Networks","authors":"V. A. Es’kin,&nbsp;D. V. Davydov,&nbsp;E. D. Egorova,&nbsp;A. O. Malkhanov,&nbsp;M. A. Akhukov,&nbsp;M. E. Smorkalov","doi":"10.1134/S106456242460194X","DOIUrl":"10.1134/S106456242460194X","url":null,"abstract":"<p>A method is presented that allows to reduce a problem described by differential equations with initial and boundary conditions to a problem described only by differential equations which encapsulate initial and boundary conditions. It becomes possible to represent the loss function for physics-informed neural networks (PINNs) methodology in the form of a single term associated with modified differential equations. Thus eliminating the need to tune the scaling coefficients for the terms of loss function related to boundary and initial conditions. The weighted loss functions respecting causality were modified and new weighted loss functions, based on generalized functions, are derived. Numerical experiments have been carried out for a number of problems, demonstrating the accuracy of the proposed approaches. The neural network architecture was proposed for the Korteweg–De Vries equation, which is more relevant for this problem under consideration, and it demonstrates superior extrapolation of the solution in the space-time domain where training was not performed.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S172 - S192"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S106456242460194X.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal 3D Map Reconstruction for Intelligent Robotcs Using Neural Network-Based Methods 基于神经网络的智能机器人多模态三维地图重建方法
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602014
D. A. Yudin

Methods for constructing multimodal 3D maps are becoming increasingly important for robot navigation systems. In such maps, each 3D point or object contains, in addition to color and semantic category information, compressed vector representations of a text description or sound. This allows solving problems of moving to objects based on natural language queries, even those that do not explicitly mention the object. This article proposes an original taxonomy of methods that allow constructing multimodal 3D maps using neural network methods. It is shown that sparse methods that use a scene representation in the form of an object graph and large language models to find an answer to spatial and semantic queries demonstrate the most promising results on existing open benchmarks. Based on the analysis, recommendations are formulated for choosing certain methods for solving practical problems of intelligent robotics.

对于机器人导航系统来说,构建多模态三维地图的方法正变得越来越重要。在这类地图中,每个三维点或物体除了包含颜色和语义类别信息外,还包含文本描述或声音的压缩矢量表示。这样就可以解决根据自然语言查询移动到物体的问题,即使是那些没有明确提到物体的查询。本文提出了一种独创的方法分类法,可以利用神经网络方法构建多模态三维地图。结果表明,使用对象图形式的场景表示和大型语言模型来寻找空间和语义查询答案的稀疏方法在现有的公开基准测试中取得了最有前途的结果。根据分析结果,提出了选择某些方法解决智能机器人实际问题的建议。
{"title":"Multimodal 3D Map Reconstruction for Intelligent Robotcs Using Neural Network-Based Methods","authors":"D. A. Yudin","doi":"10.1134/S1064562424602014","DOIUrl":"10.1134/S1064562424602014","url":null,"abstract":"<p>Methods for constructing multimodal 3D maps are becoming increasingly important for robot navigation systems. In such maps, each 3D point or object contains, in addition to color and semantic category information, compressed vector representations of a text description or sound. This allows solving problems of moving to objects based on natural language queries, even those that do not explicitly mention the object. This article proposes an original taxonomy of methods that allow constructing multimodal 3D maps using neural network methods. It is shown that sparse methods that use a scene representation in the form of an object graph and large language models to find an answer to spatial and semantic queries demonstrate the most promising results on existing open benchmarks. Based on the analysis, recommendations are formulated for choosing certain methods for solving practical problems of intelligent robotics.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S117 - S125"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602014.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Search for Optimal Architecture of Physics-Informed Neural Networks Using Differential Evolution Algorithm 用差分进化算法搜索物理信息神经网络的最优结构
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602300
F. A. Buzaev, D. S. Efremenko, I. A. Chuprov, Ya. N. Khassan, E. N. Kazakov, J. Gao

The accuracy of solving partial differential equations using physics-informed neural networks (PINNs) significantly depends on their architecture and the choice of hyperparameters. However, manually searching for the optimal configuration can be difficult due to the high computational complexity. In this paper, we propose an approach for optimizing the PINN architecture using a differential evolution algorithm. We focus on optimizing over a small number of training epochs, which allows us to consider a wider range of configurations while reducing the computational cost. The number of epochs is chosen such that the accuracy of the model at the initial stage correlates with its accuracy after full training, which significantly speeds up the optimization process. To improve efficiency, we also apply a surrogate model based on a Gaussian process, which reduces the number of required PINN trainings. The paper presents the results of optimizing PINN architectures for solving various partial differential equations and offers recommendations for improving their performance.

利用物理信息神经网络(pinn)求解偏微分方程的精度在很大程度上取决于它们的结构和超参数的选择。然而,由于计算复杂度高,手动搜索最优配置可能很困难。在本文中,我们提出了一种使用差分进化算法优化PINN架构的方法。我们专注于在少量的训练周期上进行优化,这使我们能够在降低计算成本的同时考虑更广泛的配置。epoch的选择使得模型在初始阶段的精度与完全训练后的精度相关,这大大加快了优化过程。为了提高效率,我们还应用了基于高斯过程的代理模型,这减少了所需的PINN训练次数。本文介绍了求解各种偏微分方程的PINN体系结构的优化结果,并提出了改进其性能的建议。
{"title":"Search for Optimal Architecture of Physics-Informed Neural Networks Using Differential Evolution Algorithm","authors":"F. A. Buzaev,&nbsp;D. S. Efremenko,&nbsp;I. A. Chuprov,&nbsp;Ya. N. Khassan,&nbsp;E. N. Kazakov,&nbsp;J. Gao","doi":"10.1134/S1064562424602300","DOIUrl":"10.1134/S1064562424602300","url":null,"abstract":"<p>The accuracy of solving partial differential equations using physics-informed neural networks (PINNs) significantly depends on their architecture and the choice of hyperparameters. However, manually searching for the optimal configuration can be difficult due to the high computational complexity. In this paper, we propose an approach for optimizing the PINN architecture using a differential evolution algorithm. We focus on optimizing over a small number of training epochs, which allows us to consider a wider range of configurations while reducing the computational cost. The number of epochs is chosen such that the accuracy of the model at the initial stage correlates with its accuracy after full training, which significantly speeds up the optimization process. To improve efficiency, we also apply a surrogate model based on a Gaussian process, which reduces the number of required PINN trainings. The paper presents the results of optimizing PINN architectures for solving various partial differential equations and offers recommendations for improving their performance.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S8 - S14"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine Search Problem for Mathematical Expression 数学表达式的机器搜索问题
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602087
A. I. Diveev, E. A. Sofronova

The problem of searching for a mathematical expression is formulated. Classes of mathematical problems where this problem is in demand are presented. A general approach to solving this problem based on machine learning using symbolic regression methods is given. This approach allows finding not only the parameters of the desired mathematical expression, but also its structure. The general problem of machine learning by methods of symbolic regression is described and a unified approach to overcoming it based on the application of the principle of small variations of the basis solution is given.

寻找数学表达式的问题被公式化了。提出了需要这个问题的数学问题的类别。给出了一种基于机器学习的符号回归方法来解决这一问题的一般方法。这种方法不仅可以找到所需数学表达式的参数,还可以找到它的结构。描述了用符号回归方法进行机器学习的一般问题,并给出了基于基解小变分原理的统一解决方法。
{"title":"Machine Search Problem for Mathematical Expression","authors":"A. I. Diveev,&nbsp;E. A. Sofronova","doi":"10.1134/S1064562424602087","DOIUrl":"10.1134/S1064562424602087","url":null,"abstract":"<p>The problem of searching for a mathematical expression is formulated. Classes of mathematical problems where this problem is in demand are presented. A general approach to solving this problem based on machine learning using symbolic regression methods is given. This approach allows finding not only the parameters of the desired mathematical expression, but also its structure. The general problem of machine learning by methods of symbolic regression is described and a unified approach to overcoming it based on the application of the principle of small variations of the basis solution is given.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S25 - S34"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602087.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stack More LLM’s: Efficient Detection of Machine-Generated Texts via Perplexity Approximation 堆栈更多的 LLM:通过易错性近似高效检测机器生成的文本
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602075
G. M. Gritsai, I. A. Khabutdinov, A. V. Grabovoy

The development of large language models (LLMs) is currently receiving a great amount of interest, but an update of text generation methods should entail a continuous update of methods for detecting machine-generated texts. Earlier, it has been highlighted that values of perplexity and log-probability are able to capture a measure of the difference between artificial and human-written texts. Using this observation, we define a new criterion based on these two values to judge whether a passage is generated from a given LLM. In this paper, we propose a novel efficient method that enables the detection of machine-generated fragments using an approximation of the LLM perplexity value based on pre-collected statistical language models. Approximation lends a hand in achieving high performance and quality metrics also on fragments from weights-closed LLMs. A large number of pre-collected statistical dictionaries results in an increased generalisation ability and the possibility to cover text sequences from the wild. Such approach is easy to update by only adding a new dictionary with latest model text outputs. The presented method has a high performance and achieves quality with an average of 94% recall in detecting generated fragments among texts from various open-source LLMs. In addition, the method is able to perform in milliseconds, which outperforms state-of-the-art models by a factor of thousands.

大型语言模型(llm)的开发目前受到了极大的关注,但是文本生成方法的更新需要不断更新检测机器生成文本的方法。早些时候,已经强调了困惑度和对数概率值能够捕捉到人工文本和人类文本之间差异的度量。利用这一观察结果,我们基于这两个值定义了一个新的标准来判断一个段落是否由给定的LLM生成。在本文中,我们提出了一种新的有效方法,可以使用基于预先收集的统计语言模型的LLM困惑值的近似值来检测机器生成的片段。近似有助于在权重封闭llm的片段上实现高性能和质量指标。大量预先收集的统计字典提高了泛化能力,并有可能涵盖来自野外的文本序列。这种方法很容易更新,只需添加一个带有最新模型文本输出的新字典。该方法在检测各种开源llm文本中生成的片段时,具有较高的性能,平均召回率达到94%。此外,该方法能够在毫秒内执行,这比最先进的模型要好几千倍。
{"title":"Stack More LLM’s: Efficient Detection of Machine-Generated Texts via Perplexity Approximation","authors":"G. M. Gritsai,&nbsp;I. A. Khabutdinov,&nbsp;A. V. Grabovoy","doi":"10.1134/S1064562424602075","DOIUrl":"10.1134/S1064562424602075","url":null,"abstract":"<p>The development of large language models (LLMs) is currently receiving a great amount of interest, but an update of text generation methods should entail a continuous update of methods for detecting machine-generated texts. Earlier, it has been highlighted that values of perplexity and log-probability are able to capture a measure of the difference between artificial and human-written texts. Using this observation, we define a new criterion based on these two values to judge whether a passage is generated from a given LLM. In this paper, we propose a novel efficient method that enables the detection of machine-generated fragments using an approximation of the LLM perplexity value based on pre-collected statistical language models. Approximation lends a hand in achieving high performance and quality metrics also on fragments from weights-closed LLMs. A large number of pre-collected statistical dictionaries results in an increased generalisation ability and the possibility to cover text sequences from the wild. Such approach is easy to update by only adding a new dictionary with latest model text outputs. The presented method has a high performance and achieves quality with an average of 94% recall in detecting generated fragments among texts from various open-source LLMs. In addition, the method is able to perform in milliseconds, which outperforms state-of-the-art models by a factor of thousands.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S203 - S211"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602075.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Empirical Scrutinization of Four Crisp Clustering Methods with Four Distance Metrics and One Straightforward Interpretation Rule 用四种距离度量和一种直接解释规则对四种简洁聚类方法进行实证检验
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602002
T. A. Alvandyan, S. Shalileh

Clustering has always been in great demand by scientific and industrial communities. However, due to the lack of ground truth, interpreting its obtained results can be debatable. The current research provides an empirical benchmark on the efficiency of three popular and one recently proposed crisp clustering methods. To this end, we extensively analyzed these (four) methods by applying them to nine real-world and 420 synthetic datasets using four different values of p in Minkowski distance. Furthermore, we validated a previously proposed yet not well-known straightforward rule to interpret the recovered clusters. Our computations showed (i) Nesterov gradient descent clustering is the most effective clustering method using our real-world data, while K-Means had edge over it using our synthetic data; (ii) Minkowski distance with p = 1 is the most effective distance function, (iii) the investigated cluster interpretation rule is intuitive and valid.

科学界和工业界一直对集群有很大的需求。然而,由于缺乏基础真理,解释其获得的结果可能是有争议的。本研究对三种流行的聚类方法和最近提出的一种清晰聚类方法的效率提供了一个经验基准。为此,我们使用四种不同的闵可夫斯基距离p值,将这四种方法应用于9个真实数据集和420个合成数据集,对这四种方法进行了广泛的分析。此外,我们验证了先前提出的但不为人所知的简单规则来解释恢复的群集。我们的计算表明(i) Nesterov梯度下降聚类是使用我们真实世界数据的最有效的聚类方法,而K-Means在使用我们的合成数据时具有优势;(ii) p = 1的Minkowski距离是最有效的距离函数,(iii)所研究的聚类解释规则直观有效。
{"title":"An Empirical Scrutinization of Four Crisp Clustering Methods with Four Distance Metrics and One Straightforward Interpretation Rule","authors":"T. A. Alvandyan,&nbsp;S. Shalileh","doi":"10.1134/S1064562424602002","DOIUrl":"10.1134/S1064562424602002","url":null,"abstract":"<p>Clustering has always been in great demand by scientific and industrial communities. However, due to the lack of ground truth, interpreting its obtained results can be debatable. The current research provides an empirical benchmark on the efficiency of three popular and one recently proposed crisp clustering methods. To this end, we extensively analyzed these (four) methods by applying them to nine real-world and 420 synthetic datasets using four different values of <i>p</i> in Minkowski distance. Furthermore, we validated a previously proposed yet not well-known straightforward rule to interpret the recovered clusters. Our computations showed (i) Nesterov gradient descent clustering is the most effective clustering method using our real-world data, while K-Means had edge over it using our synthetic data; (ii) Minkowski distance with <i>p</i> = 1 is the most effective distance function, (iii) the investigated cluster interpretation rule is intuitive and valid.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S236 - S250"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602002.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Doklady Mathematics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1