首页 > 最新文献

Doklady Mathematics最新文献

英文 中文
Introductory Words of AI Journey Team AI之旅团队简介
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424702272
The AI Journey Team
{"title":"Introductory Words of AI Journey Team","authors":"The AI Journey Team","doi":"10.1134/S1064562424702272","DOIUrl":"10.1134/S1064562424702272","url":null,"abstract":"","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S1 - S1"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stratified Statistical Models in Hardware Reliability Analysis 分层统计模型在硬件可靠性分析中的应用
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424601963
I. A. Vasilev, I. O. Filimonova, M. I. Petrovskiy, I. V. Mashechkin

Reliability analysis is becoming paramount to the successful operation of systems. This paper considers the problem of hardware failure using hard disc drives and solid state drives as examples. Survivability analysis methods are used to predict hardware degradation by estimating the probability of an event occurring over time. Also, survival models account for incomplete data about the true time to event for censored observations. However, popular statistical methods do not account for features of real data such as the presence of outliers and categorical variables. In this paper, we propose to extend classical survival statistical methods by introducing an interpretable stratifying tree, each leaf of which corresponds to a statistical model. The experimental study is based on evaluating the dependence of the models’ quality as the depth of the tree increases. According to the experimental results, the proposed method outperforms classical statistical models. The results of the study demonstrate the effectiveness of the proposed approach and its potential in the field of reliability of complex technical systems.

可靠性分析对系统的成功运行至关重要。本文以硬盘驱动器和固态驱动器为例研究硬件故障问题。生存性分析方法用于通过估计事件随时间发生的概率来预测硬件退化。此外,生存模型解释了关于审查观察的事件真实时间的不完整数据。然而,流行的统计方法没有考虑到真实数据的特征,如异常值和分类变量的存在。在本文中,我们提出通过引入一个可解释的分层树来扩展经典的生存统计方法,该树的每个叶子对应一个统计模型。实验研究是基于评估模型质量随树深度增加的依赖性。实验结果表明,该方法优于经典统计模型。研究结果证明了该方法的有效性及其在复杂技术系统可靠性研究领域的潜力。
{"title":"Stratified Statistical Models in Hardware Reliability Analysis","authors":"I. A. Vasilev,&nbsp;I. O. Filimonova,&nbsp;M. I. Petrovskiy,&nbsp;I. V. Mashechkin","doi":"10.1134/S1064562424601963","DOIUrl":"10.1134/S1064562424601963","url":null,"abstract":"<p>Reliability analysis is becoming paramount to the successful operation of systems. This paper considers the problem of hardware failure using hard disc drives and solid state drives as examples. Survivability analysis methods are used to predict hardware degradation by estimating the probability of an event occurring over time. Also, survival models account for incomplete data about the true time to event for censored observations. However, popular statistical methods do not account for features of real data such as the presence of outliers and categorical variables. In this paper, we propose to extend classical survival statistical methods by introducing an interpretable stratifying tree, each leaf of which corresponds to a statistical model. The experimental study is based on evaluating the dependence of the models’ quality as the depth of the tree increases. According to the experimental results, the proposed method outperforms classical statistical models. The results of the study demonstrate the effectiveness of the proposed approach and its potential in the field of reliability of complex technical systems.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S103 - S109"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424601963.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neural Network Image Classifiers Informed by Factor Analyzers 基于因子分析的神经网络图像分类器
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S106456242460204X
A. M. Dostovalova, A. K. Gorshenin

The paper develops an approach to probability informing deep neural networks, that is, improving their results by using various probability models within architectural elements. We introduce factor analyzers with additive and impulse noise components as such models. The identifiability of the model is proved. The relationship between the parameter estimates by the methods of least squares and maximum likelihood is established, which actually means that the estimates of the parameters of the factor analyzer obtained within the informed block are unbiased and consistent. A mathematical model is used to create a new architectural element that implements the fusion of multiscale image features to improve classification accuracy in the case of a small volume of training data. This problem is typical for various applied tasks, including remote sensing data analysis. Various widely used neural network classifiers (EfficientNet, MobileNet, and Xception), both with and without a new informed block, are tested. It is demonstrated that on the open datasets UC Merced (remote sensing data) and Oxford Flowers (flower images), informed neural networks achieve a significant increase in accuracy for this class of tasks: the largest improvement in Top-1 accuracy was 6.67% (mean accuracy without informing equals 87.3%), while Top-5 accuracy increased by 1.49% (mean base accuracy value is 96.27%).

本文开发了一种概率通知深度神经网络的方法,即通过在架构元素中使用各种概率模型来改进其结果。我们介绍了以加性和脉冲噪声成分为模型的因子分析仪。证明了模型的可辨识性。建立了最小二乘法和最大似然法的参数估计之间的关系,这实际上意味着在知情块内得到的因子分析仪的参数估计是无偏和一致的。利用数学模型创建新的架构元素,实现多尺度图像特征的融合,在训练数据量小的情况下提高分类精度。这个问题是各种应用任务的典型问题,包括遥感数据分析。测试了各种广泛使用的神经网络分类器(EfficientNet, MobileNet和Xception),无论是否有新的通知块。结果表明,在开放数据集UC Merced(遥感数据)和Oxford Flowers(花卉图像)上,知情神经网络在这类任务上的准确率显著提高:Top-1的准确率最大提高了6.67%(未告知的平均准确率为87.3%),Top-5的准确率提高了1.49%(平均基础准确率值为96.27%)。
{"title":"Neural Network Image Classifiers Informed by Factor Analyzers","authors":"A. M. Dostovalova,&nbsp;A. K. Gorshenin","doi":"10.1134/S106456242460204X","DOIUrl":"10.1134/S106456242460204X","url":null,"abstract":"<p>The paper develops an approach to probability informing deep neural networks, that is, improving their results by using various probability models within architectural elements. We introduce factor analyzers with additive and impulse noise components as such models. The identifiability of the model is proved. The relationship between the parameter estimates by the methods of least squares and maximum likelihood is established, which actually means that the estimates of the parameters of the factor analyzer obtained within the informed block are unbiased and consistent. A mathematical model is used to create a new architectural element that implements the fusion of multiscale image features to improve classification accuracy in the case of a small volume of training data. This problem is typical for various applied tasks, including remote sensing data analysis. Various widely used neural network classifiers (EfficientNet, MobileNet, and Xception), both with and without a new informed block, are tested. It is demonstrated that on the open datasets UC Merced (remote sensing data) and Oxford Flowers (flower images), informed neural networks achieve a significant increase in accuracy for this class of tasks: the largest improvement in Top-1 accuracy was 6.67% (mean accuracy without informing equals 87.3%), while Top-5 accuracy increased by 1.49% (mean base accuracy value is 96.27%).</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S35 - S41"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S106456242460204X.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Solution of the Multimode Nonlinear Schrödinger Equation Using Physics-Informed Neural Networks 利用物理信息神经网络求解多模非线性Schrödinger方程
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602105
I. A. Chuprov, J. Gao, D. S. Efremenko, F. A. Buzaev, V. V. Zemlyakov

Single-mode optical fibers (SMFs) have become the foundation of modern communication systems. However, their capacity is expected to reach its theoretical limit in the near future. The use of multimode fibers (MMF) is seen as one of the most promising solutions to address this capacity deficit. The multimode nonlinear Schrödinger equation (MMNLSE) describing light propagation in MMF is significantly more complex than the equations for SMF, making numerical simulations of MMF-based systems computationally costly and impractical for most realistic scenarios. In this paper, we apply physics-informed neural networks (PINNs) to solve the MMNLSE. We show that a simple implementation of PINNs does not yield satisfactory results. We investigate the convergence of PINN and propose a novel scaling transformation for the zeroth-order dispersion coefficient that allows PINN to account for all important physical effects. Our calculations show good agreement with the Split-Step Fourier (SSF) method for fiber lengths of up to several hundred meters.

单模光纤(SMFs)已成为现代通信系统的基础。然而,它们的容量预计将在不久的将来达到理论极限。多模光纤(MMF)的使用被认为是解决这一容量不足的最有希望的解决方案之一。描述光在MMF中传播的多模非线性Schrödinger方程(MMNLSE)比SMF方程要复杂得多,这使得基于MMF的系统的数值模拟在大多数实际情况下计算成本高昂且不切实际。在本文中,我们应用物理通知神经网络(pinn)来解决MMNLSE问题。我们证明了pin n的简单实现不会产生令人满意的结果。我们研究了PINN的收敛性,并提出了一种新的零阶色散系数的缩放变换,该变换允许PINN考虑所有重要的物理效应。对于长度达几百米的光纤,我们的计算结果与分步傅里叶(SSF)方法很好地吻合。
{"title":"Solution of the Multimode Nonlinear Schrödinger Equation Using Physics-Informed Neural Networks","authors":"I. A. Chuprov,&nbsp;J. Gao,&nbsp;D. S. Efremenko,&nbsp;F. A. Buzaev,&nbsp;V. V. Zemlyakov","doi":"10.1134/S1064562424602105","DOIUrl":"10.1134/S1064562424602105","url":null,"abstract":"<p>Single-mode optical fibers (SMFs) have become the foundation of modern communication systems. However, their capacity is expected to reach its theoretical limit in the near future. The use of multimode fibers (MMF) is seen as one of the most promising solutions to address this capacity deficit. The multimode nonlinear Schrödinger equation (MMNLSE) describing light propagation in MMF is significantly more complex than the equations for SMF, making numerical simulations of MMF-based systems computationally costly and impractical for most realistic scenarios. In this paper, we apply physics-informed neural networks (PINNs) to solve the MMNLSE. We show that a simple implementation of PINNs does not yield satisfactory results. We investigate the convergence of PINN and propose a novel scaling transformation for the zeroth-order dispersion coefficient that allows PINN to account for all important physical effects. Our calculations show good agreement with the Split-Step Fourier (SSF) method for fiber lengths of up to several hundred meters.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S15 - S24"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602105.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deforming Implicit Neural Representation Generative Adversarial Network for Unsupervised Appearence Editing 变形隐式神经表示生成对抗网络的无监督外观编辑
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602373
S. Ignatiev, V. Egiazarian, R. Rakhimov, E. Burnaev

In this work, we present a new deep generative model for disentangling image shape from its appearance through differentiable warping. We propose to use implicit neural representations for modeling the deformation field and show that coordinate-based representations hold the necessary inductive bias. Unlike the previous warping-based approaches, which tend to model only local and small-scale displacements, our method is able to learn complex deformations and is not restricted to reversible mappings. We study the convergence of warping-based generative models and find that the high-frequency nature of the textures leads to shattered learning gradients, slow convergence, and suboptimal solutions. To cope with this problem, we propose to use invertible blurring, which smooths the gradients and leads to improved results. As a way to further facilitate the convergence of warping, we train the deformation module jointly as a vanilla GAN generator to guide the learning process in a self-distillation manner. Our complete pipeline shows decent results on the LSUN churches dataset. Finally, we demonstrate various applications of our model, like composable texture editing, controllable deformation editing, and keypoint detection.

在这项工作中,我们提出了一种新的深度生成模型,通过可微扭曲将图像形状与其外观分离开来。我们建议使用隐式神经表征来建模变形场,并表明基于坐标的表征具有必要的归纳偏差。与之前的基于翘曲的方法不同,这些方法倾向于只对局部和小尺度位移进行建模,我们的方法能够学习复杂的变形,并且不局限于可逆映射。我们研究了基于翘曲的生成模型的收敛性,发现纹理的高频特性导致了破碎的学习梯度、缓慢的收敛和次优解。为了解决这个问题,我们提出使用可逆模糊,平滑梯度,导致改善的结果。为了进一步促进翘曲的收敛,我们将变形模块作为香草GAN生成器联合训练,以自蒸馏的方式指导学习过程。我们完整的管道在LSUN教堂数据集上显示了不错的结果。最后,我们演示了模型的各种应用,如可组合纹理编辑、可控变形编辑和关键点检测。
{"title":"Deforming Implicit Neural Representation Generative Adversarial Network for Unsupervised Appearence Editing","authors":"S. Ignatiev,&nbsp;V. Egiazarian,&nbsp;R. Rakhimov,&nbsp;E. Burnaev","doi":"10.1134/S1064562424602373","DOIUrl":"10.1134/S1064562424602373","url":null,"abstract":"<p>In this work, we present a new deep generative model for disentangling image shape from its appearance through differentiable warping. We propose to use implicit neural representations for modeling the deformation field and show that coordinate-based representations hold the necessary inductive bias. Unlike the previous warping-based approaches, which tend to model only local and small-scale displacements, our method is able to learn complex deformations and is not restricted to reversible mappings. We study the convergence of warping-based generative models and find that the high-frequency nature of the textures leads to shattered learning gradients, slow convergence, and suboptimal solutions. To cope with this problem, we propose to use invertible blurring, which smooths the gradients and leads to improved results. As a way to further facilitate the convergence of warping, we train the deformation module jointly as a vanilla GAN generator to guide the learning process in a self-distillation manner. Our complete pipeline shows decent results on the LSUN churches dataset. Finally, we demonstrate various applications of our model, like composable texture editing, controllable deformation editing, and keypoint detection.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S299 - S311"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602373.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CRAFT: Cultural Russian-Oriented Dataset Adaptation for Focused Text-to-Image Generation CRAFT:以俄语文化为导向的数据集改编,用于文本到图像的集中生成
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602324
V. A. Vasilev, V. S. Arkhipkin, J. D. Agafonova, T. V. Nikulina, E. O. Mironova, A. A. Shichanina, N. A. Gerasimenko, M. A. Shoytov, D. V. Dimitrov

Despite the fact that popular text-to-image generation models cope well with international and general cultural queries, they have a significant knowledge gap regarding individual cultures. This is due to the content of existing large training datasets collected on the Internet, which are predominantly based on Western European or American popular culture. Meanwhile, the lack of cultural adaptation of the model can lead to incorrect results, a decrease in the generation quality, and the spread of stereotypes and offensive content. In an effort to address this issue, we examine the concept of cultural code and recognize the critical importance of its understanding by modern image generation models, an issue that has not been sufficiently addressed in the research community to date. We propose the methodology for collecting and processing the data necessary to form a dataset based on the cultural code, in particular the Russian one. We explore how the collected data affects the quality of generations in the national domain and analyze the effectiveness of our approach using the Kandinsky 3.1 text-to-image model. Human evaluation results demonstrate an increase in the level of awareness of Russian culture in the model.

尽管流行的文本到图像生成模型可以很好地处理国际和一般文化查询,但它们在个体文化方面存在显著的知识差距。这是由于在互联网上收集的现有大型训练数据集的内容主要基于西欧或美国流行文化。同时,缺乏对模型的文化适应,会导致不正确的结果,导致生成质量下降,以及刻板印象和攻击性内容的传播。为了解决这个问题,我们研究了文化代码的概念,并认识到现代图像生成模型对其理解的重要性,这是一个迄今为止在研究界尚未得到充分解决的问题。我们提出了收集和处理必要数据的方法,以形成基于文化代码的数据集,特别是俄罗斯文化代码。我们探讨了收集的数据如何影响国家领域的世代质量,并使用Kandinsky 3.1文本到图像模型分析了我们方法的有效性。人的评价结果表明,在该模型中,俄罗斯文化的意识水平有所提高。
{"title":"CRAFT: Cultural Russian-Oriented Dataset Adaptation for Focused Text-to-Image Generation","authors":"V. A. Vasilev,&nbsp;V. S. Arkhipkin,&nbsp;J. D. Agafonova,&nbsp;T. V. Nikulina,&nbsp;E. O. Mironova,&nbsp;A. A. Shichanina,&nbsp;N. A. Gerasimenko,&nbsp;M. A. Shoytov,&nbsp;D. V. Dimitrov","doi":"10.1134/S1064562424602324","DOIUrl":"10.1134/S1064562424602324","url":null,"abstract":"<p>Despite the fact that popular text-to-image generation models cope well with international and general cultural queries, they have a significant knowledge gap regarding individual cultures. This is due to the content of existing large training datasets collected on the Internet, which are predominantly based on Western European or American popular culture. Meanwhile, the lack of cultural adaptation of the model can lead to incorrect results, a decrease in the generation quality, and the spread of stereotypes and offensive content. In an effort to address this issue, we examine the concept of cultural code and recognize the critical importance of its understanding by modern image generation models, an issue that has not been sufficiently addressed in the research community to date. We propose the methodology for collecting and processing the data necessary to form a dataset based on the cultural code, in particular the Russian one. We explore how the collected data affects the quality of generations in the national domain and analyze the effectiveness of our approach using the Kandinsky 3.1 text-to-image model. Human evaluation results demonstrate an increase in the level of awareness of Russian culture in the model.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S137 - S150"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rethinking Graph Classification Problem in Presence of Isomorphism 存在同构的图分类问题的再思考
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602385
S. Ivanov, S. Sviridov, E. Burnaev

There is an increasing interest in developing new models for graph classification problem that serves as a common benchmark for evaluation and comparison of GNNs and graph kernels. To ensure a fair comparison of the models several commonly used datasets exist and current assessments and conclusions rely on the validity of these datasets. However, as we show in this paper majority of these datasets contain isomorphic copies of the data points, which can lead to misleading conclusions. For example, the relative ranking of the graph models can change substantially if we remove isomorphic graphs in the test set.

To mitigate this we present several results. We show that explicitly incorporating the knowledge of isomorphism in the datasets can significantly boost the performance of any graph model. Finally, we re-evaluate commonly used graph models on refined graph datasets and provide recommendations for designing new datasets and metrics for graph classification problem.

人们对开发图分类问题的新模型越来越感兴趣,这些模型可以作为评估和比较gnn和图核的通用基准。为了确保模型之间的公平比较,存在几种常用的数据集,目前的评估和结论依赖于这些数据集的有效性。然而,正如我们在本文中所展示的,这些数据集中的大多数包含数据点的同构副本,这可能导致误导性的结论。例如,如果我们在测试集中删除同构图,图模型的相对排名会发生很大的变化。为了减轻这种情况,我们提出了几个结果。我们表明,在数据集中显式地结合同构知识可以显着提高任何图模型的性能。最后,我们在改进的图数据集上重新评估了常用的图模型,并为图分类问题设计新的数据集和度量提供了建议。
{"title":"Rethinking Graph Classification Problem in Presence of Isomorphism","authors":"S. Ivanov,&nbsp;S. Sviridov,&nbsp;E. Burnaev","doi":"10.1134/S1064562424602385","DOIUrl":"10.1134/S1064562424602385","url":null,"abstract":"<p>There is an increasing interest in developing new models for graph classification problem that serves as a common benchmark for evaluation and comparison of GNNs and graph kernels. To ensure a fair comparison of the models several commonly used datasets exist and current assessments and conclusions rely on the validity of these datasets. However, as we show in this paper majority of these datasets contain isomorphic copies of the data points, which can lead to misleading conclusions. For example, the relative ranking of the graph models can change substantially if we remove isomorphic graphs in the test set.</p><p>To mitigate this we present several results. We show that explicitly incorporating the knowledge of isomorphism in the datasets can significantly boost the performance of any graph model. Finally, we re-evaluate commonly used graph models on refined graph datasets and provide recommendations for designing new datasets and metrics for graph classification problem.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S312 - S331"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602385.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RuSciBench: Open Benchmark for Russian and English Scientific Document Representations RuSciBench:俄语和英语科学文献表示的开放基准
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602191
A. Vatolin, N. Gerasimenko, A. Ianina, K. Vorontsov

Sharing scientific knowledge in the community is an important endeavor. However, most papers are written in English, which makes dissemination of knowledge in countries where English is not spoken by the majority of people harder. Nowadays, machine translation and language models may help to solve this problem, but it is still complicated to train and evaluate models in languages other than English with no or little data in the required language. To address this, we propose the first benchmark for evaluating models on scientific texts in Russian. It consists of papers from Russian electronic library of scientific publications. We also present a set of tasks which can be used to fine-tune various models on our data and provide a detailed comparison between state-of-the-art models on our benchmark.

在社区中分享科学知识是一项重要的努力。然而,大多数论文都是用英语写的,这使得在大多数人不讲英语的国家传播知识变得更加困难。如今,机器翻译和语言模型可能有助于解决这个问题,但在没有或很少有所需语言数据的情况下,用英语以外的语言训练和评估模型仍然很复杂。为了解决这个问题,我们提出了评估俄语科学文本模型的第一个基准。它由来自俄罗斯科学出版物电子图书馆的论文组成。我们还提出了一组任务,可用于对数据上的各种模型进行微调,并在基准上提供最先进模型之间的详细比较。
{"title":"RuSciBench: Open Benchmark for Russian and English Scientific Document Representations","authors":"A. Vatolin,&nbsp;N. Gerasimenko,&nbsp;A. Ianina,&nbsp;K. Vorontsov","doi":"10.1134/S1064562424602191","DOIUrl":"10.1134/S1064562424602191","url":null,"abstract":"<p>Sharing scientific knowledge in the community is an important endeavor. However, most papers are written in English, which makes dissemination of knowledge in countries where English is not spoken by the majority of people harder. Nowadays, machine translation and language models may help to solve this problem, but it is still complicated to train and evaluate models in languages other than English with no or little data in the required language. To address this, we propose the first benchmark for evaluating models on scientific texts in Russian. It consists of papers from Russian electronic library of scientific publications. We also present a set of tasks which can be used to fine-tune various models on our data and provide a detailed comparison between state-of-the-art models on our benchmark.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S251 - S260"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602191.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hiding Backdoors within Event Sequence Data via Poisoning Attacks 通过中毒攻击在事件序列数据中隐藏后门
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602221
A. Ermilova, E. Kovtun, D. Berestnev, A. Zaytsev

Deep learning’s emerging role in the financial sector’s decision-making introduces risks of adversarial attacks. A specific threat is a poisoning attack that modifies the training sample to develop a backdoor that persists during model usage. However, data cleaning procedures and routine model checks are easy-to-implement actions that prevent the usage of poisoning attacks. The problem is even more challenging for event sequence models, for which it is hard to design an attack due to the discrete nature of the data. We start with a general investigation of the possibility of poisoning for event sequence models. Then, we propose a concealed poisoning attack that can bypass natural banks’ defences. The empirical investigation shows that the developed poisoned model trained on contaminated data passes the check procedure, being similar to a clean model, and simultaneously contains a simple to-implement backdoor.

深度学习在金融部门决策中的新兴作用带来了对抗性攻击的风险。一个特定的威胁是中毒攻击,它修改训练样本以开发一个在模型使用期间持续存在的后门。但是,数据清理过程和常规模型检查是易于实现的操作,可以防止使用中毒攻击。对于事件序列模型来说,这个问题更具挑战性,由于数据的离散性,很难设计攻击。我们从事件序列模型中毒可能性的一般调查开始。然后,我们提出了一种隐蔽的投毒攻击,可以绕过天然银行的防御。实证研究表明,在污染数据上训练的开发的有毒模型通过了检查程序,与干净模型相似,同时包含一个易于实现的后门。
{"title":"Hiding Backdoors within Event Sequence Data via Poisoning Attacks","authors":"A. Ermilova,&nbsp;E. Kovtun,&nbsp;D. Berestnev,&nbsp;A. Zaytsev","doi":"10.1134/S1064562424602221","DOIUrl":"10.1134/S1064562424602221","url":null,"abstract":"<p>Deep learning’s emerging role in the financial sector’s decision-making introduces risks of adversarial attacks. A specific threat is a poisoning attack that modifies the training sample to develop a backdoor that persists during model usage. However, data cleaning procedures and routine model checks are easy-to-implement actions that prevent the usage of poisoning attacks. The problem is even more challenging for event sequence models, for which it is hard to design an attack due to the discrete nature of the data. We start with a general investigation of the possibility of poisoning for event sequence models. Then, we propose a concealed poisoning attack that can bypass natural banks’ defences. The empirical investigation shows that the developed poisoned model trained on contaminated data passes the check procedure, being similar to a clean model, and simultaneously contains a simple to-implement backdoor.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S288 - S298"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602221.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation SwiftDepth++:用于精确深度估计的高效轻量级模型
IF 0.5 4区 数学 Q3 MATHEMATICS Pub Date : 2025-03-22 DOI: 10.1134/S1064562424602038
Y. Dayoub, I. Makarov

Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.

深度估计是各个领域的关键任务,但收集标记深度数据的高成本导致人们对自监督单目深度估计方法的兴趣越来越大。在本文中,我们介绍了SwiftDepth++,这是一个轻量级的深度估计模型,在保持低计算预算的同时提供有竞争力的结果。SwiftDepth++的核心创新在于其新颖的深度解码器,在保留重要信息的同时,通过快速压缩特征来提高效率。此外,我们结合了一个师生知识蒸馏框架,指导学生模型改进其预测。我们在KITTI和NYU数据集上对swiftdeep++进行了评估,在没有微调的情况下,它在KITTI数据集上的绝对相对误差(Abs_rel)为10.2%,在NYU数据集上的绝对相对误差(Abs_rel)为22%,所有参数都有大约600万个参数。这些结果表明,SwiftDepth++不仅满足现代深度估计任务的要求,而且显著降低了计算复杂度,使其成为现实应用的实用选择。
{"title":"SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation","authors":"Y. Dayoub,&nbsp;I. Makarov","doi":"10.1134/S1064562424602038","DOIUrl":"10.1134/S1064562424602038","url":null,"abstract":"<p>Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S162 - S171"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602038.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Doklady Mathematics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1