首页 > 最新文献

Pattern Analysis and Applications最新文献

英文 中文
Learning sample representativeness for class-imbalanced multi-label classification 学习类别不平衡多标签分类的样本代表性
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01209-8
Yu Zhang, Sichen Cao, Siya Mi, Yali Bian
{"title":"Learning sample representativeness for class-imbalanced multi-label classification","authors":"Yu Zhang, Sichen Cao, Siya Mi, Yali Bian","doi":"10.1007/s10044-024-01209-8","DOIUrl":"https://doi.org/10.1007/s10044-024-01209-8","url":null,"abstract":"","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140418872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-supervised fuzzy broad learning system based on mean-teacher model 基于平均教师模型的半监督模糊广泛学习系统
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01217-8
Zizhu Fan, Yijing Huang, Chao Xi, Cheng Peng, Shitong Wang

Fuzzy broad learning system (FBLS) is a newly proposed fuzzy system, which introduces Takagi–Sugeno fuzzy model into broad learning system. It has shown that FBLS has better nonlinear fitting ability and faster calculation speed than the most of fuzzy neural networks proposed earlier. At the same time, compared to other fuzzy neural networks, FBLS has fewer rules and lower cost of training time. However, label errors or missing are prone to appear in large-scale dataset, which will greatly reduce the performance of FBLS. Therefore, how to use limited label information to train a powerful classifier is an important challenge. In order to address this problem, we introduce Mean-Teacher model for the fuzzy broad learning system. We use the Mean-Teacher model to rebuild the weights of the output layer of FBLS, and use the Teacher–Student model to train FBLS. The proposed model is an implementation of semi-supervised learning which integrates fuzzy logic and broad learning system in the Mean-Teacher-based knowledge distillation framework. Finally, we have proved the great performance of Mean-Teacher-based fuzzy broad learning system (MT-FBLS) through a large number of experiments.

模糊广义学习系统(FBLS)是一种新提出的模糊系统,它在广义学习系统中引入了高木-菅野模糊模型。研究表明,与之前提出的大多数模糊神经网络相比,FBLS 具有更好的非线性拟合能力和更快的计算速度。同时,与其他模糊神经网络相比,FBLS 的规则更少,训练时间成本更低。但是,在大规模数据集中容易出现标签错误或缺失,这将大大降低 FBLS 的性能。因此,如何利用有限的标签信息训练出强大的分类器是一个重要的挑战。为了解决这个问题,我们为模糊广义学习系统引入了 Mean-Teacher 模型。我们使用 Mean-Teacher 模型重建 FBLS 输出层的权重,并使用 Teacher-Student 模型训练 FBLS。所提出的模型是半监督学习的一种实现,它在基于平均-教师的知识提炼框架中整合了模糊逻辑和广义学习系统。最后,我们通过大量实验证明了基于中值-教师的模糊广义学习系统(MT-FBLS)的卓越性能。
{"title":"Semi-supervised fuzzy broad learning system based on mean-teacher model","authors":"Zizhu Fan, Yijing Huang, Chao Xi, Cheng Peng, Shitong Wang","doi":"10.1007/s10044-024-01217-8","DOIUrl":"https://doi.org/10.1007/s10044-024-01217-8","url":null,"abstract":"<p>Fuzzy broad learning system (FBLS) is a newly proposed fuzzy system, which introduces Takagi–Sugeno fuzzy model into broad learning system. It has shown that FBLS has better nonlinear fitting ability and faster calculation speed than the most of fuzzy neural networks proposed earlier. At the same time, compared to other fuzzy neural networks, FBLS has fewer rules and lower cost of training time. However, label errors or missing are prone to appear in large-scale dataset, which will greatly reduce the performance of FBLS. Therefore, how to use limited label information to train a powerful classifier is an important challenge. In order to address this problem, we introduce Mean-Teacher model for the fuzzy broad learning system. We use the Mean-Teacher model to rebuild the weights of the output layer of FBLS, and use the Teacher–Student model to train FBLS. The proposed model is an implementation of semi-supervised learning which integrates fuzzy logic and broad learning system in the Mean-Teacher-based knowledge distillation framework. Finally, we have proved the great performance of Mean-Teacher-based fuzzy broad learning system (MT-FBLS) through a large number of experiments.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Selective bin model for reversible data hiding in encrypted images 用于加密图像中可逆数据隐藏的选择性 bin 模型
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01220-z
Ruchi Agarwal, Sara Ahmed, Manoj Kumar

In tandem with the fast-growing technology, the issue of secure data transmission over the Internet has achieved increasing importance. In digital media, enclosing data in images is one of the most common methods for communicating confidential information. A novel reversible data hiding in the encrypted images scheme based on selective bin models is proposed in this paper. The scheme focuses on enhancing the embedding capacity while ensuring the security of images with the help of encryption and the proposed data hiding process. For data embedding, lossless compression is utilized and the image is classified into three bins. Then, marker bits are assigned to these bins for distinguishing between embeddable and non-embeddable regions. The proposed method shows a satisfactory embedding rate for smooth images as well as complex ones due to its selective bin approach. Also, the method is separable in nature, i.e., data extraction and image recovery can be performed independently. Furthermore, the experimental results demonstrate the strategy’s effectiveness when compared with others.

随着技术的飞速发展,在互联网上安全传输数据的问题也变得越来越重要。在数字媒体中,将数据封装在图像中是传递机密信息最常用的方法之一。本文提出了一种基于选择性 bin 模型的新型加密图像可逆数据隐藏方案。该方案的重点是提高嵌入能力,同时借助加密和拟议的数据隐藏过程确保图像的安全性。为了进行数据嵌入,采用了无损压缩技术,并将图像分为三个分区。然后,为这些分区分配标记位,以区分可嵌入和不可嵌入区域。由于采用了选择性分仓方法,所提出的方法对于平滑图像和复杂图像都能显示出令人满意的嵌入率。此外,该方法还具有可分性,即数据提取和图像复原可以独立进行。此外,与其他方法相比,实验结果证明了该策略的有效性。
{"title":"Selective bin model for reversible data hiding in encrypted images","authors":"Ruchi Agarwal, Sara Ahmed, Manoj Kumar","doi":"10.1007/s10044-024-01220-z","DOIUrl":"https://doi.org/10.1007/s10044-024-01220-z","url":null,"abstract":"<p>In tandem with the fast-growing technology, the issue of secure data transmission over the Internet has achieved increasing importance. In digital media, enclosing data in images is one of the most common methods for communicating confidential information. A novel reversible data hiding in the encrypted images scheme based on selective bin models is proposed in this paper. The scheme focuses on enhancing the embedding capacity while ensuring the security of images with the help of encryption and the proposed data hiding process. For data embedding, lossless compression is utilized and the image is classified into three bins. Then, marker bits are assigned to these bins for distinguishing between embeddable and non-embeddable regions. The proposed method shows a satisfactory embedding rate for smooth images as well as complex ones due to its selective bin approach. Also, the method is separable in nature, i.e., data extraction and image recovery can be performed independently. Furthermore, the experimental results demonstrate the strategy’s effectiveness when compared with others.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140011526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nonparametric K-means clustering-based adaptive unsupervised colour image segmentation 基于非参数 K 均值聚类的自适应无监督彩色图像分割
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01228-5
Zubair Khan, Jie Yang
{"title":"Nonparametric K-means clustering-based adaptive unsupervised colour image segmentation","authors":"Zubair Khan, Jie Yang","doi":"10.1007/s10044-024-01228-5","DOIUrl":"https://doi.org/10.1007/s10044-024-01228-5","url":null,"abstract":"","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140419388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Subdomain adaptation via correlation alignment with entropy minimization for unsupervised domain adaptation 通过相关性对齐和熵最小化实现无监督域适应的子域适应
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01232-9
Obsa Gilo, Jimson Mathew, Samrat Mondal, Rakesh Kumar Sandoniya

Unsupervised domain adaptation (UDA) is a well-explored domain in transfer learning, finding applications across various real-world scenarios. The central challenge in UDA lies in addressing the domain shift between training (source) and testing (target) data distributions. This study focuses on image classification tasks within UDA, where label spaces are shared, but the target domain lacks labeled samples. Our primary objective revolves around mitigating the domain discrepancies between the source and target domains, ultimately facilitating robust generalization in the target domains. Domain adaptation techniques have traditionally concentrated on the global feature distribution to minimize disparities. However, these methods often need to pay more attention to crucial, domain-specific subdomain information within identical classification categories, challenging achieving the desired performance without fine-grained data. To tackle these challenges, we propose a unified framework, Subdomain Adaptation via Correlation Alignment with Entropy Minimization, for unsupervised domain adaptation. Our approach incorporates three advanced techniques: (1) Local Maximum Mean Discrepancy, which aligns the means of local feature subsets, capturing intrinsic subdomain alignments often missed by global alignment, (2) correlation alignment aimed at minimizing the correlation between domain distributions, and (3) entropy regularization applied to target domains to encourage low-density separation between categories. We validate our proposed methods through rigorous experimental evaluations and ablation studies on standard benchmark datasets. The results consistently demonstrate the superior performance of our approaches compared to existing state-of-the-art domain adaptation methods.

无监督领域适应(UDA)是迁移学习中一个被广泛探索的领域,在现实世界的各种场景中都有应用。UDA 的核心挑战在于解决训练(源)和测试(目标)数据分布之间的领域转换问题。本研究的重点是 UDA 中的图像分类任务,在这种任务中,标签空间是共享的,但目标域缺乏有标签的样本。我们的主要目标是减少源域和目标域之间的差异,最终促进目标域的稳健泛化。领域适应技术传统上集中在全局特征分布上,以尽量减少差异。然而,这些方法往往需要更多地关注相同分类类别中关键的、特定域的子域信息,这就对在没有细粒度数据的情况下实现理想性能提出了挑战。为了应对这些挑战,我们提出了一个统一的框架,即通过相关性对齐与熵最小化实现子域适应,用于无监督域适应。我们的方法融合了三种先进技术:(1) 局部最大均值差异(Local Maximum Mean Discrepancy),它对局部特征子集的均值进行对齐,捕捉全局对齐经常忽略的内在子域对齐;(2) 相关性对齐(Correlation Alignment),旨在最小化域分布之间的相关性;(3) 熵正则化(entropy regularization),应用于目标域,鼓励类别之间的低密度分离。我们在标准基准数据集上进行了严格的实验评估和消融研究,验证了我们提出的方法。结果一致表明,与现有的最先进的域适应方法相比,我们的方法具有卓越的性能。
{"title":"Subdomain adaptation via correlation alignment with entropy minimization for unsupervised domain adaptation","authors":"Obsa Gilo, Jimson Mathew, Samrat Mondal, Rakesh Kumar Sandoniya","doi":"10.1007/s10044-024-01232-9","DOIUrl":"https://doi.org/10.1007/s10044-024-01232-9","url":null,"abstract":"<p>Unsupervised domain adaptation (UDA) is a well-explored domain in transfer learning, finding applications across various real-world scenarios. The central challenge in UDA lies in addressing the domain shift between training (source) and testing (target) data distributions. This study focuses on image classification tasks within UDA, where label spaces are shared, but the target domain lacks labeled samples. Our primary objective revolves around mitigating the domain discrepancies between the source and target domains, ultimately facilitating robust generalization in the target domains. Domain adaptation techniques have traditionally concentrated on the global feature distribution to minimize disparities. However, these methods often need to pay more attention to crucial, domain-specific subdomain information within identical classification categories, challenging achieving the desired performance without fine-grained data. To tackle these challenges, we propose a unified framework, Subdomain Adaptation via Correlation Alignment with Entropy Minimization, for unsupervised domain adaptation. Our approach incorporates three advanced techniques: (1) Local Maximum Mean Discrepancy, which aligns the means of local feature subsets, capturing intrinsic subdomain alignments often missed by global alignment, (2) correlation alignment aimed at minimizing the correlation between domain distributions, and (3) entropy regularization applied to target domains to encourage low-density separation between categories. We validate our proposed methods through rigorous experimental evaluations and ablation studies on standard benchmark datasets. The results consistently demonstrate the superior performance of our approaches compared to existing state-of-the-art domain adaptation methods.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nonlinear dimensionality reduction with q-Gaussian distribution 利用 q 高斯分布进行非线性降维
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01210-1
Motoshi Abe, Yuichiro Nomura, Takio Kurita

In recent years, the dimensionality reduction has become more important as the number of dimensions of data used in various tasks such as regression and classification has increased. As popular nonlinear dimensionality reduction methods, t-distributed stochastic neighbor embedding (t-SNE) and uniform manifold approximation and projection (UMAP) have been proposed. However, the former outputs only one low-dimensional space determined by the t-distribution and the latter is difficult to control the distribution of distance between each pair of samples in low-dimensional space. To tackle these issues, we propose novel t-SNE and UMAP extended by q-Gaussian distribution, called q-Gaussian-distributed stochastic neighbor embedding (q-SNE) and q-Gaussian-distributed uniform manifold approximation and projection (q-UMAP). The q-Gaussian distribution is a probability distribution derived by maximizing the tsallis entropy by escort distribution with mean and variance, and a generalized version of Gaussian distribution with a hyperparameter q. Since the shape of the q-Gaussian distribution can be tuned smoothly by the hyperparameter q, q-SNE and q-UMAP can in- tuitively derive different embedding spaces. To show the quality of the proposed method, we compared the visualization of the low-dimensional embedding space and the classification accuracy by k-NN in the low-dimensional space. Empirical results on MNIST, COIL-20, OliverttiFaces and FashionMNIST demonstrate that the q-SNE and q-UMAP can derive better embedding spaces than t-SNE and UMAP.

近年来,随着用于回归和分类等各种任务的数据维数的增加,降维变得越来越重要。作为流行的非线性降维方法,t 分布随机邻域嵌入(t-SNE)和均匀流形逼近与投影(UMAP)已被提出。然而,前者只能输出一个由 t 分布决定的低维空间,而后者则难以控制低维空间中每对样本之间的距离分布。为了解决这些问题,我们提出了由 q-Gaussian 分布扩展的新型 t-SNE 和 UMAP,即 q-Gaussian 分布随机邻域嵌入(q-SNE)和 q-Gaussian 分布均匀流形逼近与投影(q-UMAP)。由于 q-Gaussian 分布的形状可以通过超参数 q 平滑调整,因此 q-SNE 和 q-UMAP 可以直观地推导出不同的嵌入空间。为了证明所提方法的质量,我们比较了低维嵌入空间的可视化和 k-NN 在低维空间中的分类精度。在 MNIST、COIL-20、OliverttiFaces 和 FashionMNIST 上的实证结果表明,q-SNE 和 q-UMAP 能比 t-SNE 和 UMAP 得出更好的嵌入空间。
{"title":"Nonlinear dimensionality reduction with q-Gaussian distribution","authors":"Motoshi Abe, Yuichiro Nomura, Takio Kurita","doi":"10.1007/s10044-024-01210-1","DOIUrl":"https://doi.org/10.1007/s10044-024-01210-1","url":null,"abstract":"<p>In recent years, the dimensionality reduction has become more important as the number of dimensions of data used in various tasks such as regression and classification has increased. As popular nonlinear dimensionality reduction methods, t-distributed stochastic neighbor embedding (t-SNE) and uniform manifold approximation and projection (UMAP) have been proposed. However, the former outputs only one low-dimensional space determined by the t-distribution and the latter is difficult to control the distribution of distance between each pair of samples in low-dimensional space. To tackle these issues, we propose novel t-SNE and UMAP extended by q-Gaussian distribution, called q-Gaussian-distributed stochastic neighbor embedding (q-SNE) and q-Gaussian-distributed uniform manifold approximation and projection (q-UMAP). The q-Gaussian distribution is a probability distribution derived by maximizing the tsallis entropy by escort distribution with mean and variance, and a generalized version of Gaussian distribution with a hyperparameter q. Since the shape of the q-Gaussian distribution can be tuned smoothly by the hyperparameter q, q-SNE and q-UMAP can in- tuitively derive different embedding spaces. To show the quality of the proposed method, we compared the visualization of the low-dimensional embedding space and the classification accuracy by k-NN in the low-dimensional space. Empirical results on MNIST, COIL-20, OliverttiFaces and FashionMNIST demonstrate that the q-SNE and q-UMAP can derive better embedding spaces than t-SNE and UMAP.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Information theory divergences in principal component analysis 主成分分析中的信息论分歧
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01215-w
Eduardo K. Nakao, Alexandre L. M. Levada

The metric learning area studies methodologies to find the most appropriate distance function for a given dataset. It was shown that dimensionality reduction algorithms are closely related to metric learning because, in addition to obtaining a more compact representation of the data, such methods also implicitly derive a distance function that best represents similarity between a pair of objects in the collection. Principal Component Analysis is a traditional linear dimensionality reduction algorithm that is still widely used by researchers. However, its procedure faithfully represents outliers in the generated space, which can be an undesirable characteristic in pattern recognition applications. With this is mind, it was proposed the replacement of the traditional punctual approach by a contextual one based on the data samples neighborhoods. This approach implements a mapping from the usual feature space to a parametric feature space, where the difference between two samples is defined by the vector whose scalar coordinates are given by the statistical divergence between two probability distributions. It was demonstrated for some divergences that the new approach outperforms several existing dimensionality reduction algorithms in a wide range of datasets. Although, it is important to investigate the framework divergence sensitivity. Experiments using Total Variation, Renyi, Sharma-Mittal and Tsallis divergences are exhibited in this paper and the results evidence the method robustness.

度量学习领域研究的是为给定数据集找到最合适的距离函数的方法。研究表明,降维算法与度量学习密切相关,因为除了获得更紧凑的数据表示外,这些方法还隐含地推导出最能体现集合中一对对象之间相似性的距离函数。主成分分析是一种传统的线性降维算法,目前仍被研究人员广泛使用。然而,其程序忠实地反映了生成空间中的异常值,这在模式识别应用中可能是一个不理想的特征。有鉴于此,有人提出用一种基于数据样本邻域的上下文方法来替代传统的定时方法。这种方法实现了从常规特征空间到参数特征空间的映射,其中两个样本之间的差异由向量定义,而向量的标量坐标由两个概率分布之间的统计发散给出。研究表明,对于某些发散,新方法在大量数据集中的表现优于现有的几种降维算法。不过,研究框架的发散敏感性也很重要。本文使用总变异、Renyi、Sharma-Mittal 和 Tsallis 发散进行了实验,结果证明了该方法的鲁棒性。
{"title":"Information theory divergences in principal component analysis","authors":"Eduardo K. Nakao, Alexandre L. M. Levada","doi":"10.1007/s10044-024-01215-w","DOIUrl":"https://doi.org/10.1007/s10044-024-01215-w","url":null,"abstract":"<p>The metric learning area studies methodologies to find the most appropriate distance function for a given dataset. It was shown that dimensionality reduction algorithms are closely related to metric learning because, in addition to obtaining a more compact representation of the data, such methods also implicitly derive a distance function that best represents similarity between a pair of objects in the collection. Principal Component Analysis is a traditional linear dimensionality reduction algorithm that is still widely used by researchers. However, its procedure faithfully represents outliers in the generated space, which can be an undesirable characteristic in pattern recognition applications. With this is mind, it was proposed the replacement of the traditional punctual approach by a contextual one based on the data samples neighborhoods. This approach implements a mapping from the usual feature space to a parametric feature space, where the difference between two samples is defined by the vector whose scalar coordinates are given by the statistical divergence between two probability distributions. It was demonstrated for some divergences that the new approach outperforms several existing dimensionality reduction algorithms in a wide range of datasets. Although, it is important to investigate the framework divergence sensitivity. Experiments using Total Variation, Renyi, Sharma-Mittal and Tsallis divergences are exhibited in this paper and the results evidence the method robustness.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A deep learning approach to censored regression 删减回归的深度学习方法
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01216-9
Vlad-Rareş Dănăilă, Cătălin Buiu

In censored regression, the outcomes are a mixture of known values (uncensored) and open intervals (censored), meaning that the outcome is either known with precision or is an unknown value above or below a known threshold. The use of censored data is widespread, and correctly modeling it is essential for many applications. Although the literature on censored regression is vast, deep learning approaches have been less frequently applied. This paper proposes three loss functions for training neural networks on censored data using gradient backpropagation: the tobit likelihood, the censored mean squared error, and the censored mean absolute error. We experimented with three variations in the tobit likelihood that arose from different ways of modeling the standard deviation variable: as a fixed value, a reparametrization, and an estimation using a separate neural network for heteroscedastic data. The tobit model yielded better results, but the other two losses are simpler to implement. Another central idea of our research was that data are often censored and truncated simultaneously. The proposed losses can handle simultaneous censoring and truncation at arbitrary values from above and below.

在有删减回归中,结果是已知值(无删减)和开放区间(有删减)的混合物,这意味着结果要么是精确的已知值,要么是高于或低于已知阈值的未知值。有删减数据的使用非常广泛,对其进行正确建模对许多应用都至关重要。虽然有关删减回归的文献浩如烟海,但深度学习方法的应用却并不频繁。本文提出了使用梯度反向传播对删减数据训练神经网络的三种损失函数:tobit 概率、删减均方误差和删减平均绝对误差。我们尝试了托比特似然的三种变化,这些变化源于对标准差变量的不同建模方法:固定值、重拟态以及使用单独的神经网络对异方差数据进行估计。tobit模型取得了更好的结果,但其他两种损失实现起来更简单。我们研究的另一个核心思想是,数据通常会同时被删减和截断。所提出的损耗可以同时处理剔除和截断,截断值可以是任意的上下值。
{"title":"A deep learning approach to censored regression","authors":"Vlad-Rareş Dănăilă, Cătălin Buiu","doi":"10.1007/s10044-024-01216-9","DOIUrl":"https://doi.org/10.1007/s10044-024-01216-9","url":null,"abstract":"<p>In censored regression, the outcomes are a mixture of known values (uncensored) and open intervals (censored), meaning that the outcome is either known with precision or is an unknown value above or below a known threshold. The use of censored data is widespread, and correctly modeling it is essential for many applications. Although the literature on censored regression is vast, deep learning approaches have been less frequently applied. This paper proposes three loss functions for training neural networks on censored data using gradient backpropagation: the tobit likelihood, the censored mean squared error, and the censored mean absolute error. We experimented with three variations in the tobit likelihood that arose from different ways of modeling the standard deviation variable: as a fixed value, a reparametrization, and an estimation using a separate neural network for heteroscedastic data. The tobit model yielded better results, but the other two losses are simpler to implement. Another central idea of our research was that data are often censored and truncated simultaneously. The proposed losses can handle simultaneous censoring and truncation at arbitrary values from above and below.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Local complex features learned by randomized neural networks for texture analysis 用随机神经网络学习局部复杂特征进行纹理分析
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01230-x

Abstract

Texture is a visual attribute largely used in many problems of image analysis. Many methods that use learning techniques have been proposed for texture discrimination, achieving improved performance over previous handcrafted methods. In this paper, we present a new approach that combines a learning technique and the complex network (CN) theory for texture analysis. This method takes advantage of the representation capacity of CN to model a texture image as a directed network and then uses the topological information of vertices to train a randomized neural network. This neural network has a single hidden layer and uses a fast learning algorithm to learn local CN patterns for texture characterization. Thus, we use the weights of the trained neural network to compose a feature vector. These feature vectors are evaluated in a classification experiment in four widely used image databases. Experimental results show a high classification performance of the proposed method compared to other methods, indicating that our approach can be used in many image analysis problems.

摘要 纹理是一种视觉属性,在许多图像分析问题中得到广泛应用。许多利用学习技术进行纹理判别的方法已被提出,与以前的手工方法相比,这些方法的性能有所提高。在本文中,我们提出了一种结合学习技术和复杂网络(CN)理论进行纹理分析的新方法。该方法利用复杂网络的表示能力,将纹理图像建模为有向网络,然后利用顶点的拓扑信息训练随机神经网络。该神经网络只有一个隐藏层,使用快速学习算法来学习局部 CN 模式,从而进行纹理表征。因此,我们使用训练好的神经网络的权重来组成特征向量。我们在四个广泛使用的图像数据库中对这些特征向量进行了分类评估。实验结果表明,与其他方法相比,所提出的方法具有很高的分类性能,这表明我们的方法可用于许多图像分析问题。
{"title":"Local complex features learned by randomized neural networks for texture analysis","authors":"","doi":"10.1007/s10044-024-01230-x","DOIUrl":"https://doi.org/10.1007/s10044-024-01230-x","url":null,"abstract":"<h3>Abstract</h3> <p>Texture is a visual attribute largely used in many problems of image analysis. Many methods that use learning techniques have been proposed for texture discrimination, achieving improved performance over previous handcrafted methods. In this paper, we present a new approach that combines a learning technique and the complex network (CN) theory for texture analysis. This method takes advantage of the representation capacity of CN to model a texture image as a directed network and then uses the topological information of vertices to train a randomized neural network. This neural network has a single hidden layer and uses a fast learning algorithm to learn local CN patterns for texture characterization. Thus, we use the weights of the trained neural network to compose a feature vector. These feature vectors are evaluated in a classification experiment in four widely used image databases. Experimental results show a high classification performance of the proposed method compared to other methods, indicating that our approach can be used in many image analysis problems.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Big topic modeling based on a two-level hierarchical latent Beta-Liouville allocation for large-scale data and parameter streaming 基于两级潜在 Beta-Liouville 分配的大主题建模,适用于大规模数据和参数流
IF 3.9 4区 计算机科学 Q2 Computer Science Pub Date : 2024-02-28 DOI: 10.1007/s10044-024-01213-y
Koffi Eddy Ihou, Nizar Bouguila

As an extension to the standard symmetric latent Dirichlet allocation topic model, we implement asymmetric Beta-Liouville as a conjugate prior to the multinomial and therefore propose the maximum a posteriori for latent Beta-Liouville allocation as an alternative to maximum likelihood estimator for models such as probabilistic latent semantic indexing, unigrams, and mixture of unigrams. Since most Bayesian posteriors, for complex models, are intractable in general, we propose a point estimate (the mode) that offers a much tractable solution. The maximum a posteriori hypotheses using point estimates are much easier than full Bayesian analysis that integrates over the entire parameter space. We show that the proposed maximum a posteriori reduces the three-level hierarchical latent Beta-Liouville allocation to two-level topic mixture as we marginalize out the latent variables. In each document, the maximum a posteriori provides a soft assignment and constructs dense expectation–maximization probabilities over each word (responsibilities) for accurate estimates. For simplicity, we present a stochastic at word-level online expectation–maximization algorithm as an optimization method for maximum a posteriori latent Beta-Liouville allocation estimation whose unnormalized reparameterization is equivalent to a stochastic collapsed variational Bayes. This implicit connection between the collapsed space and expectation–maximization-based maximum a posteriori latent Beta-Liouville allocation shows its flexibility and helps in providing alternative to model selection. We characterize efficiency in the proposed approach for its ability to simultaneously stream both large-scale data and parameters seamlessly. The performance of the model using predictive perplexities as evaluation method shows the robustness of the proposed technique with text document datasets.

作为标准对称潜狄利克特分配主题模型的扩展,我们将非对称贝塔-利乌维尔作为多项式的共轭先验来实现,并因此提出了潜贝塔-利乌维尔分配的最大后验,作为概率潜语义索引、单词和单词混合等模型的最大似然估计的替代方法。由于对于复杂模型来说,大多数贝叶斯后验一般都很难处理,因此我们提出了一种点估计(模式),提供了一种更易处理的解决方案。使用点估计的最大后验假设比整合整个参数空间的完全贝叶斯分析要容易得多。我们的研究表明,当我们将潜在变量边际化时,所提出的最大后验假设将三级分层潜在 Beta-Liouville 分配减少为两级主题混合。在每篇文档中,最大后验法都会提供软分配,并在每个词(责任)上构建密集的期望最大化概率,以获得准确的估计值。为简单起见,我们提出了一种词级随机在线期望最大化算法,作为最大后验潜变量 Beta-Liouville 分配估计的优化方法,其非规范化重参数化等同于随机坍塌变分贝叶斯。坍缩空间与基于期望最大化的最大后验潜在 Beta-Liouville 分配之间的这种隐含联系显示了它的灵活性,并有助于提供模型选择的替代方法。我们对所提出的方法的效率进行了描述,因为它能够同时无缝流式处理大规模数据和参数。使用预测复杂度作为评估方法的模型性能表明了所提技术在文本文档数据集上的鲁棒性。
{"title":"Big topic modeling based on a two-level hierarchical latent Beta-Liouville allocation for large-scale data and parameter streaming","authors":"Koffi Eddy Ihou, Nizar Bouguila","doi":"10.1007/s10044-024-01213-y","DOIUrl":"https://doi.org/10.1007/s10044-024-01213-y","url":null,"abstract":"<p>As an extension to the standard symmetric latent Dirichlet allocation topic model, we implement asymmetric Beta-Liouville as a conjugate prior to the multinomial and therefore propose the maximum a posteriori for latent Beta-Liouville allocation as an alternative to maximum likelihood estimator for models such as probabilistic latent semantic indexing, unigrams, and mixture of unigrams. Since most Bayesian posteriors, for complex models, are intractable in general, we propose a point estimate (the mode) that offers a much tractable solution. The maximum a posteriori hypotheses using point estimates are much easier than full Bayesian analysis that integrates over the entire parameter space. We show that the proposed maximum a posteriori reduces the three-level hierarchical latent Beta-Liouville allocation to two-level topic mixture as we marginalize out the latent variables. In each document, the maximum a posteriori provides a soft assignment and constructs dense expectation–maximization probabilities over each word (responsibilities) for accurate estimates. For simplicity, we present a stochastic at word-level online expectation–maximization algorithm as an optimization method for maximum a posteriori latent Beta-Liouville allocation estimation whose unnormalized reparameterization is equivalent to a stochastic collapsed variational Bayes. This implicit connection between the collapsed space and expectation–maximization-based maximum a posteriori latent Beta-Liouville allocation shows its flexibility and helps in providing alternative to model selection. We characterize efficiency in the proposed approach for its ability to simultaneously stream both large-scale data and parameters seamlessly. The performance of the model using predictive perplexities as evaluation method shows the robustness of the proposed technique with text document datasets.</p>","PeriodicalId":54639,"journal":{"name":"Pattern Analysis and Applications","volume":null,"pages":null},"PeriodicalIF":3.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140009371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Pattern Analysis and Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1