首页 > 最新文献

Advances in computational intelligence最新文献

英文 中文
Advances in Computational Intelligence: 21st Mexican International Conference on Artificial Intelligence, MICAI 2022, Monterrey, Mexico, October 24–29, 2022, Proceedings, Part I 计算智能的进展:第21届墨西哥国际人工智能会议,MICAI 2022,蒙特雷,墨西哥,2022年10月24日至29日,会议录,第一部分
Pub Date : 2022-01-01 DOI: 10.1007/978-3-031-19493-1
{"title":"Advances in Computational Intelligence: 21st Mexican International Conference on Artificial Intelligence, MICAI 2022, Monterrey, Mexico, October 24–29, 2022, Proceedings, Part I","authors":"","doi":"10.1007/978-3-031-19493-1","DOIUrl":"https://doi.org/10.1007/978-3-031-19493-1","url":null,"abstract":"","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"51 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87132071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Advances in Computational Intelligence: 21st Mexican International Conference on Artificial Intelligence, MICAI 2022, Monterrey, Mexico, October 24–29, 2022, Proceedings, Part II 计算智能的进展:第21届墨西哥国际人工智能会议,MICAI 2022,蒙特雷,墨西哥,2022年10月24日至29日,会议录,第二部分
Pub Date : 2022-01-01 DOI: 10.1007/978-3-031-19496-2
{"title":"Advances in Computational Intelligence: 21st Mexican International Conference on Artificial Intelligence, MICAI 2022, Monterrey, Mexico, October 24–29, 2022, Proceedings, Part II","authors":"","doi":"10.1007/978-3-031-19496-2","DOIUrl":"https://doi.org/10.1007/978-3-031-19496-2","url":null,"abstract":"","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"6 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89961185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A support vector approach based on penalty function method 一种基于罚函数法的支持向量方法
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00026-4
Songfeng Zheng

Support vector machine (SVM) models are usually trained by solving the dual of a quadratic programming, which is time consuming. Using the idea of penalty function method from optimization theory, this paper combines the objective function and the constraints in the dual, obtaining an unconstrained optimization problem, which could be solved by a generalized Newton method, yielding an approximate solution to the original model. Extensive experiments on pattern classification were conducted, and compared to the quadratic programming-based models, the proposed approach is much more computationally efficient (tens to hundreds of times faster) and yields similar performance in terms of receiver operating characteristic curve. Furthermore, the proposed method and quadratic programming-based models extract almost the same set of support vectors.

支持向量机(SVM)模型通常通过求解二次规划的对偶来训练,这是耗时的。利用优化理论中罚函数法的思想,将对偶中的目标函数和约束条件相结合,得到了一个无约束优化问题,该问题可以用广义牛顿法求解,得到了原始模型的近似解。对模式分类进行了广泛的实验,与基于二次规划的模型相比,所提出的方法在计算上更高效(速度快几十到几百倍),并且在接收机工作特性曲线方面产生了类似的性能。此外,所提出的方法和基于二次规划的模型提取了几乎相同的支持向量集。
{"title":"A support vector approach based on penalty function method","authors":"Songfeng Zheng","doi":"10.1007/s43674-021-00026-4","DOIUrl":"10.1007/s43674-021-00026-4","url":null,"abstract":"<div><p>Support vector machine (SVM) models are usually trained by solving the dual of a quadratic programming, which is time consuming. Using the idea of penalty function method from optimization theory, this paper combines the objective function and the constraints in the dual, obtaining an unconstrained optimization problem, which could be solved by a generalized Newton method, yielding an approximate solution to the original model. Extensive experiments on pattern classification were conducted, and compared to the quadratic programming-based models, the proposed approach is much more computationally efficient (tens to hundreds of times faster) and yields similar performance in terms of receiver operating characteristic curve. Furthermore, the proposed method and quadratic programming-based models extract almost the same set of support vectors.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00026-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Solutions of Yang Baxter equation of symplectic Jordan superalgebras 辛Jordan超代数的Yang-Baxter方程的解
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00017-5
Amir Baklouti, Warda Bensalah, Khaled Al-Motairi

We establish in this paper the equivalence between the existence of a solution of the Yang Baxter equation of a Jordan superalgebras and that of symplectic form on Jordan superalgebras.

本文建立了Jordan超代数的Yang-Baxter方程的一个解的存在性与Jordan超代数上辛形式的解的等价性。
{"title":"Solutions of Yang Baxter equation of symplectic Jordan superalgebras","authors":"Amir Baklouti,&nbsp;Warda Bensalah,&nbsp;Khaled Al-Motairi","doi":"10.1007/s43674-021-00017-5","DOIUrl":"10.1007/s43674-021-00017-5","url":null,"abstract":"<div><p>We establish in this paper the equivalence between the existence of a solution of the Yang Baxter equation of a Jordan superalgebras and that of symplectic form on Jordan superalgebras.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An eigenvector approach for obtaining scale and orientation invariant classification in convolutional neural networks 卷积神经网络中获得尺度和方向不变分类的特征向量方法
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00023-7
Swetha Velluva Chathoth, Asish Kumar Mishra, Deepak Mishra, Subrahmanyam Gorthi R. K. Sai

The convolution neural networks are well known for their efficiency in detecting and classifying objects once adequately trained. Though they address shift in-variance up to a limit, appreciable rotation and scale in-variances are not guaranteed by many of the existing CNN architectures, making them sensitive towards input image or feature map rotation and scale variations. Many attempts have been made in the past to acquire rotation and scale in-variances in CNNs. In this paper, an efficient approach is proposed for incorporating rotation and scale in-variances in CNN-based classifications, based on eigenvectors and eigenvalues of the image covariance matrix. Without demanding any training data augmentation or CNN architectural change, the proposed method, ‘Scale and Orientation Corrected Networks (SOCN)’, achieves better rotation and scale-invariant performances. SOCN proposes a scale and orientation correction step for images before baseline CNN training and testing. Being a generalized approach, SOCN can be combined with any baseline CNN to improve its rotational and scale in-variance performances. We demonstrate the proposed approach’s scale and orientation invariant classification ability with several real cases ranging from scale and orientation invariant character recognition to orientation invariant image classification, with different suitable baseline architectures. The proposed approach of SOCN, though is simple, outperforms the current state of the art scale and orientation invariant classifiers comparatively with minimal training and testing time.

卷积神经网络以其在充分训练后检测和分类对象的效率而闻名。尽管它们在一定程度上解决了方差的变化,但许多现有的CNN架构并不能保证方差的显著旋转和缩放,这使得它们对输入图像或特征图的旋转和缩放变化很敏感。过去已经进行了许多尝试来获得细胞神经网络的轮换和方差规模。在本文中,基于图像协方差矩阵的特征向量和特征值,提出了一种在基于CNN的分类中结合方差中的旋转和尺度的有效方法。在不需要任何训练数据扩充或CNN架构更改的情况下,所提出的“尺度和方向校正网络(SOCN)”方法实现了更好的旋转和尺度不变性能。SOCN提出了在基线CNN训练和测试之前对图像进行尺度和方向校正的步骤。作为一种通用方法,SOCN可以与任何基线CNN相结合,以提高其旋转和方差尺度性能。我们在从尺度和方向不变的字符识别到方向不变的图像分类的几个实际案例中,用不同的合适的基线架构,证明了所提出的方法的尺度和方向无关的分类能力。所提出的SOCN方法虽然简单,但与现有技术的尺度和方向不变分类器相比,在最小的训练和测试时间下,其性能要好。
{"title":"An eigenvector approach for obtaining scale and orientation invariant classification in convolutional neural networks","authors":"Swetha Velluva Chathoth,&nbsp;Asish Kumar Mishra,&nbsp;Deepak Mishra,&nbsp;Subrahmanyam Gorthi R. K. Sai","doi":"10.1007/s43674-021-00023-7","DOIUrl":"10.1007/s43674-021-00023-7","url":null,"abstract":"<div><p>The convolution neural networks are well known for their efficiency in detecting and classifying objects once adequately trained. Though they address shift in-variance up to a limit, appreciable rotation and scale in-variances are not guaranteed by many of the existing CNN architectures, making them sensitive towards input image or feature map rotation and scale variations. Many attempts have been made in the past to acquire rotation and scale in-variances in CNNs. In this paper, an efficient approach is proposed for incorporating rotation and scale in-variances in CNN-based classifications, based on eigenvectors and eigenvalues of the image covariance matrix. Without demanding any training data augmentation or CNN architectural change, the proposed method, <b>‘Scale and Orientation Corrected Networks (SOCN)’</b>, achieves better rotation and scale-invariant performances. <b>SOCN</b> proposes a scale and orientation correction step for images before baseline CNN training and testing. Being a generalized approach, <b>SOCN</b> can be combined with any baseline CNN to improve its rotational and scale in-variance performances. We demonstrate the proposed approach’s scale and orientation invariant classification ability with several real cases ranging from scale and orientation invariant character recognition to orientation invariant image classification, with different suitable baseline architectures. The proposed approach of <b>SOCN</b>, though is simple, outperforms the current state of the art scale and orientation invariant classifiers comparatively with minimal training and testing time.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00023-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
BCK codes BCK代码
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00018-4
Hashem Bordbar

In this paper, we initiate the study of the notion of the BCK-function on an arbitrary set A, and providing connections with x-functions and x-subsets for (x in X) where X is a BCK-algebra. Moreover, using the notion of order in a BCK-algebra, the BCK-code C is introduced and besides a new structure of order in C is investigated. Finally, we show that the structure of the BCK-algebra X and the BCK-code C which is generated by X, with their related orders are the same.

在本文中,我们开始研究任意集A上BCK函数的概念,并为(x In x)提供了x函数和x子集的连接,其中x是BCK代数。此外,利用BCK代数中阶的概念,引入了BCK码C,并研究了C中一种新的阶结构。最后,我们证明了BCK代数X和由X生成的BCK码C的结构及其相关阶是相同的。
{"title":"BCK codes","authors":"Hashem Bordbar","doi":"10.1007/s43674-021-00018-4","DOIUrl":"10.1007/s43674-021-00018-4","url":null,"abstract":"<div><p>In this paper, we initiate the study of the notion of the <i>BCK</i>-function on an arbitrary set <i>A</i>, and providing connections with <i>x</i>-functions and <i>x</i>-subsets for <span>(x in X)</span> where <i>X</i> is a <i>BCK</i>-algebra. Moreover, using the notion of order in a <i>BCK</i>-algebra, the <i>BCK</i>-code <i>C</i> is introduced and besides a new structure of order in <i>C</i> is investigated. Finally, we show that the structure of the <i>BCK</i>-algebra <i>X</i> and the <i>BCK</i>-code <i>C</i> which is generated by <i>X</i>, with their related orders are the same.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00018-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Caristi type mappings and characterization of completeness of Archimedean type fuzzy metric spaces Caristi型映射与阿基米德型模糊度量空间完备性的刻画
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00014-8
J. Martínez-Moreno, D. Gopal, Vladimir Rakočević, A. S. Ranadive, R. P. Pant

This paper deals with some issues of fixed point concerning Caristi type mappings introduced by Abbasi and Golshan (Kybernetika 52:929–942, 2016) in fuzzy metric spaces. We enlarge this class of mappings and prove completeness characterization of corresponding fuzzy metric space. The paper includes a comprehensive set of examples showing the generality of our results and an open question.

本文讨论了Abbasi和Golshan(Kybernetika 52:929–9421916)在模糊度量空间中引入的Caristi型映射的不动点的一些问题。我们扩大了这类映射,并证明了相应的模糊度量空间的完备性刻画。本文包括一组全面的例子,显示了我们的结果的普遍性和一个悬而未决的问题。
{"title":"Caristi type mappings and characterization of completeness of Archimedean type fuzzy metric spaces","authors":"J. Martínez-Moreno,&nbsp;D. Gopal,&nbsp;Vladimir Rakočević,&nbsp;A. S. Ranadive,&nbsp;R. P. Pant","doi":"10.1007/s43674-021-00014-8","DOIUrl":"10.1007/s43674-021-00014-8","url":null,"abstract":"<div><p>This paper deals with some issues of fixed point concerning Caristi type mappings introduced by Abbasi and Golshan (Kybernetika 52:929–942, 2016) in fuzzy metric spaces. We enlarge this class of mappings and prove completeness characterization of corresponding fuzzy metric space. The paper includes a comprehensive set of examples showing the generality of our results and an open question.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Feature selection based on min-redundancy and max-consistency 基于最小冗余和最大一致性的特征选择
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00021-9
Yanting Guo, Meng Hu, Eric C. C. Tsang, Degang Chen, Weihua Xu

Feature selection can effectively eliminate irrelevant or redundant features without changing features semantics, so as to improve the performance of learning and reduce the training time. In most of the existing feature selection methods based on rough sets, eliminating the redundant features between features and decisions, and deleting the redundant features between features are performed separately. This will greatly increase the search time of feature subset. To quickly remove redundant features, we define a series of feature evaluation functions that consider both the consistency between features and decisions, and redundancy between features, then propose a novel feature selection method based on min-redundancy and max-consistency. Firstly, we define the consistency of features with respect to decisions and the redundancy between features from neighborhood information granules. Then we propose a combined criterion to measure the importance of features and design a feature selection algorithm based on minimal-redundancy-maximal-consistency (mRMC). Finally, on UCI data sets, mRMC is compared with three other popular feature selection algorithms based on neighborhood idea, from classification accuracy, the number of selected features and running time. The experimental comparison shows that mRMC can quickly delete redundant features and select useful features while ensuring classification accuracy.

特征选择可以在不改变特征语义的情况下有效地消除不相关或冗余的特征,从而提高学习性能,减少训练时间。在现有的大多数基于粗糙集的特征选择方法中,消除特征与决策之间的冗余特征和删除特征之间的冗余特性是分开进行的。这将大大增加特征子集的搜索时间。为了快速去除冗余特征,我们定义了一系列既考虑特征与决策之间的一致性,又考虑特征之间的冗余性的特征评估函数,然后提出了一种基于最小冗余和最大一致性的新特征选择方法。首先,我们定义了特征与决策的一致性,以及来自邻域信息颗粒的特征之间的冗余。然后,我们提出了一个衡量特征重要性的组合标准,并设计了一个基于最小冗余最大一致性(mRMC)的特征选择算法。最后,在UCI数据集上,从分类精度、选择特征的数量和运行时间等方面,将mRMC与其他三种流行的基于邻域思想的特征选择算法进行了比较。实验比较表明,mRMC可以在保证分类精度的同时,快速删除冗余特征,选择有用特征。
{"title":"Feature selection based on min-redundancy and max-consistency","authors":"Yanting Guo,&nbsp;Meng Hu,&nbsp;Eric C. C. Tsang,&nbsp;Degang Chen,&nbsp;Weihua Xu","doi":"10.1007/s43674-021-00021-9","DOIUrl":"10.1007/s43674-021-00021-9","url":null,"abstract":"<div><p>Feature selection can effectively eliminate irrelevant or redundant features without changing features semantics, so as to improve the performance of learning and reduce the training time. In most of the existing feature selection methods based on rough sets, eliminating the redundant features between features and decisions, and deleting the redundant features between features are performed separately. This will greatly increase the search time of feature subset. To quickly remove redundant features, we define a series of feature evaluation functions that consider both the consistency between features and decisions, and redundancy between features, then propose a novel feature selection method based on min-redundancy and max-consistency. Firstly, we define the consistency of features with respect to decisions and the redundancy between features from neighborhood information granules. Then we propose a combined criterion to measure the importance of features and design a feature selection algorithm based on minimal-redundancy-maximal-consistency (mRMC). Finally, on UCI data sets, mRMC is compared with three other popular feature selection algorithms based on neighborhood idea, from classification accuracy, the number of selected features and running time. The experimental comparison shows that mRMC can quickly delete redundant features and select useful features while ensuring classification accuracy.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00021-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A hybrid monotone decision tree model for interval-valued attributes 区间值属性的混合单调决策树模型
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00016-6
Jiankai Chen, Zhongyan Li, Xin Wang, Junhai Zhai

The existing monotonic decision tree algorithms are based on a linearly ordered constraint that certain attributes are monotonously consistent with the decision, which could be called monotonic attributes, whereas others, called non-monotonic attributes. In practice, monotonic and non-monotonic attributes coexist in most classification tasks, and some attribute values are even evaluated as interval numbers. In this paper, we proposed a fuzzy rank-inconsistent rate based on probability degree to judge the monotonicity of interval numbers. Furthermore, we devised a hybrid model composed of monotonic and non-monotonic attributes to construct a mixed monotone decision tree for interval-valued data. Experiments on artificial and real-world data sets show that the proposed hybrid model is effective.

现有的单调决策树算法基于线性有序约束,即某些属性与决策单调一致,可以称为单调属性,而另一些则称为非单调属性。在实践中,单调和非单调属性在大多数分类任务中共存,一些属性值甚至被评估为区间数。本文提出了一种基于概率度的模糊秩不一致率来判断区间数的单调性。此外,我们设计了一个由单调和非单调属性组成的混合模型来构造区间值数据的混合单调决策树。在人工和真实世界数据集上的实验表明,所提出的混合模型是有效的。
{"title":"A hybrid monotone decision tree model for interval-valued attributes","authors":"Jiankai Chen,&nbsp;Zhongyan Li,&nbsp;Xin Wang,&nbsp;Junhai Zhai","doi":"10.1007/s43674-021-00016-6","DOIUrl":"10.1007/s43674-021-00016-6","url":null,"abstract":"<div><p>The existing monotonic decision tree algorithms are based on a linearly ordered constraint that certain attributes are monotonously consistent with the decision, which could be called monotonic attributes, whereas others, called non-monotonic attributes. In practice, monotonic and non-monotonic attributes coexist in most classification tasks, and some attribute values are even evaluated as interval numbers. In this paper, we proposed a fuzzy rank-inconsistent rate based on probability degree to judge the monotonicity of interval numbers. Furthermore, we devised a hybrid model composed of monotonic and non-monotonic attributes to construct a mixed monotone decision tree for interval-valued data. Experiments on artificial and real-world data sets show that the proposed hybrid model is effective.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00016-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Toward durable representations for continual learning 实现持续学习的持久表征
Pub Date : 2021-12-17 DOI: 10.1007/s43674-021-00022-8
Alaa El Khatib, Fakhri Karray

Continual learning models are known to suffer from catastrophic forgetting. Existing regularization methods to countering forgetting operate by penalizing large changes to learned parameters. A significant downside to these methods, however, is that, by effectively freezing model parameters, they gradually suspend the capacity of a model to learn new tasks. In this paper, we explore an alternative approach to the continual learning problem that aims to circumvent this downside. In particular, we ask the question: instead of forcing continual learning models to remember the past, can we modify the learning process from the start, such that the learned representations are less susceptible to forgetting? To this end, we explore multiple methods that could potentially encourage durable representations. We demonstrate empirically that the use of unsupervised auxiliary tasks achieves significant reduction in parameter re-optimization across tasks, and consequently reduces forgetting, without explicitly penalizing forgetting. Moreover, we propose a distance metric to track internal model dynamics across tasks, and use it to gain insight into the workings of our proposed approach, as well as other recently proposed methods.

众所周知,持续学习模式会遭受灾难性的遗忘。现有的对抗遗忘的正则化方法通过惩罚学习参数的大变化来操作。然而,这些方法的一个显著缺点是,通过有效地冻结模型参数,它们会逐渐暂停模型学习新任务的能力。在本文中,我们探索了一种解决持续学习问题的替代方法,旨在避免这种不利影响。特别是,我们提出了一个问题:与其强迫持续学习模型记住过去,我们能否从一开始就修改学习过程,使学习到的表征不太容易被遗忘?为此,我们探索了多种可能鼓励持久表示的方法。我们实证证明,使用无监督辅助任务可以显著减少任务间的参数重新优化,从而减少遗忘,而不会明显惩罚遗忘。此外,我们提出了一个距离度量来跟踪任务之间的内部模型动态,并使用它来深入了解我们提出的方法以及其他最近提出的方法的工作原理。
{"title":"Toward durable representations for continual learning","authors":"Alaa El Khatib,&nbsp;Fakhri Karray","doi":"10.1007/s43674-021-00022-8","DOIUrl":"10.1007/s43674-021-00022-8","url":null,"abstract":"<div><p>Continual learning models are known to suffer from <i>catastrophic forgetting</i>. Existing regularization methods to countering forgetting operate by penalizing large changes to learned parameters. A significant downside to these methods, however, is that, by effectively freezing model parameters, they gradually suspend the capacity of a model to learn new tasks. In this paper, we explore an alternative approach to the continual learning problem that aims to circumvent this downside. In particular, we ask the question: instead of forcing continual learning models to remember the past, can we modify the learning process from the start, such that the learned representations are less susceptible to forgetting? To this end, we explore multiple methods that could potentially encourage durable representations. We demonstrate empirically that the use of unsupervised auxiliary tasks achieves significant reduction in parameter re-optimization across tasks, and consequently reduces forgetting, without explicitly penalizing forgetting. Moreover, we propose a distance metric to track internal model dynamics across tasks, and use it to gain insight into the workings of our proposed approach, as well as other recently proposed methods.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00022-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"50488901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Advances in computational intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1