首页 > 最新文献

The 2013 International Joint Conference on Neural Networks (IJCNN)最新文献

英文 中文
Analog system modeling based on a double modified complex valued neural network 基于双修正复值神经网络的模拟系统建模
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6707136
A. Luchetta, S. Manetti, M. C. Piccirilli
The aim of this work is to present a novel technique for the identification of lumped circuit models of general distributed apparatus and devices. It is based on the use of a double modified complex value neural network. The method is not oriented to a unique class of electromagnetic systems, but it gives a procedure for the complete validation of the approximated lumped model and the extraction of the electrical parameter values. The inputs of the system are the geometrical (and/or manufacturing) parameters of the considered structure, while the outputs are the lumped circuit parameters. The method follows the Frequency Response Analysis (FRA) approach for elaborating the data presented to the network.
本文的目的是提出一种新的方法来识别一般分布式设备的集总电路模型。它是基于使用双重修正复值神经网络。该方法并不是针对某一类特殊的电磁系统,而是给出了一个完整验证近似集总模型和提取电参数值的步骤。系统的输入是所考虑结构的几何(和/或制造)参数,而输出是集总电路参数。该方法遵循频率响应分析(FRA)方法来详细说明呈现给网络的数据。
{"title":"Analog system modeling based on a double modified complex valued neural network","authors":"A. Luchetta, S. Manetti, M. C. Piccirilli","doi":"10.1109/IJCNN.2013.6707136","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6707136","url":null,"abstract":"The aim of this work is to present a novel technique for the identification of lumped circuit models of general distributed apparatus and devices. It is based on the use of a double modified complex value neural network. The method is not oriented to a unique class of electromagnetic systems, but it gives a procedure for the complete validation of the approximated lumped model and the extraction of the electrical parameter values. The inputs of the system are the geometrical (and/or manufacturing) parameters of the considered structure, while the outputs are the lumped circuit parameters. The method follows the Frequency Response Analysis (FRA) approach for elaborating the data presented to the network.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134115546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A study of transformation-invariances of deep belief networks 深度信念网络的变换不变性研究
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706884
Zheng Shou, Yuhao Zhang, H. Cai
In order to learn transformation-invariant features, several effective deep architectures like hierarchical feature learning and variant Deep Belief Networks (DBN) have been proposed. Considering the complexity of those variants, people are interested in whether DBN itself has transformation-invariances. First of all, we use original DBN to test original data. Almost same error rates will be achieved, if we change weights in the bottom interlayer according to transformations occurred in testing data. It implies that weights in the bottom interlayer can store the knowledge to handle transformations such as rotation, shifting, and scaling. Along with the continuous learning ability and good storage of DBN, we present our Weight-Transformed Training Algorithm (WTTA) without augmenting other layers, units or filters to original DBN. Based upon original training method, WTTA is aiming at transforming weights and is still unsupervised. For MNIST handwritten digits recognizing experiments, we adopted 784-100-100-100 DBN to compare the differences of recognizing ability in weights-transformed ranges. Most error rates generated by WTTA were below 25% while most rates generated by original training algorithm exceeded 25%. Then we also did an experiment on part of MIT-CBCL face database, with varying illumination, and the best testing accuracy can be achieved is 87.5%. Besides, similar results can be achieved by datasets covering all kinds of transformations, but WTTA only needs original training data and transform weights after each training loop. Consequently, we can mine inherent transformation-invariances of DBN by WTTA, and DBN itself can recognize transformed data at satisfying error rates without inserting other components.
为了学习变换不变特征,人们提出了几种有效的深度结构,如层次特征学习和变体深度信念网络(DBN)。考虑到这些变量的复杂性,人们对DBN本身是否具有变换不变性很感兴趣。首先,我们使用原始DBN对原始数据进行测试。如果我们根据测试数据中发生的转换改变底层中间层的权重,将获得几乎相同的错误率。这意味着底层中间层中的权重可以存储处理旋转、移动和缩放等转换的知识。由于DBN具有持续学习能力和良好的存储能力,我们提出了一种权重转换训练算法(WTTA),该算法无需在原始DBN上增加其他层、单元或滤波器。WTTA在原有训练方法的基础上,以转换权重为目标,仍然是无监督的。在MNIST手写体数字识别实验中,我们采用784-100-100-100 DBN来比较权重变换范围内识别能力的差异。WTTA产生的错误率大多在25%以下,而原始训练算法产生的错误率大多超过25%。然后我们还在MIT-CBCL人脸数据库的一部分上进行了实验,在不同光照条件下,测试准确率达到了87.5%。此外,覆盖各种变换的数据集也可以得到类似的结果,但WTTA只需要原始的训练数据和每个训练循环后的变换权值。因此,我们可以利用WTTA挖掘DBN的固有变换不变性,并且DBN本身可以在不插入其他组件的情况下以满意的错误率识别转换后的数据。
{"title":"A study of transformation-invariances of deep belief networks","authors":"Zheng Shou, Yuhao Zhang, H. Cai","doi":"10.1109/IJCNN.2013.6706884","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706884","url":null,"abstract":"In order to learn transformation-invariant features, several effective deep architectures like hierarchical feature learning and variant Deep Belief Networks (DBN) have been proposed. Considering the complexity of those variants, people are interested in whether DBN itself has transformation-invariances. First of all, we use original DBN to test original data. Almost same error rates will be achieved, if we change weights in the bottom interlayer according to transformations occurred in testing data. It implies that weights in the bottom interlayer can store the knowledge to handle transformations such as rotation, shifting, and scaling. Along with the continuous learning ability and good storage of DBN, we present our Weight-Transformed Training Algorithm (WTTA) without augmenting other layers, units or filters to original DBN. Based upon original training method, WTTA is aiming at transforming weights and is still unsupervised. For MNIST handwritten digits recognizing experiments, we adopted 784-100-100-100 DBN to compare the differences of recognizing ability in weights-transformed ranges. Most error rates generated by WTTA were below 25% while most rates generated by original training algorithm exceeded 25%. Then we also did an experiment on part of MIT-CBCL face database, with varying illumination, and the best testing accuracy can be achieved is 87.5%. Besides, similar results can be achieved by datasets covering all kinds of transformations, but WTTA only needs original training data and transform weights after each training loop. Consequently, we can mine inherent transformation-invariances of DBN by WTTA, and DBN itself can recognize transformed data at satisfying error rates without inserting other components.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134099051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
The simultaneous coding of heading and path in primate MSTd 灵长类动物MSTd中头部和路径的同时编码
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706833
Oliver W. Layton, N. A. Browning
The spatio-temporal displacement of luminance patterns in a 2D image is called optic flow. Present biologically-inspired approaches to navigation that use optic flow largely focus on the problem of extracting the instantaneous direction of travel (heading) of a mobile agent. Computational models have demonstrated success in estimating heading in highly constrained environments whereby the agent is largely assumed to travel along straight paths. However, drivers competently steer around curved road bends and humans have been shown capable of judging their future, possibly curved, path of travel in addition to instantaneous heading. The computation of the general future path of travel, which need not be straight, is of interest to mobile robotic, autonomous vehicle driving, and path planning applications, yet no biologically-inspired neural network model exists that provides mechanisms through which the future path may be estimated. We present a biologically inspired recurrent neural network, based on brain area MSTd, that can dynamically code both instantaneous heading and path simultaneously. We show that the model performs similarly to humans in judging heading and the curvature of the future path.
二维图像中亮度模式的时空位移称为光流。目前利用光流进行导航的生物学启发方法主要集中在提取移动代理的瞬时行进方向(航向)的问题上。在高度受限的环境中,计算模型已经证明了在估计航向方面的成功,在这种环境中,智能体在很大程度上被假设沿着直线行驶。然而,驾驶员能够熟练地驾驭弯道,而且人类已经被证明能够判断他们未来的道路,可能是弯道,除了即时的方向。移动机器人、自动驾驶汽车和路径规划应用对一般未来路径的计算很感兴趣,但目前还没有一种受生物启发的神经网络模型,可以提供估计未来路径的机制。我们提出了一种基于脑区MSTd的生物学启发的递归神经网络,可以同时动态编码瞬时航向和路径。我们表明,该模型在判断方向和未来路径的曲率方面与人类相似。
{"title":"The simultaneous coding of heading and path in primate MSTd","authors":"Oliver W. Layton, N. A. Browning","doi":"10.1109/IJCNN.2013.6706833","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706833","url":null,"abstract":"The spatio-temporal displacement of luminance patterns in a 2D image is called optic flow. Present biologically-inspired approaches to navigation that use optic flow largely focus on the problem of extracting the instantaneous direction of travel (heading) of a mobile agent. Computational models have demonstrated success in estimating heading in highly constrained environments whereby the agent is largely assumed to travel along straight paths. However, drivers competently steer around curved road bends and humans have been shown capable of judging their future, possibly curved, path of travel in addition to instantaneous heading. The computation of the general future path of travel, which need not be straight, is of interest to mobile robotic, autonomous vehicle driving, and path planning applications, yet no biologically-inspired neural network model exists that provides mechanisms through which the future path may be estimated. We present a biologically inspired recurrent neural network, based on brain area MSTd, that can dynamically code both instantaneous heading and path simultaneously. We show that the model performs similarly to humans in judging heading and the curvature of the future path.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"33 9","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131806248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A topographical nonnegative matrix factorization algorithm 一种地形非负矩阵分解算法
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706849
Nicoleta Rogovschi, Lazhar Labiod, M. Nadif
We explore in this paper a novel topological organization algorithm for data clustering and visualization named TPNMF. It leads to a clustering of the data, as well as the projection of the clusters on a two-dimensional grid while preserving the topological order of the initial data. The proposed algorithm is based on a NMF (Nonnegative Matrix Factorization) formalism using a neighborhood function which take into account the topological order of the data. TPNMF was validated on variant real datasets and the experimental results show a good quality of the topological ordering and homogenous clustering.
本文探索了一种新的用于数据聚类和可视化的拓扑组织算法——TPNMF。它导致数据的聚类,以及在保持初始数据的拓扑顺序的同时在二维网格上的聚类投影。该算法基于NMF(非负矩阵分解)形式,使用考虑数据拓扑顺序的邻域函数。在不同的真实数据集上对TPNMF进行了验证,实验结果表明TPNMF具有良好的拓扑排序和同质聚类质量。
{"title":"A topographical nonnegative matrix factorization algorithm","authors":"Nicoleta Rogovschi, Lazhar Labiod, M. Nadif","doi":"10.1109/IJCNN.2013.6706849","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706849","url":null,"abstract":"We explore in this paper a novel topological organization algorithm for data clustering and visualization named TPNMF. It leads to a clustering of the data, as well as the projection of the clusters on a two-dimensional grid while preserving the topological order of the initial data. The proposed algorithm is based on a NMF (Nonnegative Matrix Factorization) formalism using a neighborhood function which take into account the topological order of the data. TPNMF was validated on variant real datasets and the experimental results show a good quality of the topological ordering and homogenous clustering.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131814417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Self-organizing maps with a single neuron 具有单个神经元的自组织地图
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706843
George M. Georgiou, K. Voigt
Self-organization is explored with a single complex-valued quadratic neuron. The output is the complex plane. A virtual grid is used to provide desired outputs for each input. Experiments have shown that training is fast. A quadratic neuron with the new training algorithm has been shown to have clustering properties. Data that are in a cluster in the input space tend to cluster on the complex plane. The speed of training and operation allows for efficient high-dimensional data exploration and for real-time critical applications.
研究了单个复值二次神经元的自组织问题。输出是复平面。虚拟网格用于为每个输入提供所需的输出。实验表明,训练是快速的。使用新训练算法的二次型神经元已被证明具有聚类特性。在输入空间中的聚类数据倾向于在复平面上聚类。训练和操作的速度允许高效的高维数据探索和实时关键应用。
{"title":"Self-organizing maps with a single neuron","authors":"George M. Georgiou, K. Voigt","doi":"10.1109/IJCNN.2013.6706843","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706843","url":null,"abstract":"Self-organization is explored with a single complex-valued quadratic neuron. The output is the complex plane. A virtual grid is used to provide desired outputs for each input. Experiments have shown that training is fast. A quadratic neuron with the new training algorithm has been shown to have clustering properties. Data that are in a cluster in the input space tend to cluster on the complex plane. The speed of training and operation allows for efficient high-dimensional data exploration and for real-time critical applications.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131006860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Behavioral economics and neuroeconomics: Cooperation, competition, preference, and decision making 行为经济学和神经经济学:合作、竞争、偏好和决策
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706709
S. Grossberg
Behavioral economics and neuroeconomics concern how humans process multiple alternatives to make their decisions, and propose how discoveries about how the brain works can inform models of economic behavior. This lecture will survey how results about cooperative-competitive and cognitive-emotional dynamics that were discovered to better understand how brains control behavior can shed light on issues of importance in economics, including results about the voting paradox, how to design stable economic markets, irrational decision making under risk (Prospect Theory), probabilistic decision making, preferences for previously unexperienced alternatives over rewarded experiences, and bounded rationality.
行为经济学和神经经济学关注人类如何处理多种选择来做出决定,并提出关于大脑如何工作的发现如何为经济行为模型提供信息。本讲座将探讨如何发现合作-竞争和认知-情绪动力学的结果,以更好地理解大脑如何控制行为,从而揭示经济学中的重要问题,包括投票悖论的结果,如何设计稳定的经济市场,风险下的非理性决策(前景理论),概率决策,对以前没有经验的替代方案的偏好,而不是奖励经验,以及有限理性。
{"title":"Behavioral economics and neuroeconomics: Cooperation, competition, preference, and decision making","authors":"S. Grossberg","doi":"10.1109/IJCNN.2013.6706709","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706709","url":null,"abstract":"Behavioral economics and neuroeconomics concern how humans process multiple alternatives to make their decisions, and propose how discoveries about how the brain works can inform models of economic behavior. This lecture will survey how results about cooperative-competitive and cognitive-emotional dynamics that were discovered to better understand how brains control behavior can shed light on issues of importance in economics, including results about the voting paradox, how to design stable economic markets, irrational decision making under risk (Prospect Theory), probabilistic decision making, preferences for previously unexperienced alternatives over rewarded experiences, and bounded rationality.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131153568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Toward a cooperative brain: Continuing the work with John Taylor 迈向合作大脑:继续约翰·泰勒的研究
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706715
B. Apolloni
I propose a three-step discussion following a research path shared in part with John Taylor where the leitmotif is to understand the cooperation between thinking agents: the pRAM architecture, the butler paradigm, and the networked intelligence. All three steps comprise keystones of European projects which one of us has coordinated. The principled philosophy is to “start simple and insert progressive complexity”. The results I discuss only go as far as the “start simple” point. The final goal is to find a bias that underpins the entire research effort. In this paper I will move within the connectionist paradigm at various scales, the largest being one that encompasses an Internet of Things instantiation.
我提出了一个三步讨论,其中的主题是理解思考主体之间的合作:pRAM架构、管家范式和网络化智能。所有这三个步骤都是我们其中一人协调的欧洲项目的基石。原则性的哲学是“从简单开始,然后逐步增加复杂性”。我所讨论的结果只涉及到“简单开始”点。最终目标是找到支撑整个研究工作的偏见。在本文中,我将在各种规模的连接主义范式中移动,最大的是包含物联网实例化的连接主义范式。
{"title":"Toward a cooperative brain: Continuing the work with John Taylor","authors":"B. Apolloni","doi":"10.1109/IJCNN.2013.6706715","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706715","url":null,"abstract":"I propose a three-step discussion following a research path shared in part with John Taylor where the leitmotif is to understand the cooperation between thinking agents: the pRAM architecture, the butler paradigm, and the networked intelligence. All three steps comprise keystones of European projects which one of us has coordinated. The principled philosophy is to “start simple and insert progressive complexity”. The results I discuss only go as far as the “start simple” point. The final goal is to find a bias that underpins the entire research effort. In this paper I will move within the connectionist paradigm at various scales, the largest being one that encompasses an Internet of Things instantiation.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133427764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Face recognition using voting technique for the Gabor and LDP features 人脸识别采用投票技术进行Gabor和LDP的特征分析
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6707094
I. Dagher, Jamal Hassanieh, Ahmad Younes
Face recognition can be described by a sophisticated mathematical representation and matching procedures. In this paper, Local Derivative Pattern (LDP) descriptors along with the Gabor feature extraction technique were used to achieve highest percentage of recognition possible. A robust comparison method, the Chi Square Distance, was used as a matching algorithm. Four databases involving different image capturing conditions: positioning, illumination and expressions were used. The best results were obtained after applying a voting technique to the Gabor and the LDP features.
人脸识别可以通过复杂的数学表示和匹配程序来描述。在本文中,使用局部导数模式(LDP)描述符和Gabor特征提取技术来实现最高的识别率。一种稳健的比较方法,卡方距离,被用作匹配算法。使用了定位、光照和表情四种不同图像捕获条件的数据库。将投票技术应用于Gabor和LDP的特征后,获得了最好的结果。
{"title":"Face recognition using voting technique for the Gabor and LDP features","authors":"I. Dagher, Jamal Hassanieh, Ahmad Younes","doi":"10.1109/IJCNN.2013.6707094","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6707094","url":null,"abstract":"Face recognition can be described by a sophisticated mathematical representation and matching procedures. In this paper, Local Derivative Pattern (LDP) descriptors along with the Gabor feature extraction technique were used to achieve highest percentage of recognition possible. A robust comparison method, the Chi Square Distance, was used as a matching algorithm. Four databases involving different image capturing conditions: positioning, illumination and expressions were used. The best results were obtained after applying a voting technique to the Gabor and the LDP features.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133107554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Speaker recognition based on SOINN and incremental learning Gaussian mixture model 基于SOINN和增量学习高斯混合模型的说话人识别
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6706863
Zelin Tang, S. Furao, Jinxi Zhao
Gaussian Mixture Models has been widely used in speaker recognition during the last decades. To deal with the dynamic growth of datasets, initial clustering problem and achieving the results of clustering effectively on incremental data, an incremental adaptation method called incremental learning Gaussian mixture model (IGMM) is proposed in this paper. It was applied to speaker recognition system based on Self Organization Incremental Learning Neural Network (SOINN) and improved EM algorithm. SOINN is a Neural Network which can reach a suitable mixture number and appropriate initial cluster for each model. First, the initial training is conducted by SOINN and EM algorithm only need a limited amount of data. Then, the model would adapt to the data available in each session to enrich itself incrementally and recursively. Experiments were taken on the 1st speech separation challenge database. The results show that IGMM outperforms GMM and classical Bayesian adaptation in most of the cases.
近几十年来,高斯混合模型在说话人识别中得到了广泛应用。为了解决数据集的动态增长、初始聚类问题以及在增量数据上实现有效聚类的结果,本文提出了一种增量适应方法——增量学习高斯混合模型(IGMM)。将其应用于基于自组织增量学习神经网络(SOINN)和改进的EM算法的说话人识别系统。SOINN是一种神经网络,可以为每个模型求得合适的混合数和合适的初始聚类。首先,初始训练由SOINN和EM算法进行,只需要有限的数据量。然后,模型将适应每个会话中可用的数据,以增量和递归方式丰富自己。在第一个语音分离挑战库上进行了实验。结果表明,IGMM在大多数情况下优于GMM和经典贝叶斯自适应。
{"title":"Speaker recognition based on SOINN and incremental learning Gaussian mixture model","authors":"Zelin Tang, S. Furao, Jinxi Zhao","doi":"10.1109/IJCNN.2013.6706863","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6706863","url":null,"abstract":"Gaussian Mixture Models has been widely used in speaker recognition during the last decades. To deal with the dynamic growth of datasets, initial clustering problem and achieving the results of clustering effectively on incremental data, an incremental adaptation method called incremental learning Gaussian mixture model (IGMM) is proposed in this paper. It was applied to speaker recognition system based on Self Organization Incremental Learning Neural Network (SOINN) and improved EM algorithm. SOINN is a Neural Network which can reach a suitable mixture number and appropriate initial cluster for each model. First, the initial training is conducted by SOINN and EM algorithm only need a limited amount of data. Then, the model would adapt to the data available in each session to enrich itself incrementally and recursively. Experiments were taken on the 1st speech separation challenge database. The results show that IGMM outperforms GMM and classical Bayesian adaptation in most of the cases.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115470994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Sparse similarity matrix learning for visual object retrieval 稀疏相似矩阵学习用于视觉对象检索
Pub Date : 2013-08-01 DOI: 10.1109/IJCNN.2013.6707063
Zhicheng Yan, Yizhou Yu
Tf-idf weighting scheme is adopted by state-of-the-art object retrieval systems to reflect the difference in discriminability between visual words. However, we argue it is only suboptimal by noting that tf-idf weighting scheme does not take quantization error into account and exploit word correlation. We view tf-idf weights as an example of diagonal Mahalanobis-type similarity matrix and generalize it into a sparse one by selectively activating off-diagonal elements. Our goal is to separate similarity of relevant images from that of irrelevant ones by a safe margin. We satisfy such similarity constraints by learning an optimal similarity metric from labeled data. An effective scheme is developed to collect training data with an emphasis on cases where the tf-idf weights violates the relative relevance constraints. Experimental results on benchmark datasets indicate the learnt similarity metric consistently and significantly outperforms the tf-idf weighting scheme.
最先进的目标检索系统采用Tf-idf加权方案来反映视觉词之间的区别。然而,我们认为它只是次优的,因为我们注意到tf-idf加权方案没有考虑量化误差和利用词相关性。我们将tf-idf权值作为对角马氏相似矩阵的一个例子,并通过选择性激活非对角元素将其推广为一个稀疏矩阵。我们的目标是在安全范围内将相关图像与不相关图像的相似性分开。我们通过从标记数据中学习最优相似性度量来满足这些相似性约束。开发了一种有效的方案来收集训练数据,重点是在tf-idf权重违反相对相关性约束的情况下。在基准数据集上的实验结果表明,学习到的相似度度量一致且显著优于tf-idf加权方案。
{"title":"Sparse similarity matrix learning for visual object retrieval","authors":"Zhicheng Yan, Yizhou Yu","doi":"10.1109/IJCNN.2013.6707063","DOIUrl":"https://doi.org/10.1109/IJCNN.2013.6707063","url":null,"abstract":"Tf-idf weighting scheme is adopted by state-of-the-art object retrieval systems to reflect the difference in discriminability between visual words. However, we argue it is only suboptimal by noting that tf-idf weighting scheme does not take quantization error into account and exploit word correlation. We view tf-idf weights as an example of diagonal Mahalanobis-type similarity matrix and generalize it into a sparse one by selectively activating off-diagonal elements. Our goal is to separate similarity of relevant images from that of irrelevant ones by a safe margin. We satisfy such similarity constraints by learning an optimal similarity metric from labeled data. An effective scheme is developed to collect training data with an emphasis on cases where the tf-idf weights violates the relative relevance constraints. Experimental results on benchmark datasets indicate the learnt similarity metric consistently and significantly outperforms the tf-idf weighting scheme.","PeriodicalId":376975,"journal":{"name":"The 2013 International Joint Conference on Neural Networks (IJCNN)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115553290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
The 2013 International Joint Conference on Neural Networks (IJCNN)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1