首页 > 最新文献

IEEE transactions on neural networks最新文献

英文 中文
A nonlinear control method based on ANFIS and multiple models for a class of SISO nonlinear systems and its application. 一类SISO非线性系统基于ANFIS和多模型的非线性控制方法及其应用。
Pub Date : 2011-11-01 Epub Date: 2011-09-29 DOI: 10.1109/TNN.2011.2166561
Yajun Zhang, Tianyou Chai, Hong Wang

This paper presents a novel nonlinear control strategy for a class of uncertain single-input and single-output discrete-time nonlinear systems with unstable zero-dynamics. The proposed method combines adaptive-network-based fuzzy inference system (ANFIS) with multiple models, where a linear robust controller, an ANFIS-based nonlinear controller and a switching mechanism are integrated using multiple models technique. It has been shown that the linear controller can ensure the boundedness of the input and output signals and the nonlinear controller can improve the dynamic performance of the closed loop system. Moreover, it has also been shown that the use of the switching mechanism can simultaneously guarantee the closed loop stability and improve its performance. As a result, the controller has the following three outstanding features compared with existing control strategies. First, this method relaxes the assumption of commonly-used uniform boundedness on the unmodeled dynamics and thus enhances its applicability. Second, since ANFIS is used to estimate and compensate the effect caused by the unmodeled dynamics, the convergence rate of neural network learning has been increased. Third, a "one-to-one mapping" technique is adapted to guarantee the universal approximation property of ANFIS. The proposed controller is applied to a numerical example and a pulverizing process of an alumina sintering system, respectively, where its effectiveness has been justified.

针对一类不确定单输入单输出不稳定零动力学离散非线性系统,提出了一种新的非线性控制策略。该方法将基于自适应网络的模糊推理系统(ANFIS)与多模型相结合,采用多模型技术将线性鲁棒控制器、基于ANFIS的非线性控制器和切换机制集成在一起。研究表明,线性控制器可以保证输入输出信号的有界性,非线性控制器可以改善闭环系统的动态性能。此外,还表明,使用开关机构可以同时保证闭环的稳定性和提高其性能。因此,与现有的控制策略相比,该控制器具有以下三个突出特点:首先,该方法放宽了对未建模动力学的一致有界性假设,提高了其适用性。其次,由于使用ANFIS来估计和补偿未建模的动态所造成的影响,提高了神经网络学习的收敛速度。第三,采用“一对一映射”技术保证了ANFIS的普遍逼近性。将所提出的控制器分别应用于一个数值算例和一个氧化铝烧结系统的制粉过程,验证了其有效性。
{"title":"A nonlinear control method based on ANFIS and multiple models for a class of SISO nonlinear systems and its application.","authors":"Yajun Zhang,&nbsp;Tianyou Chai,&nbsp;Hong Wang","doi":"10.1109/TNN.2011.2166561","DOIUrl":"https://doi.org/10.1109/TNN.2011.2166561","url":null,"abstract":"<p><p>This paper presents a novel nonlinear control strategy for a class of uncertain single-input and single-output discrete-time nonlinear systems with unstable zero-dynamics. The proposed method combines adaptive-network-based fuzzy inference system (ANFIS) with multiple models, where a linear robust controller, an ANFIS-based nonlinear controller and a switching mechanism are integrated using multiple models technique. It has been shown that the linear controller can ensure the boundedness of the input and output signals and the nonlinear controller can improve the dynamic performance of the closed loop system. Moreover, it has also been shown that the use of the switching mechanism can simultaneously guarantee the closed loop stability and improve its performance. As a result, the controller has the following three outstanding features compared with existing control strategies. First, this method relaxes the assumption of commonly-used uniform boundedness on the unmodeled dynamics and thus enhances its applicability. Second, since ANFIS is used to estimate and compensate the effect caused by the unmodeled dynamics, the convergence rate of neural network learning has been increased. Third, a \"one-to-one mapping\" technique is adapted to guarantee the universal approximation property of ANFIS. The proposed controller is applied to a numerical example and a pulverizing process of an alumina sintering system, respectively, where its effectiveness has been justified.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1783-95"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2166561","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30181108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Learning speaker-specific characteristics with a deep neural architecture. 使用深度神经结构学习讲话者的特定特征。
Pub Date : 2011-11-01 Epub Date: 2011-09-26 DOI: 10.1109/TNN.2011.2167240
Ke Chen, Ahmad Salman

Speech signals convey various yet mixed information ranging from linguistic to speaker-specific information. However, most of acoustic representations characterize all different kinds of information as whole, which could hinder either a speech or a speaker recognition (SR) system from producing a better performance. In this paper, we propose a novel deep neural architecture (DNA) especially for learning speaker-specific characteristics from mel-frequency cepstral coefficients, an acoustic representation commonly used in both speech recognition and SR, which results in a speaker-specific overcomplete representation. In order to learn intrinsic speaker-specific characteristics, we come up with an objective function consisting of contrastive losses in terms of speaker similarity/dissimilarity and data reconstruction losses used as regularization to normalize the interference of non-speaker-related information. Moreover, we employ a hybrid learning strategy for learning parameters of the deep neural networks: i.e., local yet greedy layerwise unsupervised pretraining for initialization and global supervised learning for the ultimate discriminative goal. With four Linguistic Data Consortium (LDC) benchmarks and two non-English corpora, we demonstrate that our overcomplete representation is robust in characterizing various speakers, no matter whether their utterances have been used in training our DNA, and highly insensitive to text and languages spoken. Extensive comparative studies suggest that our approach yields favorite results in speaker verification and segmentation. Finally, we discuss several issues concerning our proposed approach.

语音信号传递各种各样的混合信息,从语言信息到说话人特定的信息。然而,大多数声学表示将所有不同类型的信息作为一个整体来描述,这可能会阻碍语音或说话人识别(SR)系统产生更好的性能。在本文中,我们提出了一种新的深度神经结构(DNA),特别是用于从梅尔频率倒谱系数中学习说话人特定特征,这是语音识别和SR中常用的声学表示,导致说话人特定的过完全表示。为了学习说话人固有的特征,我们提出了一个由说话人相似/不相似度的对比损失和数据重建损失组成的目标函数,用于正则化非说话人相关信息的干扰。此外,我们采用了一种混合学习策略来学习深度神经网络的参数:即初始化的局部贪婪分层无监督预训练和最终判别目标的全局监督学习。通过四个语言数据联盟(LDC)基准和两个非英语语料库,我们证明了我们的过完备表示在描述不同说话者的特征方面是稳健的,无论他们的话语是否被用于训练我们的DNA,并且对所讲的文本和语言高度不敏感。大量的比较研究表明,我们的方法在说话人验证和分割方面取得了令人满意的结果。最后,我们讨论了与我们提出的方法有关的几个问题。
{"title":"Learning speaker-specific characteristics with a deep neural architecture.","authors":"Ke Chen,&nbsp;Ahmad Salman","doi":"10.1109/TNN.2011.2167240","DOIUrl":"https://doi.org/10.1109/TNN.2011.2167240","url":null,"abstract":"<p><p>Speech signals convey various yet mixed information ranging from linguistic to speaker-specific information. However, most of acoustic representations characterize all different kinds of information as whole, which could hinder either a speech or a speaker recognition (SR) system from producing a better performance. In this paper, we propose a novel deep neural architecture (DNA) especially for learning speaker-specific characteristics from mel-frequency cepstral coefficients, an acoustic representation commonly used in both speech recognition and SR, which results in a speaker-specific overcomplete representation. In order to learn intrinsic speaker-specific characteristics, we come up with an objective function consisting of contrastive losses in terms of speaker similarity/dissimilarity and data reconstruction losses used as regularization to normalize the interference of non-speaker-related information. Moreover, we employ a hybrid learning strategy for learning parameters of the deep neural networks: i.e., local yet greedy layerwise unsupervised pretraining for initialization and global supervised learning for the ultimate discriminative goal. With four Linguistic Data Consortium (LDC) benchmarks and two non-English corpora, we demonstrate that our overcomplete representation is robust in characterizing various speakers, no matter whether their utterances have been used in training our DNA, and highly insensitive to text and languages spoken. Extensive comparative studies suggest that our approach yields favorite results in speaker verification and segmentation. Finally, we discuss several issues concerning our proposed approach.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1744-56"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2167240","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30026908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 114
Delay-independent stability of genetic regulatory networks. 基因调控网络的延迟无关稳定性。
Pub Date : 2011-11-01 Epub Date: 2011-09-06 DOI: 10.1109/TNN.2011.2165556
Fang-Xiang Wu

Genetic regulatory networks can be described by nonlinear differential equations with time delays. In this paper, we study both locally and globally delay-independent stability of genetic regulatory networks, taking messenger ribonucleic acid alternative splicing into consideration. Based on nonnegative matrix theory, we first develop necessary and sufficient conditions for locally delay-independent stability of genetic regulatory networks with multiple time delays. Compared to the previous results, these conditions are easy to verify. Then we develop sufficient conditions for global delay-independent stability for genetic regulatory networks. Compared to the previous results, this sufficient condition is less conservative. To illustrate theorems developed in this paper, we analyze delay-independent stability of two genetic regulatory networks: a real-life repressilatory network with three genes and three proteins, and a synthetic gene regulatory network with five genes and seven proteins. The simulation results show that the theorems developed in this paper can effectively determine the delay-independent stability of genetic regulatory networks.

遗传调控网络可以用非线性时滞微分方程来描述。在本文中,我们研究了遗传调控网络的局部和全局延迟无关稳定性,并考虑了信使核糖核酸选择性剪接。基于非负矩阵理论,首先给出了具有多时滞的遗传调控网络局部时滞无关稳定性的充分必要条件。与之前的结果相比,这些条件很容易验证。然后给出了遗传调控网络全局时滞无关稳定性的充分条件。与以往的结果相比,该充分条件具有较小的保守性。为了说明本文提出的定理,我们分析了两种基因调控网络的延迟无关稳定性:一种是由三个基因和三个蛋白质组成的现实生活中的抑制网络,另一种是由五个基因和七个蛋白质组成的合成基因调控网络。仿真结果表明,本文提出的定理可以有效地确定遗传调控网络的延迟无关稳定性。
{"title":"Delay-independent stability of genetic regulatory networks.","authors":"Fang-Xiang Wu","doi":"10.1109/TNN.2011.2165556","DOIUrl":"https://doi.org/10.1109/TNN.2011.2165556","url":null,"abstract":"<p><p>Genetic regulatory networks can be described by nonlinear differential equations with time delays. In this paper, we study both locally and globally delay-independent stability of genetic regulatory networks, taking messenger ribonucleic acid alternative splicing into consideration. Based on nonnegative matrix theory, we first develop necessary and sufficient conditions for locally delay-independent stability of genetic regulatory networks with multiple time delays. Compared to the previous results, these conditions are easy to verify. Then we develop sufficient conditions for global delay-independent stability for genetic regulatory networks. Compared to the previous results, this sufficient condition is less conservative. To illustrate theorems developed in this paper, we analyze delay-independent stability of two genetic regulatory networks: a real-life repressilatory network with three genes and three proteins, and a synthetic gene regulatory network with five genes and seven proteins. The simulation results show that the theorems developed in this paper can effectively determine the delay-independent stability of genetic regulatory networks.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1685-93"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2165556","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30125870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
Decentralized optimal control of a class of interconnected nonlinear discrete-time systems by using online Hamilton-Jacobi-Bellman formulation. 用在线Hamilton-Jacobi-Bellman公式求解一类互连非线性离散系统的分散最优控制。
Pub Date : 2011-11-01 Epub Date: 2011-09-29 DOI: 10.1109/TNN.2011.2160968
Shahab Mehraeen, Sarangapani Jagannathan

In this paper, the direct neural dynamic programming technique is utilized to solve the Hamilton-Jacobi-Bellman equation forward-in-time for the decentralized near optimal regulation of a class of nonlinear interconnected discrete-time systems with unknown internal subsystem and interconnection dynamics, while the input gain matrix is considered known. Even though the unknown interconnection terms are considered weak and functions of the entire state vector, the decentralized control is attempted under the assumption that only the local state vector is measurable. The decentralized nearly optimal controller design for each subsystem consists of two neural networks (NNs), an action NN that is aimed to provide a nearly optimal control signal, and a critic NN which evaluates the performance of the overall system. All NN parameters are tuned online for both the NNs. By using Lyapunov techniques it is shown that all subsystems signals are uniformly ultimately bounded and that the synthesized subsystems inputs approach their corresponding nearly optimal control inputs with bounded error. Simulation results are included to show the effectiveness of the approach.

本文利用直接神经动态规划技术,在输入增益矩阵已知的情况下,求解了一类内部子系统和互联动态未知的非线性互联离散系统的分散近最优调节的Hamilton-Jacobi-Bellman方程。尽管未知互连项被认为是弱的,并且是整个状态向量的函数,但在假设只有局部状态向量是可测量的情况下,尝试分散控制。每个子系统的分散近最优控制器设计由两个神经网络(NN)组成,一个是旨在提供近最优控制信号的动作神经网络,另一个是评估整个系统性能的批评神经网络。两个NN的所有参数都是在线调优的。利用李雅普诺夫技术证明了所有子系统信号最终是一致有界的,并且合成子系统的输入接近其相应的具有有界误差的近最优控制输入。仿真结果表明了该方法的有效性。
{"title":"Decentralized optimal control of a class of interconnected nonlinear discrete-time systems by using online Hamilton-Jacobi-Bellman formulation.","authors":"Shahab Mehraeen,&nbsp;Sarangapani Jagannathan","doi":"10.1109/TNN.2011.2160968","DOIUrl":"https://doi.org/10.1109/TNN.2011.2160968","url":null,"abstract":"<p><p>In this paper, the direct neural dynamic programming technique is utilized to solve the Hamilton-Jacobi-Bellman equation forward-in-time for the decentralized near optimal regulation of a class of nonlinear interconnected discrete-time systems with unknown internal subsystem and interconnection dynamics, while the input gain matrix is considered known. Even though the unknown interconnection terms are considered weak and functions of the entire state vector, the decentralized control is attempted under the assumption that only the local state vector is measurable. The decentralized nearly optimal controller design for each subsystem consists of two neural networks (NNs), an action NN that is aimed to provide a nearly optimal control signal, and a critic NN which evaluates the performance of the overall system. All NN parameters are tuned online for both the NNs. By using Lyapunov techniques it is shown that all subsystems signals are uniformly ultimately bounded and that the synthesized subsystems inputs approach their corresponding nearly optimal control inputs with bounded error. Simulation results are included to show the effectiveness of the approach.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1757-69"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2160968","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30181105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
Energy-efficient FastICA implementation for biomedical signal separation. 生物医学信号分离的节能FastICA实现。
Pub Date : 2011-11-01 Epub Date: 2011-10-03 DOI: 10.1109/TNN.2011.2166979
Lan-Da Van, Di-You Wu, Chien-Shiun Chen

This paper presents an energy-efficient fast independent component analysis (FastICA) implementation with an early determination scheme for eight-channel electroencephalogram (EEG) signal separation. The main contributions are as follows: (1) energy-efficient FastICA using the proposed early determination scheme and the corresponding architecture; (2) cost-effective FastICA using the proposed preprocessing unit architecture with one coordinate rotation digital computer-based eigenvalue decomposition processor and the proposed one-unit architecture with the hardware reuse scheme; and (3) low-computation-time FastICA using the four parallel one-units architecture. The resulting power dissipation of the FastICA implementation for eight-channel EEG signal separation is 16.35 mW at 100 MHz at 1.0 V. Compared with the design without early determination, the proposed FastICA architecture implemented in united microelectronics corporation 90 nm 1P9M complementary metal-oxide-semiconductor process with a core area of 1.221 × 1.218 mm2 can achieve average energy reduction by 47.63%. From the post-layout simulation results, the maximum computation time is 0.29 s.

提出了一种高效、快速的独立分量分析(FastICA)实现方法,该方法具有八通道脑电图(EEG)信号分离的早期确定方案。主要贡献如下:(1)采用提出的早期确定方案和相应的体系结构的节能FastICA;(2)采用基于一个坐标旋转数字计算机特征值分解处理器的预处理单元体系结构和基于硬件复用方案的单单元体系结构的高性价比FastICA;(3)采用四并行单单元架构的低计算时间FastICA。FastICA实现的8通道脑电信号分离的功耗为16.35 mW,频率为100 MHz,电压为1.0 V。采用联合微电子公司90 nm 1P9M互补型金属氧化物半导体工艺,芯面积为1.221 × 1.218 mm2,与未进行前期设计相比,FastICA架构的平均能耗降低了47.63%。从布局后的仿真结果来看,最大计算时间为0.29 s。
{"title":"Energy-efficient FastICA implementation for biomedical signal separation.","authors":"Lan-Da Van,&nbsp;Di-You Wu,&nbsp;Chien-Shiun Chen","doi":"10.1109/TNN.2011.2166979","DOIUrl":"https://doi.org/10.1109/TNN.2011.2166979","url":null,"abstract":"<p><p>This paper presents an energy-efficient fast independent component analysis (FastICA) implementation with an early determination scheme for eight-channel electroencephalogram (EEG) signal separation. The main contributions are as follows: (1) energy-efficient FastICA using the proposed early determination scheme and the corresponding architecture; (2) cost-effective FastICA using the proposed preprocessing unit architecture with one coordinate rotation digital computer-based eigenvalue decomposition processor and the proposed one-unit architecture with the hardware reuse scheme; and (3) low-computation-time FastICA using the four parallel one-units architecture. The resulting power dissipation of the FastICA implementation for eight-channel EEG signal separation is 16.35 mW at 100 MHz at 1.0 V. Compared with the design without early determination, the proposed FastICA architecture implemented in united microelectronics corporation 90 nm 1P9M complementary metal-oxide-semiconductor process with a core area of 1.221 × 1.218 mm2 can achieve average energy reduction by 47.63%. From the post-layout simulation results, the maximum computation time is 0.29 s.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1809-22"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2166979","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30182517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Direct parallel perceptrons (DPPs): fast analytical calculation of the parallel perceptrons weights with margin control for classification tasks. 直接并行感知器(DPPs):用于分类任务的具有裕度控制的并行感知器权重的快速分析计算。
Pub Date : 2011-11-01 Epub Date: 2011-10-06 DOI: 10.1109/TNN.2011.2169086
Manuel Fernandez-Delgado, Jorge Ribeiro, Eva Cernadas, Senén Barro Ameneiro

Parallel perceptrons (PPs) are very simple and efficient committee machines (a single layer of perceptrons with threshold activation functions and binary outputs, and a majority voting decision scheme), which nevertheless behave as universal approximators. The parallel delta (P-Delta) rule is an effective training algorithm, which, following the ideas of statistical learning theory used by the support vector machine (SVM), raises its generalization ability by maximizing the difference between the perceptron activations for the training patterns and the activation threshold (which corresponds to the separating hyperplane). In this paper, we propose an analytical closed-form expression to calculate the PPs' weights for classification tasks. Our method, called Direct Parallel Perceptrons (DPPs), directly calculates (without iterations) the weights using the training patterns and their desired outputs, without any search or numeric function optimization. The calculated weights globally minimize an error function which simultaneously takes into account the training error and the classification margin. Given its analytical and noniterative nature, DPPs are computationally much more efficient than other related approaches (P-Delta and SVM), and its computational complexity is linear in the input dimensionality. Therefore, DPPs are very appealing, in terms of time complexity and memory consumption, and are very easy to use for high-dimensional classification tasks. On real benchmark datasets with two and multiple classes, DPPs are competitive with SVM and other approaches but they also allow online learning and, as opposed to most of them, have no tunable parameters.

并行感知器(PPs)是非常简单和高效的委员会机器(单层感知器,具有阈值激活函数和二进制输出,以及多数投票决策方案),但其表现为通用逼近器。并行delta (P-Delta)规则是一种有效的训练算法,它遵循支持向量机(SVM)使用的统计学习理论的思想,通过最大化训练模式的感知器激活与激活阈值(对应于分离超平面)之间的差异来提高其泛化能力。在本文中,我们提出了一个解析的封闭表达式来计算分类任务的PPs的权重。我们的方法,称为直接并行感知器(DPPs),直接计算(不迭代)使用训练模式及其期望输出的权重,而不需要任何搜索或数值函数优化。计算出的权重全局最小化误差函数,同时考虑训练误差和分类余量。考虑到它的解析性和非迭代性,dpp在计算上比其他相关方法(P-Delta和SVM)要高效得多,而且它的计算复杂度在输入维度上是线性的。因此,就时间复杂度和内存消耗而言,dpp非常吸引人,并且非常易于用于高维分类任务。在具有两个或多个类的真实基准数据集上,dpp与SVM和其他方法竞争,但它们也允许在线学习,并且与大多数方法相反,没有可调参数。
{"title":"Direct parallel perceptrons (DPPs): fast analytical calculation of the parallel perceptrons weights with margin control for classification tasks.","authors":"Manuel Fernandez-Delgado,&nbsp;Jorge Ribeiro,&nbsp;Eva Cernadas,&nbsp;Senén Barro Ameneiro","doi":"10.1109/TNN.2011.2169086","DOIUrl":"https://doi.org/10.1109/TNN.2011.2169086","url":null,"abstract":"<p><p>Parallel perceptrons (PPs) are very simple and efficient committee machines (a single layer of perceptrons with threshold activation functions and binary outputs, and a majority voting decision scheme), which nevertheless behave as universal approximators. The parallel delta (P-Delta) rule is an effective training algorithm, which, following the ideas of statistical learning theory used by the support vector machine (SVM), raises its generalization ability by maximizing the difference between the perceptron activations for the training patterns and the activation threshold (which corresponds to the separating hyperplane). In this paper, we propose an analytical closed-form expression to calculate the PPs' weights for classification tasks. Our method, called Direct Parallel Perceptrons (DPPs), directly calculates (without iterations) the weights using the training patterns and their desired outputs, without any search or numeric function optimization. The calculated weights globally minimize an error function which simultaneously takes into account the training error and the classification margin. Given its analytical and noniterative nature, DPPs are computationally much more efficient than other related approaches (P-Delta and SVM), and its computational complexity is linear in the input dimensionality. Therefore, DPPs are very appealing, in terms of time complexity and memory consumption, and are very easy to use for high-dimensional classification tasks. On real benchmark datasets with two and multiple classes, DPPs are competitive with SVM and other approaches but they also allow online learning and, as opposed to most of them, have no tunable parameters.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1837-48"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2169086","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30196886","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Multistability of second-order competitive neural networks with nondecreasing saturated activation functions. 非递减饱和激活函数二阶竞争神经网络的多稳定性。
Pub Date : 2011-11-01 Epub Date: 2011-09-06 DOI: 10.1109/TNN.2011.2164934
Xiaobing Nie, Jinde Cao

In this paper, second-order interactions are introduced into competitive neural networks (NNs) and the multistability is discussed for second-order competitive NNs (SOCNNs) with nondecreasing saturated activation functions. Firstly, based on decomposition of state space, Cauchy convergence principle, and inequality technique, some sufficient conditions ensuring the local exponential stability of 2N equilibrium points are derived. Secondly, some conditions are obtained for ascertaining equilibrium points to be locally exponentially stable and to be located in any designated region. Thirdly, the theory is extended to more general saturated activation functions with 2r corner points and a sufficient criterion is given under which the SOCNNs can have (r+1)N locally exponentially stable equilibrium points. Even if there is no second-order interactions, the obtained results are less restrictive than those in some recent works. Finally, three examples with their simulations are presented to verify the theoretical analysis.

将二阶相互作用引入竞争神经网络,讨论了具有非递减饱和激活函数的二阶竞争神经网络的多稳定性问题。首先,利用状态空间分解、柯西收敛原理和不等式技术,导出了保证2N个平衡点局部指数稳定的充分条件;其次,给出了确定平衡点局部指数稳定并位于任意指定区域的若干条件;第三,将该理论推广到具有2r个角点的更一般的饱和激活函数,并给出了SOCNNs具有(r+1)N个局部指数稳定平衡点的充分准则。即使不存在二阶相互作用,所得到的结果也比最近一些研究的结果限制性更小。最后给出了三个算例并进行了仿真,验证了理论分析的正确性。
{"title":"Multistability of second-order competitive neural networks with nondecreasing saturated activation functions.","authors":"Xiaobing Nie,&nbsp;Jinde Cao","doi":"10.1109/TNN.2011.2164934","DOIUrl":"https://doi.org/10.1109/TNN.2011.2164934","url":null,"abstract":"<p><p>In this paper, second-order interactions are introduced into competitive neural networks (NNs) and the multistability is discussed for second-order competitive NNs (SOCNNs) with nondecreasing saturated activation functions. Firstly, based on decomposition of state space, Cauchy convergence principle, and inequality technique, some sufficient conditions ensuring the local exponential stability of 2N equilibrium points are derived. Secondly, some conditions are obtained for ascertaining equilibrium points to be locally exponentially stable and to be located in any designated region. Thirdly, the theory is extended to more general saturated activation functions with 2r corner points and a sufficient criterion is given under which the SOCNNs can have (r+1)N locally exponentially stable equilibrium points. Even if there is no second-order interactions, the obtained results are less restrictive than those in some recent works. Finally, three examples with their simulations are presented to verify the theoretical analysis.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 11","pages":"1694-708"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2164934","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30127259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 64
Nonlinear regularization path for quadratic loss support vector machines. 二次损失支持向量机的非线性正则化路径。
Pub Date : 2011-10-01 Epub Date: 2011-08-30 DOI: 10.1109/TNN.2011.2164265
Masayuki Karasuyama, Ichiro Takeuchi

Regularization path algorithms have been proposed to deal with model selection problem in several machine learning approaches. These algorithms allow computation of the entire path of solutions for every value of regularization parameter using the fact that their solution paths have piecewise linear form. In this paper, we extend the applicability of regularization path algorithm to a class of learning machines that have quadratic loss and quadratic penalty term. This class contains several important learning machines such as squared hinge loss support vector machine (SVM) and modified Huber loss SVM. We first show that the solution paths of this class of learning machines have piecewise nonlinear form, and piecewise segments between two breakpoints are characterized by a class of rational functions. Then we develop an algorithm that can efficiently follow the piecewise nonlinear path by solving these rational equations. To solve these rational equations, we use rational approximation technique with quadratic convergence rate, and thus, our algorithm can follow the nonlinear path much more precisely than existing approaches such as predictor-corrector type nonlinear-path approximation. We show the algorithm performance on some artificial and real data sets.

正则化路径算法已被提出用于处理几种机器学习方法中的模型选择问题。这些算法利用其解路径具有分段线性形式的事实,允许计算每个正则化参数值的整个解路径。本文将正则化路径算法的适用性扩展到一类具有二次损失和二次惩罚项的学习机。本课程包含了几种重要的学习机,如平方铰链损失支持向量机(squared hinge loss support vector machine, SVM)和改进Huber损失支持向量机。我们首先证明了这类学习机的解路径具有分段非线性形式,并且两个断点之间的分段用一类有理函数来表征。然后通过求解这些有理方程,提出了一种能有效跟踪分段非线性路径的算法。为了求解这些有理数方程,我们使用二次收敛率的有理数近似技术,因此,我们的算法可以比现有的预测-校正型非线性路径近似方法更精确地跟踪非线性路径。我们在一些人工数据集和真实数据集上展示了算法的性能。
{"title":"Nonlinear regularization path for quadratic loss support vector machines.","authors":"Masayuki Karasuyama,&nbsp;Ichiro Takeuchi","doi":"10.1109/TNN.2011.2164265","DOIUrl":"https://doi.org/10.1109/TNN.2011.2164265","url":null,"abstract":"<p><p>Regularization path algorithms have been proposed to deal with model selection problem in several machine learning approaches. These algorithms allow computation of the entire path of solutions for every value of regularization parameter using the fact that their solution paths have piecewise linear form. In this paper, we extend the applicability of regularization path algorithm to a class of learning machines that have quadratic loss and quadratic penalty term. This class contains several important learning machines such as squared hinge loss support vector machine (SVM) and modified Huber loss SVM. We first show that the solution paths of this class of learning machines have piecewise nonlinear form, and piecewise segments between two breakpoints are characterized by a class of rational functions. Then we develop an algorithm that can efficiently follow the piecewise nonlinear path by solving these rational equations. To solve these rational equations, we use rational approximation technique with quadratic convergence rate, and thus, our algorithm can follow the nonlinear path much more precisely than existing approaches such as predictor-corrector type nonlinear-path approximation. We show the algorithm performance on some artificial and real data sets.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 10","pages":"1613-25"},"PeriodicalIF":0.0,"publicationDate":"2011-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2164265","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30110076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Textual and visual content-based anti-phishing: a Bayesian approach. 基于文本和视觉内容的反网络钓鱼:贝叶斯方法。
Pub Date : 2011-10-01 Epub Date: 2011-08-04 DOI: 10.1109/TNN.2011.2161999
Haijun Zhang, Gang Liu, Tommy W S Chow, Wenyin Liu

A novel framework using a Bayesian approach for content-based phishing web page detection is presented. Our model takes into account textual and visual contents to measure the similarity between the protected web page and suspicious web pages. A text classifier, an image classifier, and an algorithm fusing the results from classifiers are introduced. An outstanding feature of this paper is the exploration of a Bayesian model to estimate the matching threshold. This is required in the classifier for determining the class of the web page and identifying whether the web page is phishing or not. In the text classifier, the naive Bayes rule is used to calculate the probability that a web page is phishing. In the image classifier, the earth mover's distance is employed to measure the visual similarity, and our Bayesian model is designed to determine the threshold. In the data fusion algorithm, the Bayes theory is used to synthesize the classification results from textual and visual content. The effectiveness of our proposed approach was examined in a large-scale dataset collected from real phishing cases. Experimental results demonstrated that the text classifier and the image classifier we designed deliver promising results, the fusion algorithm outperforms either of the individual classifiers, and our model can be adapted to different phishing cases.

提出了一种基于贝叶斯方法的基于内容的网络钓鱼网页检测框架。我们的模型考虑了文本和视觉内容来衡量受保护网页和可疑网页之间的相似性。介绍了一种文本分类器、图像分类器和一种融合分类器结果的算法。本文的一个突出特点是探索了贝叶斯模型来估计匹配阈值。这在分类器中是必需的,用于确定网页的类别,并识别网页是否为网络钓鱼。在文本分类器中,使用朴素贝叶斯规则计算网页钓鱼的概率。在图像分类器中,采用推土机的距离来衡量视觉相似性,设计贝叶斯模型来确定阈值。在数据融合算法中,利用贝叶斯理论对文本和视觉内容的分类结果进行综合。我们提出的方法的有效性在从真实网络钓鱼案例中收集的大规模数据集中进行了检验。实验结果表明,本文设计的文本分类器和图像分类器效果良好,融合算法优于单独的分类器,并且该模型可以适应不同的网络钓鱼案例。
{"title":"Textual and visual content-based anti-phishing: a Bayesian approach.","authors":"Haijun Zhang,&nbsp;Gang Liu,&nbsp;Tommy W S Chow,&nbsp;Wenyin Liu","doi":"10.1109/TNN.2011.2161999","DOIUrl":"https://doi.org/10.1109/TNN.2011.2161999","url":null,"abstract":"<p><p>A novel framework using a Bayesian approach for content-based phishing web page detection is presented. Our model takes into account textual and visual contents to measure the similarity between the protected web page and suspicious web pages. A text classifier, an image classifier, and an algorithm fusing the results from classifiers are introduced. An outstanding feature of this paper is the exploration of a Bayesian model to estimate the matching threshold. This is required in the classifier for determining the class of the web page and identifying whether the web page is phishing or not. In the text classifier, the naive Bayes rule is used to calculate the probability that a web page is phishing. In the image classifier, the earth mover's distance is employed to measure the visual similarity, and our Bayesian model is designed to determine the threshold. In the data fusion algorithm, the Bayes theory is used to synthesize the classification results from textual and visual content. The effectiveness of our proposed approach was examined in a large-scale dataset collected from real phishing cases. Experimental results demonstrated that the text classifier and the image classifier we designed deliver promising results, the fusion algorithm outperforms either of the individual classifiers, and our model can be adapted to different phishing cases.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 10","pages":"1532-46"},"PeriodicalIF":0.0,"publicationDate":"2011-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2161999","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30063663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 193
A new formulation for feedforward neural networks. 前馈神经网络的新公式。
Pub Date : 2011-10-01 Epub Date: 2011-08-22 DOI: 10.1109/TNN.2011.2163169
Saman Razavi, Bryan A Tolson

Feedforward neural network is one of the most commonly used function approximation techniques and has been applied to a wide variety of problems arising from various disciplines. However, neural networks are black-box models having multiple challenges/difficulties associated with training and generalization. This paper initially looks into the internal behavior of neural networks and develops a detailed interpretation of the neural network functional geometry. Based on this geometrical interpretation, a new set of variables describing neural networks is proposed as a more effective and geometrically interpretable alternative to the traditional set of network weights and biases. Then, this paper develops a new formulation for neural networks with respect to the newly defined variables; this reformulated neural network (ReNN) is equivalent to the common feedforward neural network but has a less complex error response surface. To demonstrate the learning ability of ReNN, in this paper, two training methods involving a derivative-based (a variation of backpropagation) and a derivative-free optimization algorithms are employed. Moreover, a new measure of regularization on the basis of the developed geometrical interpretation is proposed to evaluate and improve the generalization ability of neural networks. The value of the proposed geometrical interpretation, the ReNN approach, and the new regularization measure are demonstrated across multiple test problems. Results show that ReNN can be trained more effectively and efficiently compared to the common neural networks and the proposed regularization measure is an effective indicator of how a network would perform in terms of generalization.

前馈神经网络是最常用的函数逼近技术之一,已广泛应用于各个学科的各种问题。然而,神经网络是具有与训练和泛化相关的多重挑战/困难的黑盒模型。本文首先研究了神经网络的内部行为,并对神经网络函数几何进行了详细的解释。基于这种几何解释,提出了一种新的描述神经网络的变量集,作为一种更有效和几何上可解释的替代传统的网络权重和偏差集。然后,针对新定义的变量,本文给出了一个新的神经网络公式;这种改进型神经网络(ReNN)相当于普通的前馈神经网络,但其误差响应面不那么复杂。为了证明ReNN的学习能力,本文采用了基于导数(反向传播的一种变化)和无导数优化算法的两种训练方法。此外,在已有的几何解释的基础上,提出了一种新的正则化测度来评价和提高神经网络的泛化能力。在多个测试问题中证明了所提出的几何解释、ReNN方法和新的正则化度量的价值。结果表明,与普通神经网络相比,ReNN可以更有效地训练,并且所提出的正则化度量是网络在泛化方面表现如何的有效指标。
{"title":"A new formulation for feedforward neural networks.","authors":"Saman Razavi,&nbsp;Bryan A Tolson","doi":"10.1109/TNN.2011.2163169","DOIUrl":"https://doi.org/10.1109/TNN.2011.2163169","url":null,"abstract":"<p><p>Feedforward neural network is one of the most commonly used function approximation techniques and has been applied to a wide variety of problems arising from various disciplines. However, neural networks are black-box models having multiple challenges/difficulties associated with training and generalization. This paper initially looks into the internal behavior of neural networks and develops a detailed interpretation of the neural network functional geometry. Based on this geometrical interpretation, a new set of variables describing neural networks is proposed as a more effective and geometrically interpretable alternative to the traditional set of network weights and biases. Then, this paper develops a new formulation for neural networks with respect to the newly defined variables; this reformulated neural network (ReNN) is equivalent to the common feedforward neural network but has a less complex error response surface. To demonstrate the learning ability of ReNN, in this paper, two training methods involving a derivative-based (a variation of backpropagation) and a derivative-free optimization algorithms are employed. Moreover, a new measure of regularization on the basis of the developed geometrical interpretation is proposed to evaluate and improve the generalization ability of neural networks. The value of the proposed geometrical interpretation, the ReNN approach, and the new regularization measure are demonstrated across multiple test problems. Results show that ReNN can be trained more effectively and efficiently compared to the common neural networks and the proposed regularization measure is an effective indicator of how a network would perform in terms of generalization.</p>","PeriodicalId":13434,"journal":{"name":"IEEE transactions on neural networks","volume":"22 10","pages":"1588-98"},"PeriodicalIF":0.0,"publicationDate":"2011-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/TNN.2011.2163169","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"30092861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 118
期刊
IEEE transactions on neural networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1