首页 > 最新文献

2015 International Joint Conference on Neural Networks (IJCNN)最新文献

英文 中文
Splitting with confidence in decision trees with application to stream mining 在决策树中进行有信心的拆分,并应用于流挖掘
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280392
R. D. Rosa, N. Cesa-Bianchi
Decision tree classifiers are a widely used tool in data stream mining. The use of confidence intervals to estimate the gain associated with each split leads to very effective methods, like the popular Hoeffding tree algorithm. From a statistical viewpoint, the analysis of decision tree classifiers in a streaming setting requires knowing when enough new information has been collected to justify splitting a leaf. Although some of the issues in the statistical analysis of Hoeffding trees have been already clarified, a general and rigorous study of confidence intervals for splitting criteria is missing. We fill this gap by deriving accurate confidence intervals to estimate the splitting gain in decision tree learning with respect to three criteria: entropy, Gini index, and a third index proposed by Kearns and Mansour. Our confidence intervals depend in a more detailed way on the tree parameters. Experiments on real and synthetic data in a streaming setting show that our trees are indeed more accurate than trees with the same number of leaves generated by other techniques.
决策树分类器是数据流挖掘中广泛使用的一种工具。使用置信区间来估计与每次分割相关的增益会产生非常有效的方法,比如流行的Hoeffding树算法。从统计学的角度来看,流设置中的决策树分类器的分析需要知道何时收集了足够的新信息来证明分割叶子是合理的。虽然Hoeffding树的统计分析中的一些问题已经得到澄清,但对分裂标准的置信区间的一般和严格的研究仍然缺失。我们通过推导准确的置信区间来估计决策树学习中的分裂增益,以三个标准来填补这一空白:熵、基尼指数和Kearns和Mansour提出的第三个指标。我们的置信区间更详细地取决于树的参数。在流式设置中对真实和合成数据进行的实验表明,我们的树确实比其他技术生成的具有相同叶子数量的树更准确。
{"title":"Splitting with confidence in decision trees with application to stream mining","authors":"R. D. Rosa, N. Cesa-Bianchi","doi":"10.1109/IJCNN.2015.7280392","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280392","url":null,"abstract":"Decision tree classifiers are a widely used tool in data stream mining. The use of confidence intervals to estimate the gain associated with each split leads to very effective methods, like the popular Hoeffding tree algorithm. From a statistical viewpoint, the analysis of decision tree classifiers in a streaming setting requires knowing when enough new information has been collected to justify splitting a leaf. Although some of the issues in the statistical analysis of Hoeffding trees have been already clarified, a general and rigorous study of confidence intervals for splitting criteria is missing. We fill this gap by deriving accurate confidence intervals to estimate the splitting gain in decision tree learning with respect to three criteria: entropy, Gini index, and a third index proposed by Kearns and Mansour. Our confidence intervals depend in a more detailed way on the tree parameters. Experiments on real and synthetic data in a streaming setting show that our trees are indeed more accurate than trees with the same number of leaves generated by other techniques.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"108 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76330867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Real-time video object recognition using convolutional neural network 基于卷积神经网络的实时视频目标识别
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280718
Byungik Ahn
A convolutional neural network (CNN) is implemented on a field-programmable gate array (FPGA) and used for recognizing objects in real-time video streams. In this system, an image pyramid is constructed by successively down-scaling the input video stream. Image blocks are extracted from the image pyramid and classified by the CNN core. The detected parts are then marked on the output video frames. The CNN core is composed of six hardware neurons and two receptor units. The hardware neurons are designed as fully-pipelined digital circuits synchronized with the system clock, and are used to compute the model neurons in a time-sharing manner. The receptor units scan the input image for local receptive fields and continuously supply data to the hardware neurons as inputs. The CNN core module is controlled according to the contents of a table describing the sequence of computational stages and containing the system parameters required to control each stage. The use of this table makes the hardware system more flexible, and various CNN configurations can be accommodated without re-designing the system. The system implemented on a mid-range FPGA achieves a computational speed greater than 170,000 classifications per second, and performs scale-invariant object recognition from a 720×480 video stream at a speed of 60 fps. This work is a part of a commercial project, and the system is targeted for recognizing any pre-trained objects with a small physical volume and low power consumption.
在现场可编程门阵列(FPGA)上实现了卷积神经网络(CNN),并将其用于实时视频流中的目标识别。在该系统中,通过对输入视频流的逐次降尺度构建图像金字塔。从图像金字塔中提取图像块,并通过CNN核心进行分类。然后在输出视频帧上标记检测到的部分。CNN核心由6个硬件神经元和2个受体单元组成。硬件神经元被设计成与系统时钟同步的全流水线数字电路,并以分时方式计算模型神经元。受体单元扫描输入图像,寻找局部接受野,并不断向硬件神经元提供数据作为输入。CNN核心模块是根据一个表的内容进行控制的,该表描述了计算阶段的顺序,并包含了控制每个阶段所需的系统参数。该表的使用使硬件系统更加灵活,无需重新设计系统即可容纳各种CNN配置。该系统在中档FPGA上实现,计算速度超过每秒170,000个分类,并以60 fps的速度从720×480视频流中执行缩放不变的目标识别。这项工作是一个商业项目的一部分,该系统的目标是识别任何物理体积小、功耗低的预训练对象。
{"title":"Real-time video object recognition using convolutional neural network","authors":"Byungik Ahn","doi":"10.1109/IJCNN.2015.7280718","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280718","url":null,"abstract":"A convolutional neural network (CNN) is implemented on a field-programmable gate array (FPGA) and used for recognizing objects in real-time video streams. In this system, an image pyramid is constructed by successively down-scaling the input video stream. Image blocks are extracted from the image pyramid and classified by the CNN core. The detected parts are then marked on the output video frames. The CNN core is composed of six hardware neurons and two receptor units. The hardware neurons are designed as fully-pipelined digital circuits synchronized with the system clock, and are used to compute the model neurons in a time-sharing manner. The receptor units scan the input image for local receptive fields and continuously supply data to the hardware neurons as inputs. The CNN core module is controlled according to the contents of a table describing the sequence of computational stages and containing the system parameters required to control each stage. The use of this table makes the hardware system more flexible, and various CNN configurations can be accommodated without re-designing the system. The system implemented on a mid-range FPGA achieves a computational speed greater than 170,000 classifications per second, and performs scale-invariant object recognition from a 720×480 video stream at a speed of 60 fps. This work is a part of a commercial project, and the system is targeted for recognizing any pre-trained objects with a small physical volume and low power consumption.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"92 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86435503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A minimal architecture for general cognition 一般认知的最小架构
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280749
Michael S. Gashler, Zachariah Kindle, Michael R. Smith
A minimalistic cognitive architecture called MANIC is presented. The MANIC architecture requires only three function approximating models, and one state machine. Even with so few major components, it is theoretically sufficient to achieve functional equivalence with all other cognitive architectures, and can be practically trained. Instead of seeking to trasfer architectural inspiration from biology into artificial intelligence, MANIC seeks to minimize novelty and follow the most well-established constructs that have evolved within various subfields of data science. From this perspective, MANIC offers an alternate approach to a long-standing objective of artificial intelligence. This paper provides a theoretical analysis of the MANIC architecture.
提出了一种极简的认知架构,称为MANIC。MANIC体系结构只需要三个函数近似模型和一个状态机。即使使用如此少的主要组件,理论上也足以实现与所有其他认知体系结构的功能等同,并且可以进行实际训练。MANIC并没有寻求将生物学的建筑灵感转移到人工智能中,而是寻求将新颖性降到最低,并遵循在数据科学的各个子领域中发展起来的最完善的结构。从这个角度来看,MANIC为人工智能的长期目标提供了另一种方法。本文对MANIC体系结构进行了理论分析。
{"title":"A minimal architecture for general cognition","authors":"Michael S. Gashler, Zachariah Kindle, Michael R. Smith","doi":"10.1109/IJCNN.2015.7280749","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280749","url":null,"abstract":"A minimalistic cognitive architecture called MANIC is presented. The MANIC architecture requires only three function approximating models, and one state machine. Even with so few major components, it is theoretically sufficient to achieve functional equivalence with all other cognitive architectures, and can be practically trained. Instead of seeking to trasfer architectural inspiration from biology into artificial intelligence, MANIC seeks to minimize novelty and follow the most well-established constructs that have evolved within various subfields of data science. From this perspective, MANIC offers an alternate approach to a long-standing objective of artificial intelligence. This paper provides a theoretical analysis of the MANIC architecture.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"52 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86470877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image segmentation using fast linking SCM 基于单片机的快速链接图像分割
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280579
K. Zhan, Jinhui Shi, Qiaoqiao Li, Jicai Teng, Mingying Wang
Spiking cortical model (SCM) is applied to image segmentation. A natural image is processed to produce a series of spike images by SCM, and the segmented result is obtained by the integration of the series of spike images. An appropriate maximum iterative times is selected to achieve an optimal threshold of SCM. In each iteration, neurons that produced spikes correspond to pixels with an intensity of the input natural image approximately. SCM synchronizes the output spikes via the fast linking synaptic modulation, which makes objects in the image as homogeneous as possible. Experimental results show that the output image not only separates objects and background well, but also pixels in each object are homogeneous. The proposed method performs well over other methods and the quantitative metrics are consistent with the visual performance.
将脉冲皮质模型(SCM)应用于图像分割。利用单片机对自然图像进行处理,产生一系列的尖峰图像,对这些尖峰图像进行积分得到分割结果。选择合适的最大迭代次数以达到SCM的最优阈值。在每次迭代中,产生尖峰的神经元与输入自然图像的强度近似对应。单片机通过快速链接突触调制同步输出尖峰,使图像中的物体尽可能均匀。实验结果表明,输出图像不仅能很好地分离物体和背景,而且每个物体的像素都是均匀的。该方法优于其他方法,且定量指标与视觉性能一致。
{"title":"Image segmentation using fast linking SCM","authors":"K. Zhan, Jinhui Shi, Qiaoqiao Li, Jicai Teng, Mingying Wang","doi":"10.1109/IJCNN.2015.7280579","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280579","url":null,"abstract":"Spiking cortical model (SCM) is applied to image segmentation. A natural image is processed to produce a series of spike images by SCM, and the segmented result is obtained by the integration of the series of spike images. An appropriate maximum iterative times is selected to achieve an optimal threshold of SCM. In each iteration, neurons that produced spikes correspond to pixels with an intensity of the input natural image approximately. SCM synchronizes the output spikes via the fast linking synaptic modulation, which makes objects in the image as homogeneous as possible. Experimental results show that the output image not only separates objects and background well, but also pixels in each object are homogeneous. The proposed method performs well over other methods and the quantitative metrics are consistent with the visual performance.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"24 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83979014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Spatio-temporal Map Formation based on a Potential Function 基于势函数的时空地图生成
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280399
Prayag Gowgi, S. G. Srinivasa
We revisit the problem of temporal self organization using activity diffusion based on the neural gas (NGAS) algorithm. Using a potential function formulation motivated by a spatio-temporal metric, we derive an adaptation rule for dynamic vector quantization of data. Simulations results show that our algorithm learns the input distribution and time correlation much faster compared to the static neural gas method over the same data sequence under similar training conditions.
我们重新研究了基于神经气体(NGAS)算法的活动扩散的时间自组织问题。利用时空度量驱动的势函数公式,推导了数据动态矢量量化的自适应规则。仿真结果表明,在相似的训练条件下,在相同的数据序列上,我们的算法学习输入分布和时间相关性的速度比静态神经气体方法快得多。
{"title":"Spatio-temporal Map Formation based on a Potential Function","authors":"Prayag Gowgi, S. G. Srinivasa","doi":"10.1109/IJCNN.2015.7280399","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280399","url":null,"abstract":"We revisit the problem of temporal self organization using activity diffusion based on the neural gas (NGAS) algorithm. Using a potential function formulation motivated by a spatio-temporal metric, we derive an adaptation rule for dynamic vector quantization of data. Simulations results show that our algorithm learns the input distribution and time correlation much faster compared to the static neural gas method over the same data sequence under similar training conditions.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"449 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82951418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Recognizing visual composite in real images 识别真实图像中的视觉合成
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280523
Lin Bai, Kan Li, Shuai Jiang
Automatically discovering and recognizing the main structured visual pattern of an image is a challenging problem. The most difficulties are how to find the component objects and how to recognize the interaction among these objects. The component objects of the structured visual pattern have consistent 3D spatial co-occurrence layout across images, which manifest themselves as a predictable pattern called visual composite. In this paper, we propose a visual composite recognition model to automatically discover and recognize the visual composite of an image. Our model firstly learns 3D spatial co-occurrence statistics among objects to discover the potential structured visual pattern of an image so that it captures the component objects of visual composite. Secondly, we construct a feedforward architecture using the proposed factored three-way interaction machine to recognize the visual composite, which casts the recognition problem as a structured prediction task. It predicts the visual composite by maximizing the probability of the correct structured label given the component objects and their 3D spatial context. Experiments conducted on a six-class sports dataset and a phrasal recognition dataset respectively demonstrate the encouraging performance of our model in discovery precision and recognition accuracy compared with competing approaches.
自动发现和识别图像的主要结构视觉模式是一个具有挑战性的问题。最困难的是如何找到组件对象以及如何识别这些对象之间的交互。结构化视觉模式的组件对象在图像之间具有一致的三维空间共现布局,这表现为一种可预测的模式,称为视觉复合。本文提出了一种视觉合成识别模型,用于自动发现和识别图像的视觉合成。我们的模型首先学习对象之间的三维空间共现统计,发现图像潜在的结构化视觉模式,从而捕获视觉复合的组成对象。其次,我们利用提出的因子三向交互机器构造前馈结构来识别视觉组合,将识别问题转化为结构化预测任务。它通过最大化给定组件对象及其3D空间上下文的正确结构化标签的概率来预测视觉组合。在一个六类运动数据集和一个短语识别数据集上进行的实验表明,与竞争对手的方法相比,我们的模型在发现精度和识别精度方面都有令人鼓舞的表现。
{"title":"Recognizing visual composite in real images","authors":"Lin Bai, Kan Li, Shuai Jiang","doi":"10.1109/IJCNN.2015.7280523","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280523","url":null,"abstract":"Automatically discovering and recognizing the main structured visual pattern of an image is a challenging problem. The most difficulties are how to find the component objects and how to recognize the interaction among these objects. The component objects of the structured visual pattern have consistent 3D spatial co-occurrence layout across images, which manifest themselves as a predictable pattern called visual composite. In this paper, we propose a visual composite recognition model to automatically discover and recognize the visual composite of an image. Our model firstly learns 3D spatial co-occurrence statistics among objects to discover the potential structured visual pattern of an image so that it captures the component objects of visual composite. Secondly, we construct a feedforward architecture using the proposed factored three-way interaction machine to recognize the visual composite, which casts the recognition problem as a structured prediction task. It predicts the visual composite by maximizing the probability of the correct structured label given the component objects and their 3D spatial context. Experiments conducted on a six-class sports dataset and a phrasal recognition dataset respectively demonstrate the encouraging performance of our model in discovery precision and recognition accuracy compared with competing approaches.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"1 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82959361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Polyphonic sound event detection using multi label deep neural networks 基于多标签深度神经网络的复音事件检测
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280624
Emre Çakir, T. Heittola, H. Huttunen, T. Virtanen
In this paper, the use of multi label neural networks are proposed for detection of temporally overlapping sound events in realistic environments. Real-life sound recordings typically have many overlapping sound events, making it hard to recognize each event with the standard sound event detection methods. Frame-wise spectral-domain features are used as inputs to train a deep neural network for multi label classification in this work. The model is evaluated with recordings from realistic everyday environments and the obtained overall accuracy is 63.8%. The method is compared against a state-of-the-art method using non-negative matrix factorization as a pre-processing stage and hidden Markov models as a classifier. The proposed method improves the accuracy by 19% percentage points overall.
本文提出使用多标签神经网络来检测现实环境中时间重叠的声音事件。现实生活中的录音通常有许多重叠的声音事件,因此很难用标准的声音事件检测方法识别每个事件。在这项工作中,使用帧频谱域特征作为输入来训练用于多标签分类的深度神经网络。用真实的日常环境记录对模型进行了评估,得到的总体准确率为63.8%。该方法与使用非负矩阵分解作为预处理阶段和隐马尔可夫模型作为分类器的最先进方法进行了比较。该方法总体上提高了19%的准确率。
{"title":"Polyphonic sound event detection using multi label deep neural networks","authors":"Emre Çakir, T. Heittola, H. Huttunen, T. Virtanen","doi":"10.1109/IJCNN.2015.7280624","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280624","url":null,"abstract":"In this paper, the use of multi label neural networks are proposed for detection of temporally overlapping sound events in realistic environments. Real-life sound recordings typically have many overlapping sound events, making it hard to recognize each event with the standard sound event detection methods. Frame-wise spectral-domain features are used as inputs to train a deep neural network for multi label classification in this work. The model is evaluated with recordings from realistic everyday environments and the obtained overall accuracy is 63.8%. The method is compared against a state-of-the-art method using non-negative matrix factorization as a pre-processing stage and hidden Markov models as a classifier. The proposed method improves the accuracy by 19% percentage points overall.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"6 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80707065","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 269
Joint adaptive loss and l2/l0-norm minimization for unsupervised feature selection 联合自适应损失和l2/ 10范数最小化的无监督特征选择
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280307
Mingjie Qian, ChengXiang Zhai
Unsupervised feature selection is a useful tool for reducing the complexity and improving the generalization performance of data mining tasks. In this paper, we propose an Adaptive Unsupervised Feature Selection (AUFS) algorithm with explicit l2/l0-norm minimization. We use a joint adaptive loss for data fitting and a l2/l0 minimization for feature selection. We solve the optimization problem with an efficient iterative algorithm and prove that all the expected properties of unsupervised feature selection can be preserved. We also show that the computational complexity and memory use is only linear to the number of instances and square to the number of clusters. Experiments show that our algorithm outperforms the state-of-the-arts on seven different benchmark data sets.
无监督特征选择是降低数据挖掘任务复杂性和提高数据挖掘泛化性能的有效工具。在本文中,我们提出了一种具有显式l2/ 10范数最小化的自适应无监督特征选择(AUFS)算法。我们使用联合自适应损失进行数据拟合,并使用l2/ 10最小化进行特征选择。我们用一种高效的迭代算法解决了优化问题,并证明了无监督特征选择的所有预期性质都可以保持。我们还表明,计算复杂度和内存使用仅与实例数量成线性关系,与集群数量成平方关系。实验表明,我们的算法在7个不同的基准数据集上的性能优于目前最先进的算法。
{"title":"Joint adaptive loss and l2/l0-norm minimization for unsupervised feature selection","authors":"Mingjie Qian, ChengXiang Zhai","doi":"10.1109/IJCNN.2015.7280307","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280307","url":null,"abstract":"Unsupervised feature selection is a useful tool for reducing the complexity and improving the generalization performance of data mining tasks. In this paper, we propose an Adaptive Unsupervised Feature Selection (AUFS) algorithm with explicit l2/l0-norm minimization. We use a joint adaptive loss for data fitting and a l2/l0 minimization for feature selection. We solve the optimization problem with an efficient iterative algorithm and prove that all the expected properties of unsupervised feature selection can be preserved. We also show that the computational complexity and memory use is only linear to the number of instances and square to the number of clusters. Experiments show that our algorithm outperforms the state-of-the-arts on seven different benchmark data sets.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"52 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80769416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
DeepSign: Deep learning for automatic malware signature generation and classification DeepSign:用于自动恶意软件签名生成和分类的深度学习
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280815
Omid David, N. Netanyahu
This paper presents a novel deep learning based method for automatic malware signature generation and classification. The method uses a deep belief network (DBN), implemented with a deep stack of denoising autoencoders, generating an invariant compact representation of the malware behavior. While conventional signature and token based methods for malware detection do not detect a majority of new variants for existing malware, the results presented in this paper show that signatures generated by the DBN allow for an accurate classification of new malware variants. Using a dataset containing hundreds of variants for several major malware families, our method achieves 98.6% classification accuracy using the signatures generated by the DBN. The presented method is completely agnostic to the type of malware behavior that is logged (e.g., API calls and their parameters, registry entries, websites and ports accessed, etc.), and can use any raw input from a sandbox to successfully train the deep neural network which is used to generate malware signatures.
提出了一种基于深度学习的恶意软件签名自动生成与分类方法。该方法使用深度信念网络(DBN),由深度去噪自编码器堆栈实现,生成恶意软件行为的不变紧凑表示。虽然传统的基于签名和令牌的恶意软件检测方法不能检测到现有恶意软件的大多数新变体,但本文提出的结果表明,DBN生成的签名允许对新的恶意软件变体进行准确分类。使用包含几个主要恶意软件家族的数百个变体的数据集,我们的方法使用DBN生成的签名实现了98.6%的分类准确率。所提出的方法是完全不可知的类型的恶意软件行为记录(例如,API调用及其参数,注册表项,网站和端口访问等),并可以使用任何原始输入从沙箱成功训练深度神经网络,这是用来生成恶意软件签名。
{"title":"DeepSign: Deep learning for automatic malware signature generation and classification","authors":"Omid David, N. Netanyahu","doi":"10.1109/IJCNN.2015.7280815","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280815","url":null,"abstract":"This paper presents a novel deep learning based method for automatic malware signature generation and classification. The method uses a deep belief network (DBN), implemented with a deep stack of denoising autoencoders, generating an invariant compact representation of the malware behavior. While conventional signature and token based methods for malware detection do not detect a majority of new variants for existing malware, the results presented in this paper show that signatures generated by the DBN allow for an accurate classification of new malware variants. Using a dataset containing hundreds of variants for several major malware families, our method achieves 98.6% classification accuracy using the signatures generated by the DBN. The presented method is completely agnostic to the type of malware behavior that is logged (e.g., API calls and their parameters, registry entries, websites and ports accessed, etc.), and can use any raw input from a sandbox to successfully train the deep neural network which is used to generate malware signatures.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"61 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79602876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 194
Si elegans: Hardware architecture and communications protocol 硬件架构和通信协议
Pub Date : 2015-07-12 DOI: 10.1109/IJCNN.2015.7280771
Pedro Machado, Kofi Appiah, T. McGinnity, J. Wade
The hardware layer of the Si elegans EU FP7 project is a massively parallel architecture designed to accurately emulate the C. elegans nematode in biological real-time. The C. elegans nematode is one of the simplest and well characterized Biological Nervous Systems (BNS) yet many questions related to basic functions such as movement and learning remain unanswered. The hardware layer includes a Hardware Neural Network (HNN) composed of 302 FPGAs (one per neuron), a Hardware Muscle Network (HMN) composed of 27 FPGAs (one per 5 muscles) and one Interface Manager FPGA, which is physically connected through 2 Local Area Networks (LANs) and through an innovative 3D optical connectome. Neuron structures (gap junctions and synapses) and muscles are modelled in the design environment of the software layer and their simulation data (spikes, variable values and parameters) generate data packets sent across the Local Area Networks (LAN). Furthermore, a software layer gives the user a set of design tools giving the required flexibility and high level hardware abstraction to design custom neuronal models. In this paper the authors present an overview of the hardware layer, connections infrastructure and communication protocol.
秀丽隐杆线虫EU FP7项目的硬件层是一个大规模并行架构,旨在精确模拟秀丽隐杆线虫的生物实时。秀丽隐杆线虫是最简单的生物神经系统之一,但许多与运动和学习等基本功能相关的问题仍未得到解答。硬件层包括一个由302个FPGA(每个神经元一个)组成的硬件神经网络(HNN),一个由27个FPGA(每5个肌肉一个)组成的硬件肌肉网络(HMN)和一个接口管理器FPGA,它通过2个局域网(lan)和一个创新的3D光学连接体物理连接。神经元结构(间隙连接和突触)和肌肉在软件层的设计环境中建模,它们的模拟数据(尖峰、变量值和参数)生成数据包,通过局域网(LAN)发送。此外,软件层为用户提供了一组设计工具,为设计自定义神经元模型提供了所需的灵活性和高级硬件抽象。在本文中,作者概述了硬件层、连接基础结构和通信协议。
{"title":"Si elegans: Hardware architecture and communications protocol","authors":"Pedro Machado, Kofi Appiah, T. McGinnity, J. Wade","doi":"10.1109/IJCNN.2015.7280771","DOIUrl":"https://doi.org/10.1109/IJCNN.2015.7280771","url":null,"abstract":"The hardware layer of the Si elegans EU FP7 project is a massively parallel architecture designed to accurately emulate the C. elegans nematode in biological real-time. The C. elegans nematode is one of the simplest and well characterized Biological Nervous Systems (BNS) yet many questions related to basic functions such as movement and learning remain unanswered. The hardware layer includes a Hardware Neural Network (HNN) composed of 302 FPGAs (one per neuron), a Hardware Muscle Network (HMN) composed of 27 FPGAs (one per 5 muscles) and one Interface Manager FPGA, which is physically connected through 2 Local Area Networks (LANs) and through an innovative 3D optical connectome. Neuron structures (gap junctions and synapses) and muscles are modelled in the design environment of the software layer and their simulation data (spikes, variable values and parameters) generate data packets sent across the Local Area Networks (LAN). Furthermore, a software layer gives the user a set of design tools giving the required flexibility and high level hardware abstraction to design custom neuronal models. In this paper the authors present an overview of the hardware layer, connections infrastructure and communication protocol.","PeriodicalId":6539,"journal":{"name":"2015 International Joint Conference on Neural Networks (IJCNN)","volume":"27 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2015-07-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83389000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2015 International Joint Conference on Neural Networks (IJCNN)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1