首页 > 最新文献

2022 International Joint Conference on Neural Networks (IJCNN)最新文献

英文 中文
Robust Cross-Modal Retrieval by Adversarial Training 基于对抗训练的鲁棒跨模态检索
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892637
Tao Zhang, Shiliang Sun, Jing Zhao
Cross-modal retrieval is usually implemented based on cross-modal representation learning, which is used to extract semantic information from cross-modal data. Recent work shows that cross-modal representation learning is vulnerable to adversarial attacks, even using large-scale pre-trained networks. By attacking the representation, it can be simple to attack the downstream tasks, especially for cross-modal retrieval tasks. Adversarial attacks on any modality will easily lead to obvious retrieval errors, which brings the challenge to improve the adversarial robustness of cross-modal retrieval. In this paper, we propose a robust cross-modal retrieval method (RoCMR), which generates adversarial examples for both the query modality and candidate modality and performs adversarial training for cross-modal retrieval. Specifically, we generate adversarial examples for both image and text modalities and train the model with benign and adversarial examples in the framework of contrastive learning. We evaluate the proposed RoCMR on two datasets and show its effectiveness in defending against gradient-based attacks.
跨模态检索通常基于跨模态表示学习来实现,该学习用于从跨模态数据中提取语义信息。最近的研究表明,即使使用大规模预训练的网络,跨模态表示学习也容易受到对抗性攻击。通过攻击表示,可以很容易地攻击下游任务,特别是跨模态检索任务。针对任何模态的对抗性攻击都容易导致明显的检索错误,这给提高跨模态检索的对抗性鲁棒性带来了挑战。在本文中,我们提出了一种鲁棒跨模态检索方法(RoCMR),该方法为查询模态和候选模态生成对抗性示例,并对跨模态检索进行对抗性训练。具体来说,我们为图像和文本模式生成对抗示例,并在对比学习的框架中使用良性和对抗示例训练模型。我们在两个数据集上评估了所提出的RoCMR,并展示了它在防御基于梯度的攻击方面的有效性。
{"title":"Robust Cross-Modal Retrieval by Adversarial Training","authors":"Tao Zhang, Shiliang Sun, Jing Zhao","doi":"10.1109/IJCNN55064.2022.9892637","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892637","url":null,"abstract":"Cross-modal retrieval is usually implemented based on cross-modal representation learning, which is used to extract semantic information from cross-modal data. Recent work shows that cross-modal representation learning is vulnerable to adversarial attacks, even using large-scale pre-trained networks. By attacking the representation, it can be simple to attack the downstream tasks, especially for cross-modal retrieval tasks. Adversarial attacks on any modality will easily lead to obvious retrieval errors, which brings the challenge to improve the adversarial robustness of cross-modal retrieval. In this paper, we propose a robust cross-modal retrieval method (RoCMR), which generates adversarial examples for both the query modality and candidate modality and performs adversarial training for cross-modal retrieval. Specifically, we generate adversarial examples for both image and text modalities and train the model with benign and adversarial examples in the framework of contrastive learning. We evaluate the proposed RoCMR on two datasets and show its effectiveness in defending against gradient-based attacks.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134460170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Exploring Attribute Space with Word Embedding for Zero-shot Learning 基于词嵌入的零学习属性空间探索
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892132
Zhaocheng Zhang, Gang Yang
With the purpose of addressing the scarcity of attribute diversity in Zero-shot Learning (ZSL), we propose to search for additional attributes in embedding space to extend the class embedding, providing a more discriminative representation of the class prototype. Meanwhile, to tackle the inherent noise behind manually annotated attributes, we apply multi-layer convolutional processing on semantic features rather than conventional linear transformation for filtering. Moreover, we employ Center Loss to assist the training stage, which helps the learned mapping be more accurate and consistent with the corresponding class's prototype. Combining these modules mentioned above, extensive experiments on several public datasets show that our method could yield decent improvements. This proposed way of extending attributes can also be migrated to other models or tasks and obtain better results.
为了解决零射击学习(Zero-shot Learning, ZSL)中属性多样性的稀缺性,我们提出在嵌入空间中寻找额外的属性来扩展类嵌入,提供一个更具判别性的类原型表示。同时,为了解决手工标注属性背后的固有噪声,我们对语义特征进行多层卷积处理,而不是传统的线性变换进行滤波。此外,我们使用中心损失来辅助训练阶段,这有助于学习到的映射更加准确,并与相应的类原型保持一致。结合上面提到的这些模块,在几个公共数据集上进行的大量实验表明,我们的方法可以产生不错的改进。这种扩展属性的方法也可以移植到其他模型或任务中,获得更好的结果。
{"title":"Exploring Attribute Space with Word Embedding for Zero-shot Learning","authors":"Zhaocheng Zhang, Gang Yang","doi":"10.1109/IJCNN55064.2022.9892132","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892132","url":null,"abstract":"With the purpose of addressing the scarcity of attribute diversity in Zero-shot Learning (ZSL), we propose to search for additional attributes in embedding space to extend the class embedding, providing a more discriminative representation of the class prototype. Meanwhile, to tackle the inherent noise behind manually annotated attributes, we apply multi-layer convolutional processing on semantic features rather than conventional linear transformation for filtering. Moreover, we employ Center Loss to assist the training stage, which helps the learned mapping be more accurate and consistent with the corresponding class's prototype. Combining these modules mentioned above, extensive experiments on several public datasets show that our method could yield decent improvements. This proposed way of extending attributes can also be migrated to other models or tasks and obtain better results.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134490477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Context-Dependent Spatial Representations in the Hippocampus using Place Cell Dendritic Computation 使用位置细胞树突计算的海马体中上下文相关的空间表征
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892401
Adedapo Alabi, D. Vanderelst, A. Minai
The hippocampus in rodents encodes physical space using place cells that show maximal firing in specific regions of space - their place fields. These place cells are reused across different contexts and environments with uncorrelated place fields. Though place fields are known to depend on distal sensory cues, even identical environments can have completely different place fields if the contexts are different. We propose a novel place cell network model for this feature using two frequently overlooked aspects of neural computation - dendritic morphology and the spatial co-location of spatiotemporally co-active afferent synapses - and show that these enable the reuse of place cells to encode different maps for environments with identical sensory cues.
啮齿类动物的海马体利用位置细胞对物理空间进行编码,这些位置细胞在空间的特定区域——它们的位置场——表现出最大的放电。这些位置细胞在不同的上下文和环境中使用不相关的位置字段进行重用。虽然已知位置场依赖于远端感官线索,但如果背景不同,即使是相同的环境也可能有完全不同的位置场。我们提出了一种新的位置细胞网络模型,利用神经计算的两个经常被忽视的方面-树突形态和时空协同活动传入突触的空间共定位-并表明这些可以重用位置细胞来编码具有相同感觉线索的环境的不同地图。
{"title":"Context-Dependent Spatial Representations in the Hippocampus using Place Cell Dendritic Computation","authors":"Adedapo Alabi, D. Vanderelst, A. Minai","doi":"10.1109/IJCNN55064.2022.9892401","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892401","url":null,"abstract":"The hippocampus in rodents encodes physical space using place cells that show maximal firing in specific regions of space - their place fields. These place cells are reused across different contexts and environments with uncorrelated place fields. Though place fields are known to depend on distal sensory cues, even identical environments can have completely different place fields if the contexts are different. We propose a novel place cell network model for this feature using two frequently overlooked aspects of neural computation - dendritic morphology and the spatial co-location of spatiotemporally co-active afferent synapses - and show that these enable the reuse of place cells to encode different maps for environments with identical sensory cues.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131675731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Adaptive Spatial-Temporal Fusion Graph Convolutional Networks for Traffic Flow Forecasting 基于自适应时空融合图卷积网络的交通流预测
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892326
Senwen Li, Liang Ge, Yongquan Lin, Bo Zeng
Traffic flow forecasting is a significant issue in the field of transportation. Early works model temporal dependencies and spatial correlations, respectively. Recently, some models are proposed to capture spatial-temporal dependencies simultaneously. However, these models have three defects. Firstly, they only use the information of road network structure to construct graph structure. It may not accurately reflect the spatial-temporal correlations among nodes. Secondly, only the correlations among nodes adjacent in time or space are considered in each graph convolutional layer. Finally, it's challenging for them to describe that future traffic flow is influenced by different scale spatial-temporal information. In this paper, we propose a model called Adaptive Spatial-Temporal Fusion Graph Convolutional Networks to address these problems. Firstly, the model can find cross-time, cross-space correlations among nodes to adjust spatial-temporal graph structure by a learnable adaptive matrix. Secondly, it can help nodes attain a larger spatiotemporal receptive field through constructing spatial-temporal graphs of different time spans. At last, the results of various spatial-temporal scale graph convolutional layers are fused to produce node embedding for prediction. It helps find the different spatial-temporal ranges' influence for various nodes. Experiments are conducted on real-world traffic datasets, and results show that our model outperforms the state-of-the-art baselines.
交通流预测是交通领域的一个重要问题。早期的作品分别模拟了时间依赖性和空间相关性。近年来,提出了一些同时捕获时空依赖关系的模型。然而,这些模型有三个缺陷。首先,他们只使用路网结构信息来构建图结构。它可能不能准确地反映节点之间的时空相关性。其次,每个图卷积层只考虑在时间或空间上相邻的节点之间的相关性。最后,如何描述未来交通流受不同尺度时空信息的影响是一个挑战。在本文中,我们提出了一种称为自适应时空融合图卷积网络的模型来解决这些问题。首先,该模型通过可学习的自适应矩阵找到节点间的跨时间、跨空间相关性,调整时空图结构;其次,通过构建不同时间跨度的时空图,帮助节点获得更大的时空接受场;最后,对各时空尺度图卷积层的结果进行融合,生成节点嵌入进行预测。它有助于发现不同时空范围对各个节点的影响。在真实世界的交通数据集上进行了实验,结果表明我们的模型优于最先进的基线。
{"title":"Adaptive Spatial-Temporal Fusion Graph Convolutional Networks for Traffic Flow Forecasting","authors":"Senwen Li, Liang Ge, Yongquan Lin, Bo Zeng","doi":"10.1109/IJCNN55064.2022.9892326","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892326","url":null,"abstract":"Traffic flow forecasting is a significant issue in the field of transportation. Early works model temporal dependencies and spatial correlations, respectively. Recently, some models are proposed to capture spatial-temporal dependencies simultaneously. However, these models have three defects. Firstly, they only use the information of road network structure to construct graph structure. It may not accurately reflect the spatial-temporal correlations among nodes. Secondly, only the correlations among nodes adjacent in time or space are considered in each graph convolutional layer. Finally, it's challenging for them to describe that future traffic flow is influenced by different scale spatial-temporal information. In this paper, we propose a model called Adaptive Spatial-Temporal Fusion Graph Convolutional Networks to address these problems. Firstly, the model can find cross-time, cross-space correlations among nodes to adjust spatial-temporal graph structure by a learnable adaptive matrix. Secondly, it can help nodes attain a larger spatiotemporal receptive field through constructing spatial-temporal graphs of different time spans. At last, the results of various spatial-temporal scale graph convolutional layers are fused to produce node embedding for prediction. It helps find the different spatial-temporal ranges' influence for various nodes. Experiments are conducted on real-world traffic datasets, and results show that our model outperforms the state-of-the-art baselines.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131675966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Pixel Rows and Columns Relationship Modeling Network based on Transformer for Retinal Vessel Segmentation 基于Transformer的视网膜血管分割像素行列关系建模网络
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892650
Zekang Qiu, J. Zhao, Chudong Shan, Jianyong Huang, Zhiyong Yuan
Performing automatic retinal vessel segmentation on fundus image can obtain clear retinal vessel structure quickly, which will assist doctors to improve the efficiency and reliability of diagnosis. In fundus image, there are many small vessels and some areas with low contrast, and there may be abnormal areas. Therefore, achieving automatic retinal vessel segmentation with high performance is still challenging. The retinal vessel in the image is a topological structure, so the distribution of retinal vessel pixels in each pixel row (or column) should have some relationship to other rows (or columns). Motivated by this observation, we propose Pixel Rows and Columns Relationship Modeling Network (PRCRM-Net) to achieve high-performance retinal vessel segmentation. PRCRM-Net separately models the relationship between different pixel rows and pixel columns of fundus image, and achieves retinal vessel segmentation by classifying the pixels in units of pixel row and pixel column. The input of PRCRM-Net is the feature map extracted by U-Net. PRCRM-Net firstly processes the input feature map into row feature sequence and column feature sequence respectively. Secondly, it models the relationship between the elements in the row feature sequence and column feature sequence respectively based on Transformer. Finally, the updated row feature sequence and column feature sequence are used to obtain row-based segmentation result and column-based segmentation result respectively. And the final segmentation result is the combination of these two types of results. To evaluate the performance of PRCRM-Net, we conduct comprehensive experiments on three representative datasets, DRIVE, STARE and CHASE_DB1. The experiment results show that the proposed PRCRM-Net achieves state-of-the-art performance.
对眼底图像进行视网膜血管自动分割,可以快速获得清晰的视网膜血管结构,有助于医生提高诊断的效率和可靠性。眼底图像可见小血管较多,部分对比度较低的区域,可能有异常区域。因此,实现高性能的视网膜血管自动分割仍然是一个挑战。图像中的视网膜血管是一种拓扑结构,因此视网膜血管像素在每个像素行(或列)中的分布应该与其他行(或列)有一定的关系。基于这一观察结果,我们提出了像素行和列关系建模网络(PRCRM-Net)来实现高性能的视网膜血管分割。prcr - net分别对眼底图像不同像素行和像素列之间的关系进行建模,以像素行和像素列为单位对像素进行分类,实现视网膜血管分割。prcr - net的输入是U-Net提取的特征图。prcr - net首先将输入的特征映射分别处理成行特征序列和列特征序列。其次,基于Transformer分别对行特征序列和列特征序列中元素之间的关系进行建模;最后,利用更新后的行特征序列和列特征序列分别获得基于行和基于列的分割结果。最后的分割结果就是这两种结果的结合。为了评估PRCRM-Net的性能,我们在DRIVE、STARE和CHASE_DB1三个具有代表性的数据集上进行了综合实验。实验结果表明,所提出的PRCRM-Net达到了最先进的性能。
{"title":"Pixel Rows and Columns Relationship Modeling Network based on Transformer for Retinal Vessel Segmentation","authors":"Zekang Qiu, J. Zhao, Chudong Shan, Jianyong Huang, Zhiyong Yuan","doi":"10.1109/IJCNN55064.2022.9892650","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892650","url":null,"abstract":"Performing automatic retinal vessel segmentation on fundus image can obtain clear retinal vessel structure quickly, which will assist doctors to improve the efficiency and reliability of diagnosis. In fundus image, there are many small vessels and some areas with low contrast, and there may be abnormal areas. Therefore, achieving automatic retinal vessel segmentation with high performance is still challenging. The retinal vessel in the image is a topological structure, so the distribution of retinal vessel pixels in each pixel row (or column) should have some relationship to other rows (or columns). Motivated by this observation, we propose Pixel Rows and Columns Relationship Modeling Network (PRCRM-Net) to achieve high-performance retinal vessel segmentation. PRCRM-Net separately models the relationship between different pixel rows and pixel columns of fundus image, and achieves retinal vessel segmentation by classifying the pixels in units of pixel row and pixel column. The input of PRCRM-Net is the feature map extracted by U-Net. PRCRM-Net firstly processes the input feature map into row feature sequence and column feature sequence respectively. Secondly, it models the relationship between the elements in the row feature sequence and column feature sequence respectively based on Transformer. Finally, the updated row feature sequence and column feature sequence are used to obtain row-based segmentation result and column-based segmentation result respectively. And the final segmentation result is the combination of these two types of results. To evaluate the performance of PRCRM-Net, we conduct comprehensive experiments on three representative datasets, DRIVE, STARE and CHASE_DB1. The experiment results show that the proposed PRCRM-Net achieves state-of-the-art performance.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124252526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-source Representation Enhancement for Wikipedia-style Entity Annotation 维基百科式实体标注的多源表示增强
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892289
Kunyuan Pang, Shasha Li, Jintao Tang, Ting Wang
Entity annotation in Wikipedia (officially named wikilinks) greatly benefits human end-users. Human editors are required to select all mentions that are most helpful to human end-users and link each mention to a Wikipedia page. We aim to design an automatic system to generate Wikipedia-style entity annotation for any plain text. However, existing research either rely heavily on mention-entity map or are restricted to named entities only. Besides, they neglect to select the appropriate mentions as Wikipedia requires. As a result, they leave out some necessary annotation and introduce excessive distracting annotation. Existing benchmarks also skirt around the coverage and selection issues. We propose a new task called Mention Detection and Se-lection for entity annotation, along with a new benchmark, WikiC, to better reflect annotation quality. The task is coined centering mentions specific to each position in high-quality human-annotated examples. We also proposed a new framework, DrWiki, to fulfill the task. We adopt a deep pre-trained span selection model inferring directly from plain text via tokens' context embedding. It can cover all possible spans and avoid limiting to mention-entity maps. In addition, information of both inarguable mention-entity pairs, and mention repeat has been introduced as token-wise representation enhancement by FLAT attention and repeat embedding respectively. Empirical results on WikiC show that, compared with often adopted and state-of-the-art Entity Linking and Entity Recognition methods, our method achieves improvement to previous methods in overall performance. Additional experiments show that DrWiki gains improvement even with a low-coverage mention-entity map.
维基百科中的实体注释(官方命名为wikilinks)极大地造福了人类最终用户。人类编辑需要选择所有对人类最终用户最有帮助的提及,并将每个提及链接到维基百科页面。我们的目标是设计一个自动系统,为任何纯文本生成维基百科风格的实体注释。然而,现有的研究要么严重依赖于提及实体图,要么仅限于命名实体。此外,他们忽略了按照维基百科的要求选择适当的提及。因此,他们省略了一些必要的注释,并引入了过多的分散注意力的注释。现有的基准也绕过了覆盖范围和选择问题。为了更好地反映标注质量,我们提出了一个名为提及检测和选择的实体标注任务,以及一个新的基准WikiC。该任务是在高质量的人工注释示例中对特定于每个位置的提及进行集中。我们还提出了一个新的框架,DrWiki,来完成这个任务。我们采用深度预训练的跨度选择模型,通过标记的上下文嵌入直接从纯文本推断。它可以覆盖所有可能的跨度,避免局限于提及实体映射。此外,引入了无可争议的提及实体对信息和提及重复信息,分别通过FLAT关注和重复嵌入作为标记智能表示增强。WikiC上的实证结果表明,与常用的实体链接和实体识别方法相比,我们的方法在整体性能上比以前的方法有所提高。额外的实验表明,即使使用低覆盖率的提及实体图,DrWiki也能获得改进。
{"title":"Multi-source Representation Enhancement for Wikipedia-style Entity Annotation","authors":"Kunyuan Pang, Shasha Li, Jintao Tang, Ting Wang","doi":"10.1109/IJCNN55064.2022.9892289","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892289","url":null,"abstract":"Entity annotation in Wikipedia (officially named wikilinks) greatly benefits human end-users. Human editors are required to select all mentions that are most helpful to human end-users and link each mention to a Wikipedia page. We aim to design an automatic system to generate Wikipedia-style entity annotation for any plain text. However, existing research either rely heavily on mention-entity map or are restricted to named entities only. Besides, they neglect to select the appropriate mentions as Wikipedia requires. As a result, they leave out some necessary annotation and introduce excessive distracting annotation. Existing benchmarks also skirt around the coverage and selection issues. We propose a new task called Mention Detection and Se-lection for entity annotation, along with a new benchmark, WikiC, to better reflect annotation quality. The task is coined centering mentions specific to each position in high-quality human-annotated examples. We also proposed a new framework, DrWiki, to fulfill the task. We adopt a deep pre-trained span selection model inferring directly from plain text via tokens' context embedding. It can cover all possible spans and avoid limiting to mention-entity maps. In addition, information of both inarguable mention-entity pairs, and mention repeat has been introduced as token-wise representation enhancement by FLAT attention and repeat embedding respectively. Empirical results on WikiC show that, compared with often adopted and state-of-the-art Entity Linking and Entity Recognition methods, our method achieves improvement to previous methods in overall performance. Additional experiments show that DrWiki gains improvement even with a low-coverage mention-entity map.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114832372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Comparative Study and a New Industrial Platform for Decentralized Anomaly Detection Using Machine Learning Algorithms 基于机器学习算法的去中心化异常检测的比较研究与新工业平台
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892939
Fabian Gerz, Tolga Renan Bastürk, Julian Kirchhoff, Joachim Denker, L. Al-Shrouf, M. Jelali
The occurrence of anomalies and unexpected, process-related faults is a major problem for manufacturing systems, which has a significant impact on product quality. Early detection of anomalies is therefore of central importance in order to create sufficient room for maneuver to take countermeasures and ensure product quality. This paper investigates the performance of machine learning (ML) algorithms for anomaly detection in sensor data streams. For this purpose, the performance of six ML algorithms (K-means, DBSCAN, Isolation Forest, OCSVM, LSTM-Network, and DeepAnt) is evaluated based on defined performance metrics. These methods are benchmarked on publicly available datasets, own synthetic datasets, and novel industrial datasets. The latter include radar sensor datasets from a hot rolling mill. Research results show a high detection performance of K-means algorithm, DBSCAN algorithm and LSTM network for punctual, collective and contextual anomalies. A decentralized strategy for (real-time) anomaly detection using sensor data streams is proposed and an industrial (Cloud-Edge Computing) platform is developed and implemented for this purpose.
异常和意外的过程相关故障的发生是制造系统的主要问题,对产品质量有重大影响。因此,早期发现异常是至关重要的,以便为采取对策和确保产品质量创造足够的机动空间。本文研究了传感器数据流中异常检测的机器学习(ML)算法的性能。为此,基于定义的性能指标评估了六种机器学习算法(K-means、DBSCAN、隔离森林、OCSVM、LSTM-Network和DeepAnt)的性能。这些方法在公开可用的数据集、自己的合成数据集和新的工业数据集上进行基准测试。后者包括热轧厂的雷达传感器数据集。研究结果表明,K-means算法、DBSCAN算法和LSTM网络对准时异常、集体异常和上下文异常具有较高的检测性能。提出了一种利用传感器数据流进行(实时)异常检测的分散策略,并为此开发和实现了一个工业(云边缘计算)平台。
{"title":"A Comparative Study and a New Industrial Platform for Decentralized Anomaly Detection Using Machine Learning Algorithms","authors":"Fabian Gerz, Tolga Renan Bastürk, Julian Kirchhoff, Joachim Denker, L. Al-Shrouf, M. Jelali","doi":"10.1109/IJCNN55064.2022.9892939","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892939","url":null,"abstract":"The occurrence of anomalies and unexpected, process-related faults is a major problem for manufacturing systems, which has a significant impact on product quality. Early detection of anomalies is therefore of central importance in order to create sufficient room for maneuver to take countermeasures and ensure product quality. This paper investigates the performance of machine learning (ML) algorithms for anomaly detection in sensor data streams. For this purpose, the performance of six ML algorithms (K-means, DBSCAN, Isolation Forest, OCSVM, LSTM-Network, and DeepAnt) is evaluated based on defined performance metrics. These methods are benchmarked on publicly available datasets, own synthetic datasets, and novel industrial datasets. The latter include radar sensor datasets from a hot rolling mill. Research results show a high detection performance of K-means algorithm, DBSCAN algorithm and LSTM network for punctual, collective and contextual anomalies. A decentralized strategy for (real-time) anomaly detection using sensor data streams is proposed and an industrial (Cloud-Edge Computing) platform is developed and implemented for this purpose.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114931615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
POPNASv2: An Efficient Multi-Objective Neural Architecture Search Technique POPNASv2:一种高效的多目标神经结构搜索技术
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892073
Andrea Falanti, Eugenio Lomurno, Stefano Samele, D. Ardagna, Matteo Matteucci
Automating the research for the best neural network model is a task that has gained more and more relevance in the last few years. In this context, Neural Architecture Search (NAS) represents the most effective technique whose results rival the state of the art hand-crafted architectures. However, this approach requires a lot of computational capabilities as well as research time, which make prohibitive its usage in many real-world scenarios. With its sequential model-based optimization strategy, Progressive Neural Architecture Search (PNAS) represents a possible step forward to face this resources issue. Despite the quality of the found network architectures, this technique is still limited in research time. A significant step in this direction has been done by Pareto-Optimal Progressive Neural Architecture Search (POPNAS), which expand PNAS with a time predictor to enable a trade-off between search time and accuracy, considering a multi-objective optimization problem. This paper proposes a new version of the Pareto-Optimal Progressive Neural Architecture Search, called POPNASv2. Our approach enhances its first version and improves its performance. We expanded the search space by adding new operators and improved the quality of both predictors to build more accurate Pareto fronts. Moreover, we introduced cell equivalence checks and enriched the search strategy with an adaptive greedy exploration step. Our efforts allow POPNASv2 to achieve PNAS-like performance with an average 4x factor search time speed-up. Code: https://doi.org/10.5281/zenodo.6574040
对最佳神经网络模型的自动化研究是近年来越来越受到重视的课题。在这种情况下,神经架构搜索(NAS)代表了最有效的技术,其结果可以与最先进的手工架构相媲美。然而,这种方法需要大量的计算能力和研究时间,这使得它在许多现实场景中的使用望而却步。渐进式神经结构搜索(Progressive Neural Architecture Search, PNAS)以其基于序列模型的优化策略为解决这一资源问题提供了可能的方法。尽管已发现的网络体系结构质量很高,但该技术的研究时间仍然有限。帕累托最优渐进式神经结构搜索(POPNAS)在这个方向上迈出了重要的一步,它用时间预测器扩展了PNAS,考虑到多目标优化问题,可以在搜索时间和精度之间进行权衡。本文提出了一个新版本的帕累托最优渐进神经结构搜索,称为POPNASv2。我们的方法增强了它的第一个版本并提高了它的性能。我们通过添加新的运算符扩展了搜索空间,并提高了两个预测器的质量,以构建更准确的帕累托前沿。此外,我们引入了单元等价性检查,并通过自适应贪婪探索步骤丰富了搜索策略。我们的努力使POPNASv2能够以平均4倍的搜索时间加速实现与pnas类似的性能。代码:https://doi.org/10.5281/zenodo.6574040
{"title":"POPNASv2: An Efficient Multi-Objective Neural Architecture Search Technique","authors":"Andrea Falanti, Eugenio Lomurno, Stefano Samele, D. Ardagna, Matteo Matteucci","doi":"10.1109/IJCNN55064.2022.9892073","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892073","url":null,"abstract":"Automating the research for the best neural network model is a task that has gained more and more relevance in the last few years. In this context, Neural Architecture Search (NAS) represents the most effective technique whose results rival the state of the art hand-crafted architectures. However, this approach requires a lot of computational capabilities as well as research time, which make prohibitive its usage in many real-world scenarios. With its sequential model-based optimization strategy, Progressive Neural Architecture Search (PNAS) represents a possible step forward to face this resources issue. Despite the quality of the found network architectures, this technique is still limited in research time. A significant step in this direction has been done by Pareto-Optimal Progressive Neural Architecture Search (POPNAS), which expand PNAS with a time predictor to enable a trade-off between search time and accuracy, considering a multi-objective optimization problem. This paper proposes a new version of the Pareto-Optimal Progressive Neural Architecture Search, called POPNASv2. Our approach enhances its first version and improves its performance. We expanded the search space by adding new operators and improved the quality of both predictors to build more accurate Pareto fronts. Moreover, we introduced cell equivalence checks and enriched the search strategy with an adaptive greedy exploration step. Our efforts allow POPNASv2 to achieve PNAS-like performance with an average 4x factor search time speed-up. Code: https://doi.org/10.5281/zenodo.6574040","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115027760","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Accelerating Motion Perception Model Mimics the Visual Neuronal Ensemble of Crab 加速运动感知模型模拟螃蟹的视觉神经元集合
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892540
Hao Luan, Mu Hua, Jigen Peng, Shigang Yue, Shengyong Chen, Qinbing Fu
In nature, crabs have a panoramic vision for the localization and perception of accelerating motion from local segments to global view in order to guide reactive behaviours including escape. The visual neuronal ensemble in crab plays crucial roles in such capability, however, has never been investigated and modelled as an artificial vision system. To bridge this gap, we propose an accelerating motion perception model (AMPM) mimicking the visual neuronal ensemble in crab. The AMPM includes two main parts, wherein the pre-synaptic network from the previous modelling work simulates 16 MLGI neurons covering the entire view to localize moving objects. The emphasis herein is laid on the original modelling of MLGIs' post-synaptic network to perceive accelerating motions from a global view, which employs a novel spatial-temporal difference encoder (STDE), and an adaptive spiking threshold temporal difference encoder (AT-TDE). Specifically, the STDE transforms “time-to-travel” between activations of two successive segments of MLG1 into excitatory post-synaptic current (EPSC), which decays with the elapse of time. The AT-TDE in two directional, i.e., counter-clockwise and clockwise accelerating detectors guarantees “non-firing” to constant movements. Accordingly, the accelerating motion can be effectively localized and perceived by the whole network. The systematic experiments verified the feasibility and robustness of the proposed method. The model responses to translational accelerating motion also fit many of the explored physiological features of direction selective neurons in the lobula complex of crab (i.e. lobula complex direction cells, LCDCs). This modelling study not only provides a reasonable hypothesis for such biological neural pathways, but is also critical for developing a new neuromorphic sensor strategy.
在自然界中,螃蟹具有从局部到全局加速运动的定位和感知全景视野,以指导包括逃跑在内的反应性行为。然而,螃蟹的视觉神经元集合在这种能力中起着至关重要的作用,但从未作为人工视觉系统进行过研究和建模。为了弥补这一差距,我们提出了一种模拟螃蟹视觉神经元集合的加速运动感知模型(AMPM)。AMPM包括两个主要部分,其中来自先前建模工作的突触前网络模拟了覆盖整个视图的16个MLGI神经元来定位运动物体。本文重点介绍了mlgi突触后网络的原始建模,从全局角度感知加速运动,该模型采用了一种新的时空差分编码器(STDE)和自适应尖峰阈值时间差分编码器(AT-TDE)。具体来说,STDE将MLG1两个连续片段激活之间的“旅行时间”转化为兴奋性突触后电流(EPSC), EPSC随着时间的流逝而衰减。AT-TDE的两个方向,即逆时针和顺时针加速探测器,保证了恒定运动的“不着火”。因此,加速运动可以被整个网络有效地定位和感知。系统实验验证了该方法的可行性和鲁棒性。该模型对平移加速运动的响应也符合螃蟹小叶复合体(即小叶复合体方向细胞,lobula complex direction cells, LCDCs)中方向选择神经元的许多生理特征。该模型研究不仅为这种生物神经通路提供了合理的假设,而且对于开发新的神经形态传感器策略也至关重要。
{"title":"Accelerating Motion Perception Model Mimics the Visual Neuronal Ensemble of Crab","authors":"Hao Luan, Mu Hua, Jigen Peng, Shigang Yue, Shengyong Chen, Qinbing Fu","doi":"10.1109/IJCNN55064.2022.9892540","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892540","url":null,"abstract":"In nature, crabs have a panoramic vision for the localization and perception of accelerating motion from local segments to global view in order to guide reactive behaviours including escape. The visual neuronal ensemble in crab plays crucial roles in such capability, however, has never been investigated and modelled as an artificial vision system. To bridge this gap, we propose an accelerating motion perception model (AMPM) mimicking the visual neuronal ensemble in crab. The AMPM includes two main parts, wherein the pre-synaptic network from the previous modelling work simulates 16 MLGI neurons covering the entire view to localize moving objects. The emphasis herein is laid on the original modelling of MLGIs' post-synaptic network to perceive accelerating motions from a global view, which employs a novel spatial-temporal difference encoder (STDE), and an adaptive spiking threshold temporal difference encoder (AT-TDE). Specifically, the STDE transforms “time-to-travel” between activations of two successive segments of MLG1 into excitatory post-synaptic current (EPSC), which decays with the elapse of time. The AT-TDE in two directional, i.e., counter-clockwise and clockwise accelerating detectors guarantees “non-firing” to constant movements. Accordingly, the accelerating motion can be effectively localized and perceived by the whole network. The systematic experiments verified the feasibility and robustness of the proposed method. The model responses to translational accelerating motion also fit many of the explored physiological features of direction selective neurons in the lobula complex of crab (i.e. lobula complex direction cells, LCDCs). This modelling study not only provides a reasonable hypothesis for such biological neural pathways, but is also critical for developing a new neuromorphic sensor strategy.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117030058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Time-Aware Neighbor Sampling on Temporal Graphs 时间图上的时间感知邻居抽样
Pub Date : 2022-07-18 DOI: 10.1109/IJCNN55064.2022.9892942
Yiwei Wang, Yujun Cai, Yuxuan Liang, Henghui Ding, Changhu Wang, Bryan Hooi
We present a new neighbor sampling method on temporal graphs. In a temporal graph, predicting different nodes' time-varying properties can require the receptive neigh-borhood of various temporal scales. In this work, we propose the TNS (Time-aware Neighbor Sampling) method: TNS learns from temporal information to provide an adaptive receptive neighborhood for every node at any time. Learning how to sample neighbors is non-trivial, since the neighbor indices in time order are discrete and not differentiable. To address this challenge, we transform neighbor indices from discrete values to continuous ones by interpolating the neighbors' messages. TNS can be flexibly incorporated into popular temporal graph networks to improve their effectiveness without increasing their time complexity. TNS can be trained in an end-to-end manner. It requires no extra supervision and is automatically and implicitly guided to sample the neighbors that are most beneficial for prediction. Empirical results on multiple standard datasets show that TNS yields significant gains on edge prediction and node classification.
提出了一种新的时间图邻域采样方法。在时间图中,预测不同节点的时变特性可能需要不同时间尺度的接受邻域。在这项工作中,我们提出了TNS(时间感知邻居采样)方法:TNS从时间信息中学习,在任何时候为每个节点提供自适应的接受邻居。学习如何采样邻居是不平凡的,因为邻居指数在时间顺序上是离散的,不可微的。为了解决这个问题,我们通过插值邻居的消息将邻居索引从离散值转换为连续值。在不增加时间复杂度的情况下,TNS可以灵活地加入到流行的时态图网络中,以提高其有效性。TNS可以以端到端方式进行训练。它不需要额外的监督,并且自动和隐含地引导对最有利于预测的邻居进行采样。在多个标准数据集上的实验结果表明,TNS在边缘预测和节点分类方面取得了显著的进步。
{"title":"Time-Aware Neighbor Sampling on Temporal Graphs","authors":"Yiwei Wang, Yujun Cai, Yuxuan Liang, Henghui Ding, Changhu Wang, Bryan Hooi","doi":"10.1109/IJCNN55064.2022.9892942","DOIUrl":"https://doi.org/10.1109/IJCNN55064.2022.9892942","url":null,"abstract":"We present a new neighbor sampling method on temporal graphs. In a temporal graph, predicting different nodes' time-varying properties can require the receptive neigh-borhood of various temporal scales. In this work, we propose the TNS (Time-aware Neighbor Sampling) method: TNS learns from temporal information to provide an adaptive receptive neighborhood for every node at any time. Learning how to sample neighbors is non-trivial, since the neighbor indices in time order are discrete and not differentiable. To address this challenge, we transform neighbor indices from discrete values to continuous ones by interpolating the neighbors' messages. TNS can be flexibly incorporated into popular temporal graph networks to improve their effectiveness without increasing their time complexity. TNS can be trained in an end-to-end manner. It requires no extra supervision and is automatically and implicitly guided to sample the neighbors that are most beneficial for prediction. Empirical results on multiple standard datasets show that TNS yields significant gains on edge prediction and node classification.","PeriodicalId":106974,"journal":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117043023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2022 International Joint Conference on Neural Networks (IJCNN)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1