首页 > 最新文献

Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)最新文献

英文 中文
TFCNs: A CNN-Transformer Hybrid Network for Medical Image Segmentation TFCNs:用于医学图像分割的CNN-Transformer混合网络
Zihan Li, Dihan Li, Cangbai Xu, Wei-Chien Wang, Qingqi Hong, Qingde Li, Jie Tian
Medical image segmentation is one of the most fundamental tasks concerning medical information analysis. Various solutions have been proposed so far, including many deep learning-based techniques, such as U-Net, FC-DenseNet, etc. However, high-precision medical image segmentation remains a highly challenging task due to the existence of inherent magnification and distortion in medical images as well as the presence of lesions with similar density to normal tissues. In this paper, we propose TFCNs (Transformers for Fully Convolutional denseNets) to tackle the problem by introducing ResLinear-Transformer (RL-Transformer) and Convolutional Linear Attention Block (CLAB) to FC-DenseNet. TFCNs is not only able to utilize more latent information from the CT images for feature extraction, but also can capture and disseminate semantic features and filter non-semantic features more effectively through the CLAB module. Our experimental results show that TFCNs can achieve state-of-the-art performance with dice scores of 83.72% on the Synapse dataset. In addition, we evaluate the robustness of TFCNs for lesion area effects on the COVID-19 public datasets. The Python code will be made publicly available on https://github.com/HUANGLIZI/TFCNs. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
医学图像分割是医学信息分析中最基本的任务之一。到目前为止,已经提出了各种解决方案,包括许多基于深度学习的技术,如U-Net, FC-DenseNet等。然而,由于医学图像存在固有的放大和畸变,以及存在与正常组织密度相似的病变,高精度医学图像分割仍然是一项极具挑战性的任务。在本文中,我们提出了TFCNs (transformer for Fully Convolutional densenet),通过在FC-DenseNet中引入ResLinear-Transformer (RL-Transformer)和Convolutional Linear Attention Block (CLAB)来解决这个问题。TFCNs不仅可以利用CT图像中更多的潜在信息进行特征提取,还可以通过CLAB模块更有效地捕获和传播语义特征,过滤非语义特征。我们的实验结果表明,TFCNs在Synapse数据集上的骰子得分为83.72%,可以达到最先进的性能。此外,我们评估了tfns在COVID-19公共数据集上对病变区域效应的鲁棒性。Python代码将在https://github.com/HUANGLIZI/TFCNs上公开提供。©2022,作者获得施普林格自然瑞士股份有限公司的独家授权。
{"title":"TFCNs: A CNN-Transformer Hybrid Network for Medical Image Segmentation","authors":"Zihan Li, Dihan Li, Cangbai Xu, Wei-Chien Wang, Qingqi Hong, Qingde Li, Jie Tian","doi":"10.48550/arXiv.2207.03450","DOIUrl":"https://doi.org/10.48550/arXiv.2207.03450","url":null,"abstract":"Medical image segmentation is one of the most fundamental tasks concerning medical information analysis. Various solutions have been proposed so far, including many deep learning-based techniques, such as U-Net, FC-DenseNet, etc. However, high-precision medical image segmentation remains a highly challenging task due to the existence of inherent magnification and distortion in medical images as well as the presence of lesions with similar density to normal tissues. In this paper, we propose TFCNs (Transformers for Fully Convolutional denseNets) to tackle the problem by introducing ResLinear-Transformer (RL-Transformer) and Convolutional Linear Attention Block (CLAB) to FC-DenseNet. TFCNs is not only able to utilize more latent information from the CT images for feature extraction, but also can capture and disseminate semantic features and filter non-semantic features more effectively through the CLAB module. Our experimental results show that TFCNs can achieve state-of-the-art performance with dice scores of 83.72% on the Synapse dataset. In addition, we evaluate the robustness of TFCNs for lesion area effects on the COVID-19 public datasets. The Python code will be made publicly available on https://github.com/HUANGLIZI/TFCNs. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"66 1","pages":"781-792"},"PeriodicalIF":0.0,"publicationDate":"2022-07-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81843600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Attention Guided Network for Salient Object Detection in Optical Remote Sensing Images 光学遥感图像中显著目标检测的注意力引导网络
Yuhan Lin, Han Sun, Ningzhong Liu, Yetong Bian, Jun Cen, Huiyu Zhou
Due to the extreme complexity of scale and shape as well as the uncertainty of the predicted location, salient object detection in optical remote sensing images (RSI-SOD) is a very difficult task. The existing SOD methods can satisfy the detection performance for natural scene images, but they are not well adapted to RSI-SOD due to the above-mentioned image characteristics in remote sensing images. In this paper, we propose a novel Attention Guided Network (AGNet) for SOD in optical RSIs, including position enhancement stage and detail refinement stage. Specifically, the position enhancement stage consists of a semantic attention module and a contextual attention module to accurately describe the approximate location of salient objects. The detail refinement stage uses the proposed self-refinement module to progressively refine the predicted results under the guidance of attention and reverse attention. In addition, the hybrid loss is applied to supervise the training of the network, which can improve the performance of the model from three perspectives of pixel, region and statistics. Extensive experiments on two popular benchmarks demonstrate that AGNet achieves competitive performance compared to other state-of-the-art methods. The code will be available at https://github.com/NuaaYH/AGNet.
由于尺度和形状的极端复杂性以及预测位置的不确定性,光学遥感图像中的显著目标检测(RSI-SOD)是一项非常困难的任务。现有的超氧化物歧化酶方法可以满足对自然场景图像的检测性能,但由于遥感图像的上述图像特征,对RSI-SOD的适应性不强。在本文中,我们提出了一种新的注意引导网络(AGNet),用于光学rsi中SOD,包括位置增强阶段和细节细化阶段。具体而言,位置增强阶段包括语义注意模块和上下文注意模块,以准确描述显著物体的大致位置。细节细化阶段使用提出的自细化模块,在注意和反向注意的指导下,逐步细化预测结果。此外,利用混合损失来监督网络的训练,可以从像素、区域和统计三个角度提高模型的性能。在两个流行的基准测试上进行的大量实验表明,与其他最先进的方法相比,AGNet实现了具有竞争力的性能。代码可在https://github.com/NuaaYH/AGNet上获得。
{"title":"Attention Guided Network for Salient Object Detection in Optical Remote Sensing Images","authors":"Yuhan Lin, Han Sun, Ningzhong Liu, Yetong Bian, Jun Cen, Huiyu Zhou","doi":"10.48550/arXiv.2207.01755","DOIUrl":"https://doi.org/10.48550/arXiv.2207.01755","url":null,"abstract":"Due to the extreme complexity of scale and shape as well as the uncertainty of the predicted location, salient object detection in optical remote sensing images (RSI-SOD) is a very difficult task. The existing SOD methods can satisfy the detection performance for natural scene images, but they are not well adapted to RSI-SOD due to the above-mentioned image characteristics in remote sensing images. In this paper, we propose a novel Attention Guided Network (AGNet) for SOD in optical RSIs, including position enhancement stage and detail refinement stage. Specifically, the position enhancement stage consists of a semantic attention module and a contextual attention module to accurately describe the approximate location of salient objects. The detail refinement stage uses the proposed self-refinement module to progressively refine the predicted results under the guidance of attention and reverse attention. In addition, the hybrid loss is applied to supervise the training of the network, which can improve the performance of the model from three perspectives of pixel, region and statistics. Extensive experiments on two popular benchmarks demonstrate that AGNet achieves competitive performance compared to other state-of-the-art methods. The code will be available at https://github.com/NuaaYH/AGNet.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"99 1","pages":"25-36"},"PeriodicalIF":0.0,"publicationDate":"2022-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81279505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Multi scale Feature Extraction and Fusion for Online Knowledge Distillation 在线知识蒸馏的多尺度特征提取与融合
Panpan Zou, Yinglei Teng, Tao Niu
Online knowledge distillation conducts knowledge transfer among all student models to alleviate the reliance on pre-trained models. However, existing online methods rely heavily on the prediction distributions and neglect the further exploration of the representational knowledge. In this paper, we propose a novel Multi-scale Feature Extraction and Fusion method (MFEF) for online knowledge distillation, which comprises three key components: Multi-scale Feature Extraction, Dual-attention and Feature Fusion, towards generating more informative feature maps for distillation. The multiscale feature extraction exploiting divide-and-concatenate in channel dimension is proposed to improve the multi-scale representation ability of feature maps. To obtain more accurate information, we design a dual-attention to strengthen the important channel and spatial regions adaptively. Moreover, we aggregate and fuse the former processed feature maps via feature fusion to assist the training of student models. Extensive experiments on CIF AR-10, CIF AR-100, and CINIC-10 show that MFEF transfers more beneficial representational knowledge for distillation and outperforms alternative methods among various network architectures
在线知识蒸馏在所有学生模型之间进行知识转移,以减轻对预训练模型的依赖。然而,现有的在线方法严重依赖于预测分布,忽视了对表征知识的进一步探索。本文提出了一种用于在线知识蒸馏的多尺度特征提取与融合方法(MFEF),该方法包括三个关键部分:多尺度特征提取、双关注和特征融合,以生成更多信息的特征映射用于蒸馏。为了提高特征映射的多尺度表示能力,提出了一种利用通道维度上的分拼接的多尺度特征提取方法。为了获得更准确的信息,我们设计了双关注自适应增强重要通道和空间区域。此外,我们通过特征融合对之前处理过的特征图进行聚合和融合,以辅助学生模型的训练。在CIF AR-10、CIF AR-100和CINIC-10上进行的大量实验表明,MFEF为蒸馏传递了更多有益的表征知识,并且在各种网络架构中优于其他方法
{"title":"Multi scale Feature Extraction and Fusion for Online Knowledge Distillation","authors":"Panpan Zou, Yinglei Teng, Tao Niu","doi":"10.48550/arXiv.2206.08224","DOIUrl":"https://doi.org/10.48550/arXiv.2206.08224","url":null,"abstract":"Online knowledge distillation conducts knowledge transfer among all student models to alleviate the reliance on pre-trained models. However, existing online methods rely heavily on the prediction distributions and neglect the further exploration of the representational knowledge. In this paper, we propose a novel Multi-scale Feature Extraction and Fusion method (MFEF) for online knowledge distillation, which comprises three key components: Multi-scale Feature Extraction, Dual-attention and Feature Fusion, towards generating more informative feature maps for distillation. The multiscale feature extraction exploiting divide-and-concatenate in channel dimension is proposed to improve the multi-scale representation ability of feature maps. To obtain more accurate information, we design a dual-attention to strengthen the important channel and spatial regions adaptively. Moreover, we aggregate and fuse the former processed feature maps via feature fusion to assist the training of student models. Extensive experiments on CIF AR-10, CIF AR-100, and CINIC-10 show that MFEF transfers more beneficial representational knowledge for distillation and outperforms alternative methods among various network architectures","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"39 3-4 1","pages":"126-138"},"PeriodicalIF":0.0,"publicationDate":"2022-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77867222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Attention Awareness Multiple Instance Neural Network 注意感知多实例神经网络
Jingjun Yi, Beichen Zhou
Multiple instance learning is qualified for many pattern recognition tasks with weakly annotated data. The combination of artificial neural network and multiple instance learning offers an end-to-end solution and has been widely utilized. However, challenges remain in two-folds. Firstly, current MIL pooling operators are usually pre-defined and lack flexibility to mine key instances. Secondly, in current solutions, the bag-level representation can be inaccurate or inaccessible. To this end, we propose an attention awareness multiple instance neural network framework in this paper. It consists of an instance-level classifier, a trainable MIL pooling operator based on spatial attention and a bag-level classification layer. Exhaustive experiments on a series of pattern recognition tasks demonstrate that our framework outperforms many state-of-the-art MIL methods and val-idates the effectiveness of our proposed attention MIL pooling operators.
多实例学习适用于许多带有弱标注数据的模式识别任务。人工神经网络与多实例学习的结合提供了端到端的解决方案,并得到了广泛的应用。然而,挑战仍然存在于两方面。首先,当前的MIL池操作符通常是预定义的,缺乏挖掘关键实例的灵活性。其次,在当前的解决方案中,包级表示可能不准确或不可访问。为此,本文提出了一种注意力感知多实例神经网络框架。它由实例级分类器、基于空间注意的可训练MIL池化算子和袋级分类层组成。在一系列模式识别任务上的详尽实验表明,我们的框架优于许多最先进的MIL方法,并验证了我们提出的注意力MIL池算子的有效性。
{"title":"Attention Awareness Multiple Instance Neural Network","authors":"Jingjun Yi, Beichen Zhou","doi":"10.48550/arXiv.2205.13750","DOIUrl":"https://doi.org/10.48550/arXiv.2205.13750","url":null,"abstract":"Multiple instance learning is qualified for many pattern recognition tasks with weakly annotated data. The combination of artificial neural network and multiple instance learning offers an end-to-end solution and has been widely utilized. However, challenges remain in two-folds. Firstly, current MIL pooling operators are usually pre-defined and lack flexibility to mine key instances. Secondly, in current solutions, the bag-level representation can be inaccurate or inaccessible. To this end, we propose an attention awareness multiple instance neural network framework in this paper. It consists of an instance-level classifier, a trainable MIL pooling operator based on spatial attention and a bag-level classification layer. Exhaustive experiments on a series of pattern recognition tasks demonstrate that our framework outperforms many state-of-the-art MIL methods and val-idates the effectiveness of our proposed attention MIL pooling operators.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"127 1","pages":"581-592"},"PeriodicalIF":0.0,"publicationDate":"2022-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75815708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A unified view on Self-Organizing Maps (SOMs) and Stochastic Neighbor Embedding (SNE) 关于自组织映射(SOMs)和随机邻居嵌入(SNE)的统一观点
Thibaut Kulak, Anthony Fillion, Franccois Blayo
We propose a unified view on two widely used data visualization techniques: Self-Organizing Maps (SOMs) and Stochastic Neighbor Embedding (SNE). We show that they can both be derived from a common mathematical framework. Leveraging this formulation, we propose to compare SOM and SNE quantitatively on two datasets, and discuss possible avenues for future work to take advantage of both approaches.
我们对两种广泛使用的数据可视化技术:自组织映射(SOMs)和随机邻居嵌入(SNE)提出了统一的观点。我们证明它们都可以从一个共同的数学框架中推导出来。利用这一公式,我们建议在两个数据集上定量地比较SOM和SNE,并讨论利用这两种方法的未来工作的可能途径。
{"title":"A unified view on Self-Organizing Maps (SOMs) and Stochastic Neighbor Embedding (SNE)","authors":"Thibaut Kulak, Anthony Fillion, Franccois Blayo","doi":"10.48550/arXiv.2205.01492","DOIUrl":"https://doi.org/10.48550/arXiv.2205.01492","url":null,"abstract":"We propose a unified view on two widely used data visualization techniques: Self-Organizing Maps (SOMs) and Stochastic Neighbor Embedding (SNE). We show that they can both be derived from a common mathematical framework. Leveraging this formulation, we propose to compare SOM and SNE quantitatively on two datasets, and discuss possible avenues for future work to take advantage of both approaches.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"139 20","pages":"458-468"},"PeriodicalIF":0.0,"publicationDate":"2022-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91403086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Jacobian Ensembles Improve Robustness Trade-offs to Adversarial Attacks 雅可比集合改进对对抗性攻击的鲁棒性权衡
Kenneth T. Co, David Martínez-Rego, Zhongyuan Hau, Emil C. Lupu
Deep neural networks have become an integral part of our software infrastructure and are being deployed in many widely-used and safety-critical applications. However, their integration into many systems also brings with it the vulnerability to test time attacks in the form of Universal Adversarial Perturbations (UAPs). UAPs are a class of perturbations that when applied to any input causes model misclassification. Although there is an ongoing effort to defend models against these adversarial attacks, it is often difficult to reconcile the trade-offs in model accuracy and robustness to adversarial attacks. Jacobian regularization has been shown to improve the robustness of models against UAPs, whilst model ensembles have been widely adopted to improve both predictive performance and model robustness. In this work, we propose a novel approach, Jacobian Ensembles-a combination of Jacobian regularization and model ensembles to significantly increase the robustness against UAPs whilst maintaining or improving model accuracy. Our results show that Jacobian Ensembles achieves previously unseen levels of accuracy and robustness, greatly improving over previous methods that tend to skew towards only either accuracy or robustness.
深度神经网络已经成为我们软件基础设施的一个组成部分,并被部署在许多广泛使用和安全关键应用中。然而,它们与许多系统的集成也带来了以通用对抗性扰动(uap)形式的测试时间攻击的脆弱性。uap是一类扰动,当应用于任何输入时都会导致模型错误分类。尽管人们正在努力保护模型免受这些对抗性攻击,但通常很难在模型准确性和对抗性攻击的鲁棒性之间进行权衡。雅可比正则化已被证明可以提高模型对uap的鲁棒性,而模型集成已被广泛采用以提高预测性能和模型鲁棒性。在这项工作中,我们提出了一种新的方法,雅可比集成-雅可比正则化和模型集成的结合,以显着增加对uap的鲁棒性,同时保持或提高模型精度。我们的研究结果表明,雅可比集成达到了以前从未见过的精度和鲁棒性水平,大大改善了以前的方法,这些方法往往只倾向于准确性或鲁棒性。
{"title":"Jacobian Ensembles Improve Robustness Trade-offs to Adversarial Attacks","authors":"Kenneth T. Co, David Martínez-Rego, Zhongyuan Hau, Emil C. Lupu","doi":"10.48550/arXiv.2204.08726","DOIUrl":"https://doi.org/10.48550/arXiv.2204.08726","url":null,"abstract":"Deep neural networks have become an integral part of our software infrastructure and are being deployed in many widely-used and safety-critical applications. However, their integration into many systems also brings with it the vulnerability to test time attacks in the form of Universal Adversarial Perturbations (UAPs). UAPs are a class of perturbations that when applied to any input causes model misclassification. Although there is an ongoing effort to defend models against these adversarial attacks, it is often difficult to reconcile the trade-offs in model accuracy and robustness to adversarial attacks. Jacobian regularization has been shown to improve the robustness of models against UAPs, whilst model ensembles have been widely adopted to improve both predictive performance and model robustness. In this work, we propose a novel approach, Jacobian Ensembles-a combination of Jacobian regularization and model ensembles to significantly increase the robustness against UAPs whilst maintaining or improving model accuracy. Our results show that Jacobian Ensembles achieves previously unseen levels of accuracy and robustness, greatly improving over previous methods that tend to skew towards only either accuracy or robustness.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"65 1","pages":"680-691"},"PeriodicalIF":0.0,"publicationDate":"2022-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79248862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Stream-based Active Learning with Verification Latency in Non-stationary Environments 非平稳环境下具有验证延迟的基于流的主动学习
Andrea Castellani, Sebastian Schmitt, Barbara Hammer
{"title":"Stream-based Active Learning with Verification Latency in Non-stationary Environments","authors":"Andrea Castellani, Sebastian Schmitt, Barbara Hammer","doi":"10.1007/978-3-031-15937-4_22","DOIUrl":"https://doi.org/10.1007/978-3-031-15937-4_22","url":null,"abstract":"","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"6 2 1","pages":"260-272"},"PeriodicalIF":0.0,"publicationDate":"2022-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90498295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Novel Approach to Train Diverse Types of Language Models for Health Mention Classification of Tweets 一种训练不同类型的推文健康提及分类语言模型的新方法
Pervaiz Iqbal Khan, Imran Razzak, A. Dengel, Sheraz Ahmed
Health mention classification deals with the disease detection in a given text containing disease words. However, non-health and figurative use of disease words adds challenges to the task. Recently, adversarial training acting as a means of regularization has gained popularity in many NLP tasks. In this paper, we propose a novel approach to train language models for health mention classification of tweets that involves adversarial training. We generate adversarial examples by adding perturbation to the representations of transformer models for tweet examples at various levels using Gaussian noise. Further, we employ contrastive loss as an additional objective function. We evaluate the proposed method on the PHM2017 dataset extended version. Results show that our proposed approach improves the performance of classifier significantly over the baseline methods. Moreover, our analysis shows that adding noise at earlier layers improves models' performance whereas adding noise at intermediate layers deteriorates models' performance. Finally, adding noise towards the final layers performs better than the middle layers noise addition.
健康提及分类处理包含疾病词的给定文本中的疾病检测。然而,非健康和比喻性的疾病词汇的使用给这项任务增加了挑战。最近,对抗训练作为一种正则化手段在许多NLP任务中得到了普及。在本文中,我们提出了一种新的方法来训练涉及对抗性训练的推文健康提及分类的语言模型。我们通过使用高斯噪声在不同级别的推文示例的变压器模型的表示中添加扰动来生成对抗性示例。此外,我们采用对比损失作为一个额外的目标函数。我们在PHM2017数据集扩展版本上对所提出的方法进行了评估。结果表明,与基线方法相比,我们提出的方法显著提高了分类器的性能。此外,我们的分析表明,在早期层添加噪声可以提高模型的性能,而在中间层添加噪声会降低模型的性能。最后,向最后一层添加噪声比中间层添加噪声效果更好。
{"title":"A Novel Approach to Train Diverse Types of Language Models for Health Mention Classification of Tweets","authors":"Pervaiz Iqbal Khan, Imran Razzak, A. Dengel, Sheraz Ahmed","doi":"10.48550/arXiv.2204.06337","DOIUrl":"https://doi.org/10.48550/arXiv.2204.06337","url":null,"abstract":"Health mention classification deals with the disease detection in a given text containing disease words. However, non-health and figurative use of disease words adds challenges to the task. Recently, adversarial training acting as a means of regularization has gained popularity in many NLP tasks. In this paper, we propose a novel approach to train language models for health mention classification of tweets that involves adversarial training. We generate adversarial examples by adding perturbation to the representations of transformer models for tweet examples at various levels using Gaussian noise. Further, we employ contrastive loss as an additional objective function. We evaluate the proposed method on the PHM2017 dataset extended version. Results show that our proposed approach improves the performance of classifier significantly over the baseline methods. Moreover, our analysis shows that adding noise at earlier layers improves models' performance whereas adding noise at intermediate layers deteriorates models' performance. Finally, adding noise towards the final layers performs better than the middle layers noise addition.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"157 1","pages":"136-147"},"PeriodicalIF":0.0,"publicationDate":"2022-04-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73694540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Examining the Proximity of Adversarial Examples to Class Manifolds in Deep Networks 深度网络中对抗性示例与类流形的接近性研究
Stefan Pócos, Iveta Becková, I. Farkaš
Deep neural networks achieve remarkable performance in multiple fields. However, after proper training they suffer from an inherent vulnerability against adversarial examples (AEs). In this work we shed light on inner representations of the AEs by analysing their activations on the hidden layers. We test various types of AEs, each crafted using a specific norm constraint, which affects their visual appearance and eventually their behavior in the trained networks. Our results in image classification tasks (MNIST and CIFAR-10) reveal qualitative differences between the individual types of AEs, when comparing their proximity to the class-specific manifolds on the inner representations. We propose two methods that can be used to compare the distances to class-specific manifolds, regardless of the changing dimensions throughout the network. Using these methods, we consistently confirm that some of the adversarials do not necessarily leave the proximity of the manifold of the correct class, not even in the last hidden layer of the neural network. Next, using UMAP visualisation technique, we project the class activations to 2D space. The results indicate that the activations of the individual AEs are entangled with the activations of the test set. This, however, does not hold for a group of crafted inputs called the rubbish class. We also confirm the entanglement of adversarials with the test set numerically using the soft nearest neighbour loss.
{"title":"Examining the Proximity of Adversarial Examples to Class Manifolds in Deep Networks","authors":"Stefan Pócos, Iveta Becková, I. Farkaš","doi":"10.48550/arXiv.2204.05764","DOIUrl":"https://doi.org/10.48550/arXiv.2204.05764","url":null,"abstract":"Deep neural networks achieve remarkable performance in multiple fields. However, after proper training they suffer from an inherent vulnerability against adversarial examples (AEs). In this work we shed light on inner representations of the AEs by analysing their activations on the hidden layers. We test various types of AEs, each crafted using a specific norm constraint, which affects their visual appearance and eventually their behavior in the trained networks. Our results in image classification tasks (MNIST and CIFAR-10) reveal qualitative differences between the individual types of AEs, when comparing their proximity to the class-specific manifolds on the inner representations. We propose two methods that can be used to compare the distances to class-specific manifolds, regardless of the changing dimensions throughout the network. Using these methods, we consistently confirm that some of the adversarials do not necessarily leave the proximity of the manifold of the correct class, not even in the last hidden layer of the neural network. Next, using UMAP visualisation technique, we project the class activations to 2D space. The results indicate that the activations of the individual AEs are entangled with the activations of the test set. This, however, does not hold for a group of crafted inputs called the rubbish class. We also confirm the entanglement of adversarials with the test set numerically using the soft nearest neighbour loss.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"17 1","pages":"645-656"},"PeriodicalIF":0.0,"publicationDate":"2022-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74886642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Learning Trajectories of Hamiltonian Systems with Neural Networks 神经网络哈密顿系统的学习轨迹
Katsiaryna Haitsiukevich, A. Ilin
Modeling of conservative systems with neural networks is an area of active research. A popular approach is to use Hamiltonian neural networks (HNNs) which rely on the assumptions that a conservative system is described with Hamilton's equations of motion. Many recent works focus on improving the integration schemes used when training HNNs. In this work, we propose to enhance HNNs with an estimation of a continuous-time trajectory of the modeled system using an additional neural network, called a deep hidden physics model in the literature. We demonstrate that the proposed integration scheme works well for HNNs, especially with low sampling rates, noisy and irregular observations.
用神经网络对保守系统进行建模是一个活跃的研究领域。一种流行的方法是使用哈密顿神经网络(HNNs),它依赖于用哈密顿运动方程描述保守系统的假设。最近的许多工作集中在改进训练hnn时使用的集成方案上。在这项工作中,我们建议通过使用一个额外的神经网络(在文献中称为深度隐藏物理模型)来估计建模系统的连续时间轨迹来增强HNNs。我们证明了所提出的集成方案对于hnn,特别是在低采样率、有噪声和不规则观测值的情况下,效果很好。
{"title":"Learning Trajectories of Hamiltonian Systems with Neural Networks","authors":"Katsiaryna Haitsiukevich, A. Ilin","doi":"10.48550/arXiv.2204.05077","DOIUrl":"https://doi.org/10.48550/arXiv.2204.05077","url":null,"abstract":"Modeling of conservative systems with neural networks is an area of active research. A popular approach is to use Hamiltonian neural networks (HNNs) which rely on the assumptions that a conservative system is described with Hamilton's equations of motion. Many recent works focus on improving the integration schemes used when training HNNs. In this work, we propose to enhance HNNs with an estimation of a continuous-time trajectory of the modeled system using an additional neural network, called a deep hidden physics model in the literature. We demonstrate that the proposed integration scheme works well for HNNs, especially with low sampling rates, noisy and irregular observations.","PeriodicalId":93416,"journal":{"name":"Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)","volume":"21 1","pages":"562-573"},"PeriodicalIF":0.0,"publicationDate":"2022-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77754708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Artificial neural networks, ICANN : international conference ... proceedings. International Conference on Artificial Neural Networks (European Neural Network Society)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1