首页 > 最新文献

ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)最新文献

英文 中文
A Convex Formulation for the Robust Estimation of Multivariate Exponential Power Models 多元指数幂模型鲁棒估计的凸公式
N. Ouzir, J. Pesquet, F. Pascal
The multivariate power exponential (MEP) distribution can model a broad range of signals. In noisy scenarios, the robust estimation of the MEP parameters has been traditionally addressed by a fixed-point approach associated with a nonconvex optimization problem. Establishing convergence properties for this approach when the distribution mean is unknown is still an open problem. As an alternative, this paper presents a novel convex formulation for robustly estimating MEP parameters in the presence of multiplicative perturbations. The proposed approach is grounded on a re-parametrization of the original likelihood function in a way that ensures convexity. We also show that this property is preserved for several typical regularization functions. Compared with the robust Tyler’s estimator, the proposed method shows a more accurate precision matrix estimation, with similar mean and covariance estimation performance.
多元幂指数(MEP)分布可以模拟大范围的信号。在噪声情况下,MEP参数的鲁棒估计传统上是通过与非凸优化问题相关的不动点方法来解决的。当分布均值未知时,如何确定该方法的收敛性仍然是一个有待解决的问题。作为一种替代方法,本文提出了一种新的凸公式,用于鲁棒估计存在乘性扰动的MEP参数。提出的方法是基于原始似然函数的重新参数化,以确保凸性。我们还证明了这一性质对于几个典型的正则化函数是保留的。与鲁棒Tyler估计方法相比,该方法具有更精确的矩阵估计精度,且具有相近的均值和协方差估计性能。
{"title":"A Convex Formulation for the Robust Estimation of Multivariate Exponential Power Models","authors":"N. Ouzir, J. Pesquet, F. Pascal","doi":"10.1109/icassp43922.2022.9747354","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9747354","url":null,"abstract":"The multivariate power exponential (MEP) distribution can model a broad range of signals. In noisy scenarios, the robust estimation of the MEP parameters has been traditionally addressed by a fixed-point approach associated with a nonconvex optimization problem. Establishing convergence properties for this approach when the distribution mean is unknown is still an open problem. As an alternative, this paper presents a novel convex formulation for robustly estimating MEP parameters in the presence of multiplicative perturbations. The proposed approach is grounded on a re-parametrization of the original likelihood function in a way that ensures convexity. We also show that this property is preserved for several typical regularization functions. Compared with the robust Tyler’s estimator, the proposed method shows a more accurate precision matrix estimation, with similar mean and covariance estimation performance.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114967629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning to Fuse Heterogeneous Features for Low-Light Image Enhancement 学习融合异构特征的弱光图像增强
Zhenyu Tang, Long Ma, Xiaoke Shang, Xin Fan
To see clearly in low-light scenarios, a series of learning-based techniques have been developed to improve visual quality. However, due to the absence of semantic-level features, the existing methods are perhaps less effective on semantic-oriented visual analysis tasks (e.g., saliency detection). To break down the limitation, we propose a new classification-driven enhancement method with heterogeneous feature fusion. Specifically, we construct a new low-light image enhancement network by integrating features acquired from the pre-trained classification network. Then, to better exploit the semantic-level information, we establish a Heterogeneous Feature Fusion (HF2) operation with channel-and-spatial attention to strength the effects of cross-domain features. HF2 acts on not only the fusion between classification and encoded features but also the fusion between encoded and decoded features. Extensive experiments are conducted to indicate our superiority against other state-of-the-art methods. The application on saliency detection further reveals our effectiveness in settling the semantic-oriented visual tasks.
为了在弱光情况下看得清楚,已经开发了一系列基于学习的技术来提高视觉质量。然而,由于缺乏语义级特征,现有的方法在面向语义的可视化分析任务(例如,显著性检测)上可能不太有效。为了突破这一局限,我们提出了一种基于异构特征融合的分类驱动增强方法。具体来说,我们通过整合从预训练的分类网络中获得的特征,构建了一个新的弱光图像增强网络。然后,为了更好地利用语义级信息,我们建立了信道和空间关注的异构特征融合(HF2)操作,以增强跨域特征的影响。HF2不仅作用于分类特征与编码特征的融合,也作用于编码特征与解码特征的融合。进行了大量的实验,以表明我们的方法优于其他最先进的方法。在显著性检测上的应用进一步揭示了我们在解决面向语义的视觉任务方面的有效性。
{"title":"Learning to Fuse Heterogeneous Features for Low-Light Image Enhancement","authors":"Zhenyu Tang, Long Ma, Xiaoke Shang, Xin Fan","doi":"10.1109/icassp43922.2022.9746255","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9746255","url":null,"abstract":"To see clearly in low-light scenarios, a series of learning-based techniques have been developed to improve visual quality. However, due to the absence of semantic-level features, the existing methods are perhaps less effective on semantic-oriented visual analysis tasks (e.g., saliency detection). To break down the limitation, we propose a new classification-driven enhancement method with heterogeneous feature fusion. Specifically, we construct a new low-light image enhancement network by integrating features acquired from the pre-trained classification network. Then, to better exploit the semantic-level information, we establish a Heterogeneous Feature Fusion (HF2) operation with channel-and-spatial attention to strength the effects of cross-domain features. HF2 acts on not only the fusion between classification and encoded features but also the fusion between encoded and decoded features. Extensive experiments are conducted to indicate our superiority against other state-of-the-art methods. The application on saliency detection further reveals our effectiveness in settling the semantic-oriented visual tasks.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114995782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Connecting Targets via Latent Topics And Contrastive Learning: A Unified Framework For Robust Zero-Shot and Few-Shot Stance Detection 通过潜在主题和对比学习连接目标:一种鲁棒零弹和少弹姿态检测的统一框架
R. Liu, Zheng Lin, Peng Fu, Yuanxin Liu, Weiping Wang
Zero-shot and few-shot stance detection (ZFSD) aims to automatically identify the users’ stance toward a wide range of continuously emerging targets without or with limited labeled data. Previous works on in-target and cross-target stance detection typically focus on extremely limited targets, which is not applicable to the zero-shot and few-shot scenarios. Additionally, existing ZFSD models are not good at modeling the relationship between seen and unseen targets. In this paper, we propose a unified end-to-end framework with a discrete latent topic variable that implicitly establishes the connections between targets. Moreover, we apply supervised contrastive learning to enhance the generalization ability of the model. Comprehensive experiments on the ZFSD task verify the effectiveness and superiority of our proposed method.
零弹和少弹姿态检测(ZFSD)的目的是在没有或有限的标记数据的情况下,自动识别用户对大范围不断出现的目标的姿态。以往的目标内和跨目标姿态检测工作通常只针对极为有限的目标,不适用于零弹和少弹场景。此外,现有的ZFSD模型并不擅长对可见目标和未可见目标之间的关系进行建模。在本文中,我们提出了一个统一的端到端框架,该框架具有离散的潜在主题变量,可以隐式地建立目标之间的联系。此外,我们应用监督对比学习来增强模型的泛化能力。在ZFSD任务上的综合实验验证了该方法的有效性和优越性。
{"title":"Connecting Targets via Latent Topics And Contrastive Learning: A Unified Framework For Robust Zero-Shot and Few-Shot Stance Detection","authors":"R. Liu, Zheng Lin, Peng Fu, Yuanxin Liu, Weiping Wang","doi":"10.1109/icassp43922.2022.9746739","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9746739","url":null,"abstract":"Zero-shot and few-shot stance detection (ZFSD) aims to automatically identify the users’ stance toward a wide range of continuously emerging targets without or with limited labeled data. Previous works on in-target and cross-target stance detection typically focus on extremely limited targets, which is not applicable to the zero-shot and few-shot scenarios. Additionally, existing ZFSD models are not good at modeling the relationship between seen and unseen targets. In this paper, we propose a unified end-to-end framework with a discrete latent topic variable that implicitly establishes the connections between targets. Moreover, we apply supervised contrastive learning to enhance the generalization ability of the model. Comprehensive experiments on the ZFSD task verify the effectiveness and superiority of our proposed method.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"516 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115347432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Deterministic Transform Based Weight Matrices for Neural Networks 基于确定性变换的神经网络权重矩阵
Pol Grau Jurado, Xinyue Liang, S. Chatterjee
We propose to use deterministic transforms as weight matrices for several feedforward neural networks. The use of deterministic transforms helps to reduce the computational complexity in two ways: (1) matrix-vector product complexity in forward pass, helping real time complexity, and (2) fully avoiding backpropagation in the training stage. For each layer of a feedforward network, we pro-pose two unsupervised methods to choose the most appropriate deterministic transform from a set of transforms (a bag of well-known transforms). Experimental results show that the use of deterministic transforms is as good as traditional random matrices in the sense of providing similar classification performance.
我们提出将确定性变换作为几个前馈神经网络的权矩阵。确定性变换的使用有助于在两个方面降低计算复杂度:(1)前向传递的矩阵向量积复杂度,有助于实时复杂度,(2)完全避免训练阶段的反向传播。对于前馈网络的每一层,我们提出了两种无监督的方法来从一组变换(一袋众所周知的变换)中选择最合适的确定性变换。实验结果表明,在提供相似分类性能的意义上,确定性变换的使用与传统随机矩阵一样好。
{"title":"Deterministic Transform Based Weight Matrices for Neural Networks","authors":"Pol Grau Jurado, Xinyue Liang, S. Chatterjee","doi":"10.1109/icassp43922.2022.9747256","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9747256","url":null,"abstract":"We propose to use deterministic transforms as weight matrices for several feedforward neural networks. The use of deterministic transforms helps to reduce the computational complexity in two ways: (1) matrix-vector product complexity in forward pass, helping real time complexity, and (2) fully avoiding backpropagation in the training stage. For each layer of a feedforward network, we pro-pose two unsupervised methods to choose the most appropriate deterministic transform from a set of transforms (a bag of well-known transforms). Experimental results show that the use of deterministic transforms is as good as traditional random matrices in the sense of providing similar classification performance.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115405950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real-World On-Board Uav Audio Data Set For Propeller Anomalies 真实世界的机载无人机音频数据集螺旋桨异常
Sai Srinadhu Katta, Kide Vuojärvi, S. Nandyala, Ulla-Maria Kovalainen, Lauren Baddeley
Detecting propeller damage in Unmanned Aerial Vehicles (UAV) is a crucial step in ensuring their operational resilience and safety. In this work, we present a novel real-world audio data set of propeller anomalies, and use several deep learning models to classify the damage. This data set consists of more than 5 hours of audio recordings, covering all configurations of intact and broken propellers in a UAV quadcopter. A microphone array was mounted onto a UAV, and numerous autonomous indoor missions were flown. Our on-board setup has provided clean audio recordings containing little background noise. We have developed classification models for this data set, using different deep learning architectures: Deep Neural Networks (DNNs), Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM), and Transformer Encoder (TrEnc). We conclude that the TrEnc outperforms other architectures, having 11k parameters, .57M Flops, 98.30% accuracy, .98 precision, and .98 recall. Finally, we make our data set publicly available here⊙.
无人机螺旋桨损伤检测是保证无人机运行弹性和安全性的关键环节。在这项工作中,我们提出了一种新的真实世界螺旋桨异常音频数据集,并使用几种深度学习模型对损伤进行分类。该数据集由超过5小时的录音组成,涵盖了无人机四轴飞行器中完整和损坏的螺旋桨的所有配置。一个麦克风阵列被安装在一架无人机上,并执行了许多自主室内任务。我们的车载设置提供了包含少量背景噪音的干净音频记录。我们为该数据集开发了分类模型,使用不同的深度学习架构:深度神经网络(dnn)、卷积神经网络(cnn)、长短期记忆(LSTM)和变压器编码器(TrEnc)。我们得出结论,TrEnc优于其他架构,具有11k参数,0.57 m Flops, 98.30%准确率,0.98精度和0.98召回率。最后,我们在这里公开了我们的数据集⊙。
{"title":"Real-World On-Board Uav Audio Data Set For Propeller Anomalies","authors":"Sai Srinadhu Katta, Kide Vuojärvi, S. Nandyala, Ulla-Maria Kovalainen, Lauren Baddeley","doi":"10.1109/ICASSP43922.2022.9747789","DOIUrl":"https://doi.org/10.1109/ICASSP43922.2022.9747789","url":null,"abstract":"Detecting propeller damage in Unmanned Aerial Vehicles (UAV) is a crucial step in ensuring their operational resilience and safety. In this work, we present a novel real-world audio data set of propeller anomalies, and use several deep learning models to classify the damage. This data set consists of more than 5 hours of audio recordings, covering all configurations of intact and broken propellers in a UAV quadcopter. A microphone array was mounted onto a UAV, and numerous autonomous indoor missions were flown. Our on-board setup has provided clean audio recordings containing little background noise. We have developed classification models for this data set, using different deep learning architectures: Deep Neural Networks (DNNs), Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM), and Transformer Encoder (TrEnc). We conclude that the TrEnc outperforms other architectures, having 11k parameters, .57M Flops, 98.30% accuracy, .98 precision, and .98 recall. Finally, we make our data set publicly available here⊙.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115447060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
FDSNeT: An Accurate Real-Time Surface Defect Segmentation Network FDSNeT:一个精确的实时表面缺陷分割网络
Jian Zhang, Runwei Ding, Miaoju Ban, Tianyu Guo
Surface defect detection is a common task for industrial quality control, which increasingly requires accuracy and real-time ability. However, the current segmentation networks are not effective in dealing with defect boundary details, local similarity of different defects and low contrast between defect and background. To this end, we propose a real-time surface defect segmentation network (FDSNet) based on two-branch architecture, in which two corresponding auxiliary tasks are introduced to encode more boundary details and semantic context. To handle the local similarity problem of different surface defects, we propose a Global Context Upsampling (GCU) module by capturing long-range context from multi-scales. Moreover, we present a representative Mobile phone screen Surface Defect (MSD) segmentation dataset to alleviate the lack of dataset in this field. Experiments on NEU-Seg, Magnetic-tile-defect-datasets and MSD dataset show that the proposed FDSNet achieves promising trade-off between accuracy and inference speed. The dataset and code are available at https://github.com/jianzhang96/fdsnet.
表面缺陷检测是工业质量控制的一项常见任务,对其准确性和实时性的要求越来越高。然而,现有的分割网络在处理缺陷边界细节、不同缺陷局部相似度以及缺陷与背景对比度低等方面效果不佳。为此,我们提出了一种基于双分支架构的实时表面缺陷分割网络(FDSNet),该网络引入了两个相应的辅助任务来编码更多的边界细节和语义上下文。为了解决不同表面缺陷的局部相似性问题,提出了一种从多尺度捕获远程上下文的全局上下文上采样(Global Context Upsampling, GCU)模块。此外,我们提出了具有代表性的手机屏幕表面缺陷(MSD)分割数据集,以缓解该领域数据集的不足。在nue - seg、磁砖-缺陷数据集和MSD数据集上的实验表明,所提出的FDSNet在准确率和推理速度之间取得了很好的平衡。数据集和代码可在https://github.com/jianzhang96/fdsnet上获得。
{"title":"FDSNeT: An Accurate Real-Time Surface Defect Segmentation Network","authors":"Jian Zhang, Runwei Ding, Miaoju Ban, Tianyu Guo","doi":"10.1109/icassp43922.2022.9747311","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9747311","url":null,"abstract":"Surface defect detection is a common task for industrial quality control, which increasingly requires accuracy and real-time ability. However, the current segmentation networks are not effective in dealing with defect boundary details, local similarity of different defects and low contrast between defect and background. To this end, we propose a real-time surface defect segmentation network (FDSNet) based on two-branch architecture, in which two corresponding auxiliary tasks are introduced to encode more boundary details and semantic context. To handle the local similarity problem of different surface defects, we propose a Global Context Upsampling (GCU) module by capturing long-range context from multi-scales. Moreover, we present a representative Mobile phone screen Surface Defect (MSD) segmentation dataset to alleviate the lack of dataset in this field. Experiments on NEU-Seg, Magnetic-tile-defect-datasets and MSD dataset show that the proposed FDSNet achieves promising trade-off between accuracy and inference speed. The dataset and code are available at https://github.com/jianzhang96/fdsnet.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"133 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123064645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Learning Deep Pathological Features for WSI-Level Cervical Cancer Grading 了解wsi级宫颈癌分级的深层病理特征
Ruixiang Geng, Qing Liu, Shuo Feng, Yixiong Liang
Fully automated cervical cancer grading on the level of Whole Slide Images (WSI) is a challenge task. As WSIs are in gigapixel resolution, it is impossible to train a deep classification neural network with the entire WSIs as inputs. To bypass this problem, we propose a two-stage learning framework. In detail, we propose to first learn patch-level deep pathological features for smear patches via a patch-level feature learning module, which is trained via leveraging the cell instance detection task. Then, we propose to learn WSI-level pathological features from patch-level features for cervical cancer grading. We conduct extensive experiments on our private dataset and make comparisons with rule-based cervical cancer grading methods. Experimental results demonstrate that our proposed deep feature-based WSI-level cervical cancer grading method achieves state-of-the-art performance.
在全幻灯片图像(WSI)水平上的全自动宫颈癌分级是一项具有挑战性的任务。由于wsi的分辨率为十亿像素,因此不可能将整个wsi作为输入来训练深度分类神经网络。为了绕过这个问题,我们提出了一个两阶段学习框架。具体而言,我们建议首先通过利用细胞实例检测任务来训练的斑块级特征学习模块来学习涂抹斑块的斑块级深度病理特征。然后,我们建议从斑块级特征中学习wsi级病理特征,用于宫颈癌分级。我们在我们的私人数据集上进行了广泛的实验,并与基于规则的宫颈癌分级方法进行了比较。实验结果表明,我们提出的基于深度特征的wsi级宫颈癌分级方法达到了最先进的性能。
{"title":"Learning Deep Pathological Features for WSI-Level Cervical Cancer Grading","authors":"Ruixiang Geng, Qing Liu, Shuo Feng, Yixiong Liang","doi":"10.1109/ICASSP43922.2022.9747112","DOIUrl":"https://doi.org/10.1109/ICASSP43922.2022.9747112","url":null,"abstract":"Fully automated cervical cancer grading on the level of Whole Slide Images (WSI) is a challenge task. As WSIs are in gigapixel resolution, it is impossible to train a deep classification neural network with the entire WSIs as inputs. To bypass this problem, we propose a two-stage learning framework. In detail, we propose to first learn patch-level deep pathological features for smear patches via a patch-level feature learning module, which is trained via leveraging the cell instance detection task. Then, we propose to learn WSI-level pathological features from patch-level features for cervical cancer grading. We conduct extensive experiments on our private dataset and make comparisons with rule-based cervical cancer grading methods. Experimental results demonstrate that our proposed deep feature-based WSI-level cervical cancer grading method achieves state-of-the-art performance.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123067676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Contrastive Translation Learning For Medical Image Segmentation 用于医学图像分割的对比翻译学习
Wankang Zeng, Wenkang Fan, Dongfang Shen, Yinran Chen, Xióngbiao Luó
Unsupervised domain adaptation commonly uses cycle generative networks to produce synthesis data from source to target domains. Unfortunately, translated samples cannot effectively preserve semantic information from input sources, resulting in bad or low adaptability of the network to segment target data. This work proposes an advantageous domain translation mechanism to improve the perceptual ability of the network for accurate unlabeled target data segmentation. Our domain translation employs patchwise contrastive learning to improve the semantic content consistency between input and translated images. Our approach was applied to unsupervised domain adaptation based abdominal organ segmentation. The experimental results demonstrate the effectiveness of our framework that outperforms other methods.
无监督域自适应通常使用循环生成网络产生从源域到目标域的综合数据。不幸的是,翻译后的样本不能有效地保留输入源的语义信息,导致网络对目标数据的分割适应性差或较低。这项工作提出了一个有利的领域翻译机制,以提高网络对准确的未标记目标数据分割的感知能力。我们的领域翻译采用补丁对比学习来提高输入和翻译图像之间的语义内容一致性。将该方法应用于基于无监督域自适应的腹部器官分割。实验结果表明,该框架的有效性优于其他方法。
{"title":"Contrastive Translation Learning For Medical Image Segmentation","authors":"Wankang Zeng, Wenkang Fan, Dongfang Shen, Yinran Chen, Xióngbiao Luó","doi":"10.1109/icassp43922.2022.9747097","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9747097","url":null,"abstract":"Unsupervised domain adaptation commonly uses cycle generative networks to produce synthesis data from source to target domains. Unfortunately, translated samples cannot effectively preserve semantic information from input sources, resulting in bad or low adaptability of the network to segment target data. This work proposes an advantageous domain translation mechanism to improve the perceptual ability of the network for accurate unlabeled target data segmentation. Our domain translation employs patchwise contrastive learning to improve the semantic content consistency between input and translated images. Our approach was applied to unsupervised domain adaptation based abdominal organ segmentation. The experimental results demonstrate the effectiveness of our framework that outperforms other methods.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123178317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FlowDT: A Flow-Aware Digital Twin for Computer Networks 面向计算机网络的流感知数字孪生
Miquel Ferriol Galmés, Xiangle Cheng, Xiang Shi, Shihan Xiao, P. Barlet-Ros, A. Cabellos-Aparicio
Network modeling is an essential tool for network planning and management. It allows network administrators to explore the performance of new protocols, mechanisms, or optimal configurations without the need for testing them in real production networks. Recently, Graph Neural Networks (GNNs) have emerged as a practical solution to produce network models that can learn and extract complex patterns from real data without making any assumptions. However, state-of-the-art GNN-based network models only work with traffic matrices, this is a very coarse and simplified representation of network traffic. Although this assumption has shown to work well in certain use-cases, it is a limiting factor because, in practice, networks operate with flows. In this paper, we present FlowDT a new DL-based solution designed to model computer networks at the fine-grained flow level. In our evaluation, we show how FlowDT can accurately predict relevant per-flow performance metrics with an error of 3.5%, FlowDT’s performance is also benchmarked against vanilla DL models as well as with Queuing Theory.
网络建模是网络规划和管理的重要工具。它允许网络管理员探索新协议、机制或最佳配置的性能,而无需在实际生产网络中进行测试。最近,图神经网络(gnn)作为一种实用的解决方案出现了,它可以在不做任何假设的情况下从真实数据中学习和提取复杂模式的网络模型。然而,最先进的基于gnn的网络模型只能处理流量矩阵,这是一个非常粗糙和简化的网络流量表示。尽管这个假设在某些用例中表现得很好,但它是一个限制因素,因为在实践中,网络与流一起操作。在本文中,我们提出了FlowDT一种新的基于dl的解决方案,旨在对细粒度流级别的计算机网络进行建模。在我们的评估中,我们展示了FlowDT如何准确地预测相关的每流性能指标,误差为3.5%,FlowDT的性能也与香草DL模型以及排队理论进行了基准测试。
{"title":"FlowDT: A Flow-Aware Digital Twin for Computer Networks","authors":"Miquel Ferriol Galmés, Xiangle Cheng, Xiang Shi, Shihan Xiao, P. Barlet-Ros, A. Cabellos-Aparicio","doi":"10.1109/icassp43922.2022.9746953","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9746953","url":null,"abstract":"Network modeling is an essential tool for network planning and management. It allows network administrators to explore the performance of new protocols, mechanisms, or optimal configurations without the need for testing them in real production networks. Recently, Graph Neural Networks (GNNs) have emerged as a practical solution to produce network models that can learn and extract complex patterns from real data without making any assumptions. However, state-of-the-art GNN-based network models only work with traffic matrices, this is a very coarse and simplified representation of network traffic. Although this assumption has shown to work well in certain use-cases, it is a limiting factor because, in practice, networks operate with flows. In this paper, we present FlowDT a new DL-based solution designed to model computer networks at the fine-grained flow level. In our evaluation, we show how FlowDT can accurately predict relevant per-flow performance metrics with an error of 3.5%, FlowDT’s performance is also benchmarked against vanilla DL models as well as with Queuing Theory.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114668895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Two-Stage Beam Training and Channel Estimation for Ris-Aided Mmwave Systems Via Fast Alternating Least Squares 基于快速交替最小二乘的ris辅助毫米波系统有效两级波束训练和信道估计
Hyeonjin Chung, Sunwoo Kim
This paper proposes a two-stage beam training and a channel estimation based on fast alternating least squares (FALS) for reconfigurable intelligent surface (RIS)-aided millimeter-wave systems. To reduce the beam training overhead, only selected columns and rows of the channel matrix are observed by two-stage beam training. This beam training produces a partly observed channel matrix with low coherence, which enables the low rank matrix completion technique to recover unobserved entries. Unobserved entries are recovered by FALS, which alternatingly updates the left and the right singular vectors that comprise the channel. Simulation results and analysis show that the proposed algorithm is computationally efficient and has superior accuracy to existing algorithms.
针对可重构智能表面(RIS)辅助毫米波系统,提出了一种基于快速交替最小二乘(FALS)的两阶段波束训练和信道估计方法。为了减少波束训练开销,两阶段波束训练只观察信道矩阵中选定的列和行。这种波束训练产生了一个低相干性的部分观察到的信道矩阵,这使得低秩矩阵补全技术能够恢复未观察到的条目。未观察到的条目由FALS恢复,FALS交替更新组成通道的左、右奇异向量。仿真结果和分析表明,该算法计算效率高,精度优于现有算法。
{"title":"Efficient Two-Stage Beam Training and Channel Estimation for Ris-Aided Mmwave Systems Via Fast Alternating Least Squares","authors":"Hyeonjin Chung, Sunwoo Kim","doi":"10.1109/icassp43922.2022.9746094","DOIUrl":"https://doi.org/10.1109/icassp43922.2022.9746094","url":null,"abstract":"This paper proposes a two-stage beam training and a channel estimation based on fast alternating least squares (FALS) for reconfigurable intelligent surface (RIS)-aided millimeter-wave systems. To reduce the beam training overhead, only selected columns and rows of the channel matrix are observed by two-stage beam training. This beam training produces a partly observed channel matrix with low coherence, which enables the low rank matrix completion technique to recover unobserved entries. Unobserved entries are recovered by FALS, which alternatingly updates the left and the right singular vectors that comprise the channel. Simulation results and analysis show that the proposed algorithm is computationally efficient and has superior accuracy to existing algorithms.","PeriodicalId":272439,"journal":{"name":"ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115828525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
全部 Geobiology Appl. Clay Sci. Geochim. Cosmochim. Acta J. Hydrol. Org. Geochem. Carbon Balance Manage. Contrib. Mineral. Petrol. Int. J. Biometeorol. IZV-PHYS SOLID EART+ J. Atmos. Chem. Acta Oceanolog. Sin. Acta Geophys. ACTA GEOL POL ACTA PETROL SIN ACTA GEOL SIN-ENGL AAPG Bull. Acta Geochimica Adv. Atmos. Sci. Adv. Meteorol. Am. J. Phys. Anthropol. Am. J. Sci. Am. Mineral. Annu. Rev. Earth Planet. Sci. Appl. Geochem. Aquat. Geochem. Ann. Glaciol. Archaeol. Anthropol. Sci. ARCHAEOMETRY ARCT ANTARCT ALP RES Asia-Pac. J. Atmos. Sci. ATMOSPHERE-BASEL Atmos. Res. Aust. J. Earth Sci. Atmos. Chem. Phys. Atmos. Meas. Tech. Basin Res. Big Earth Data BIOGEOSCIENCES Geostand. Geoanal. Res. GEOLOGY Geosci. J. Geochem. J. Geochem. Trans. Geosci. Front. Geol. Ore Deposits Global Biogeochem. Cycles Gondwana Res. Geochem. Int. Geol. J. Geophys. Prospect. Geosci. Model Dev. GEOL BELG GROUNDWATER Hydrogeol. J. Hydrol. Earth Syst. Sci. Hydrol. Processes Int. J. Climatol. Int. J. Earth Sci. Int. Geol. Rev. Int. J. Disaster Risk Reduct. Int. J. Geomech. Int. J. Geog. Inf. Sci. Isl. Arc J. Afr. Earth. Sci. J. Adv. Model. Earth Syst. J APPL METEOROL CLIM J. Atmos. Oceanic Technol. J. Atmos. Sol. Terr. Phys. J. Clim. J. Earth Sci. J. Earth Syst. Sci. J. Environ. Eng. Geophys. J. Geog. Sci. Mineral. Mag. Miner. Deposita Mon. Weather Rev. Nat. Hazards Earth Syst. Sci. Nat. Clim. Change Nat. Geosci. Ocean Dyn. Ocean and Coastal Research npj Clim. Atmos. Sci. Ocean Modell. Ocean Sci. Ore Geol. Rev. OCEAN SCI J Paleontol. J. PALAEOGEOGR PALAEOCL PERIOD MINERAL PETROLOGY+ Phys. Chem. Miner. Polar Sci. Prog. Oceanogr. Quat. Sci. Rev. Q. J. Eng. Geol. Hydrogeol. RADIOCARBON Pure Appl. Geophys. Resour. Geol. Rev. Geophys. Sediment. Geol.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1