首页 > 最新文献

2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)最新文献

英文 中文
Embedding Methods or Link-based Similarity Measures, Which is Better for Link Prediction? 嵌入方法和基于链接的相似性度量,哪个更适合链接预测?
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660590
M. Hamedani, Sang-Wook Kim
The link prediction task has attracted significant attention in the literature. Link-based similarity measures (in short, similarity measures) are the conventional methods for this task, while recently graph embedding methods (in short, embedding methods) are widely employed as well. In this paper, we extensively investigate the effectiveness of embedding methods and similarity measures (i.e., both non-recursive and recursive ones) in link prediction. Our experimental results with three real-world datasets demonstrate that 1) recursive similarity measures are not beneficial in this task than non-recursive one,2) increasing the number of dimensions in vectors may not help improve the accuracy of embedding methods, and 3) in comparison with embedding methods, Adamic/Adar, a non-recursive similarity measure, can be a useful method for link prediction since it shows promising results while being parameter-free.
链接预测任务在文献中引起了极大的关注。基于链接的相似度度量(简称相似度度量)是该任务的常规方法,而最近图嵌入方法(简称嵌入方法)也被广泛采用。在本文中,我们广泛地研究了嵌入方法和相似性度量(即非递归和递归)在链接预测中的有效性。我们对三个真实数据集的实验结果表明:1)递归相似性度量在该任务中不如非递归相似性度量有利;2)增加向量的维数可能无助于提高嵌入方法的准确性;3)与嵌入方法相比,非递归相似性度量Adamic/Adar可以是一种有用的链接预测方法,因为它在无参数的情况下显示出令人满意的结果。
{"title":"Embedding Methods or Link-based Similarity Measures, Which is Better for Link Prediction?","authors":"M. Hamedani, Sang-Wook Kim","doi":"10.1109/IC-NIDC54101.2021.9660590","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660590","url":null,"abstract":"The link prediction task has attracted significant attention in the literature. Link-based similarity measures (in short, similarity measures) are the conventional methods for this task, while recently graph embedding methods (in short, embedding methods) are widely employed as well. In this paper, we extensively investigate the effectiveness of embedding methods and similarity measures (i.e., both non-recursive and recursive ones) in link prediction. Our experimental results with three real-world datasets demonstrate that 1) recursive similarity measures are not beneficial in this task than non-recursive one,2) increasing the number of dimensions in vectors may not help improve the accuracy of embedding methods, and 3) in comparison with embedding methods, Adamic/Adar, a non-recursive similarity measure, can be a useful method for link prediction since it shows promising results while being parameter-free.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116917350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Small Range Ergodic Beamforming Method Based on Binocular Vision Positioning 基于双目视觉定位的小范围遍历波束形成方法
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660448
Bo-cheng Yu, Xin Zhang
The increasing construction of 5G dense network creates the conditions for the application of Massive MIMO system. However, with the continuous expansion of business requirements, users put forward higher requirements for the number of antennas in MIMO system. With the increase of the number of antennas, the cost of traditional MIMO beamforming algorithm for channel detection and feedback will increase rapidly, which consumes more wire-less resources and greatly increases the computational burden of the system. The use of computer vision aids provides convenience for the beamforming method to track the target accurately under LOS condition. Combined with image tracking algorithm, the position of the target in each image frame can be calculated so that the angle information of LOS path and the best beam-forming scheme can be determined directly, which can reduce the cost and calculation of the system through wireless resource measurement and feed-back. As a result, the operation speed and accuracy of the system are improved. In this paper, a beamforming method based on binocular positioning is studied. Compared with the traditional method, this method can reduce the number of codeword searches and improve the channel capacity in high-density 5G network.
5G密集网络的不断建设,为大规模MIMO系统的应用创造了条件。然而,随着业务需求的不断扩大,用户对MIMO系统中的天线数量提出了更高的要求。随着天线数量的增加,传统MIMO波束形成算法用于信道检测和反馈的成本将迅速增加,消耗更多的无线资源,大大增加了系统的计算负担。计算机视觉辅助的使用为波束形成方法在目视条件下准确跟踪目标提供了方便。结合图像跟踪算法,可以计算出目标在每帧图像中的位置,从而直接确定LOS路径的角度信息和最佳波束形成方案,通过无线资源测量和反馈减少系统的成本和计算量。从而提高了系统的运行速度和精度。本文研究了一种基于双目定位的波束形成方法。与传统方法相比,该方法可以减少码字搜索次数,提高高密度5G网络的信道容量。
{"title":"A Small Range Ergodic Beamforming Method Based on Binocular Vision Positioning","authors":"Bo-cheng Yu, Xin Zhang","doi":"10.1109/IC-NIDC54101.2021.9660448","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660448","url":null,"abstract":"The increasing construction of 5G dense network creates the conditions for the application of Massive MIMO system. However, with the continuous expansion of business requirements, users put forward higher requirements for the number of antennas in MIMO system. With the increase of the number of antennas, the cost of traditional MIMO beamforming algorithm for channel detection and feedback will increase rapidly, which consumes more wire-less resources and greatly increases the computational burden of the system. The use of computer vision aids provides convenience for the beamforming method to track the target accurately under LOS condition. Combined with image tracking algorithm, the position of the target in each image frame can be calculated so that the angle information of LOS path and the best beam-forming scheme can be determined directly, which can reduce the cost and calculation of the system through wireless resource measurement and feed-back. As a result, the operation speed and accuracy of the system are improved. In this paper, a beamforming method based on binocular positioning is studied. Compared with the traditional method, this method can reduce the number of codeword searches and improve the channel capacity in high-density 5G network.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"14 5-6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116471315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Zombie Hosts Identification Based on DNS Log 基于DNS日志识别僵尸主机
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660578
Renjie Wang, Yangsen Zhang, Ruixue Duan, Zhuofan Huang
Although the academia has done a lot of research on DNS abnormal behavior, whether from the perspective of traffic or irregular domain name recognition, the mechanism behind DNS is ignored in the pre-processing of DNS logs and other data. In addition, most studies focus on traffic anomaly detection and unconventional domain name recognition, and lack of systematic research on the combination of the two, so the proposed algorithm has no practical application. This paper proposes a clustering method based on DNS client IP address traffic characteristics, which divides DNS logs into five access modes. Then, a DNS log preprocessing algorithm is designed to preprocess the logs that may exist in zombie hosts. Finally, a two-layer GRU network detection algorithm based on domain name text features is proposed. Experimental results show that this method can effectively identify zombie hosts in DNS logs.
虽然学术界对DNS异常行为进行了大量的研究,但无论是从流量的角度还是从不规则域名识别的角度,在对DNS日志等数据进行预处理时,都忽略了DNS背后的机制。此外,大多数研究集中在流量异常检测和非常规域名识别方面,缺乏对两者结合的系统研究,因此所提出的算法没有实际应用。本文提出了一种基于DNS客户端IP地址流量特征的聚类方法,将DNS日志划分为五种访问模式。然后设计DNS日志预处理算法,对僵尸主机中可能存在的日志进行预处理。最后,提出了一种基于域名文本特征的两层GRU网络检测算法。实验结果表明,该方法可以有效识别DNS日志中的僵尸主机。
{"title":"Zombie Hosts Identification Based on DNS Log","authors":"Renjie Wang, Yangsen Zhang, Ruixue Duan, Zhuofan Huang","doi":"10.1109/IC-NIDC54101.2021.9660578","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660578","url":null,"abstract":"Although the academia has done a lot of research on DNS abnormal behavior, whether from the perspective of traffic or irregular domain name recognition, the mechanism behind DNS is ignored in the pre-processing of DNS logs and other data. In addition, most studies focus on traffic anomaly detection and unconventional domain name recognition, and lack of systematic research on the combination of the two, so the proposed algorithm has no practical application. This paper proposes a clustering method based on DNS client IP address traffic characteristics, which divides DNS logs into five access modes. Then, a DNS log preprocessing algorithm is designed to preprocess the logs that may exist in zombie hosts. Finally, a two-layer GRU network detection algorithm based on domain name text features is proposed. Experimental results show that this method can effectively identify zombie hosts in DNS logs.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123986820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Degree Matters: Assessing the Generalization of Graph Neural Network 程度问题:评估图神经网络的泛化
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660574
Hyunmok Park, Kijung Yoon
Graph neural network (GNN) is a general framework for using deep neural networks on graph data. The defining feature of a GNN is that it uses a form of neural message passing where vector messages are exchanged between nodes and updated using neural networks. The message passing operation that underlies GNNs has recently been applied to develop neural approximate inference algorithms, but little work has been done on understanding under what conditions GNNs can be used as a core module for building general inference models. To study this question, we consider the task of out-of-distribution generalization where training and test data have different distributions, by systematically investigating how the graph size and structural properties affect the inferential performance of GNNs. We find that (1) the average unique node degree is one of the key features in predicting whether GNNs can generalize to unseen graphs; (2) the graph size is not a fundamental limiting factor of the generalization in GNNs when the average node degree remains invariant across training and test distributions; (3) despite the size-invariant generalization, training GNNs on graphs of high degree (and of large size consequently) is not trivial (4) neural inference by GNNs outperforms algorithmic inferences especially when the pairwise potentials are strong, which naturally makes the inference problem harder.
图神经网络(GNN)是在图数据上使用深度神经网络的通用框架。GNN的定义特征是它使用一种神经消息传递形式,其中向量消息在节点之间交换并使用神经网络更新。作为gnn基础的消息传递操作最近被应用于开发神经近似推理算法,但在理解gnn在什么条件下可以用作构建一般推理模型的核心模块方面做的工作很少。为了研究这个问题,我们考虑了分布外泛化的任务,其中训练数据和测试数据具有不同的分布,通过系统地研究图的大小和结构性质如何影响gnn的推理性能。我们发现(1)平均唯一节点度是预测gnn能否泛化到未见图的关键特征之一;(2)当平均节点度在训练分布和测试分布之间保持不变时,图大小不是gnn泛化的基本限制因素;(3)尽管有大小不变的泛化,但在高程度(因此也有大尺寸)的图上训练gnn并非易事(4)gnn的神经推理优于算法推理,特别是当配对电位很强时,这自然使推理问题变得更加困难。
{"title":"Degree Matters: Assessing the Generalization of Graph Neural Network","authors":"Hyunmok Park, Kijung Yoon","doi":"10.1109/IC-NIDC54101.2021.9660574","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660574","url":null,"abstract":"Graph neural network (GNN) is a general framework for using deep neural networks on graph data. The defining feature of a GNN is that it uses a form of neural message passing where vector messages are exchanged between nodes and updated using neural networks. The message passing operation that underlies GNNs has recently been applied to develop neural approximate inference algorithms, but little work has been done on understanding under what conditions GNNs can be used as a core module for building general inference models. To study this question, we consider the task of out-of-distribution generalization where training and test data have different distributions, by systematically investigating how the graph size and structural properties affect the inferential performance of GNNs. We find that (1) the average unique node degree is one of the key features in predicting whether GNNs can generalize to unseen graphs; (2) the graph size is not a fundamental limiting factor of the generalization in GNNs when the average node degree remains invariant across training and test distributions; (3) despite the size-invariant generalization, training GNNs on graphs of high degree (and of large size consequently) is not trivial (4) neural inference by GNNs outperforms algorithmic inferences especially when the pairwise potentials are strong, which naturally makes the inference problem harder.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126771654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Target Detection Method Based on the Fusion Algorithm of Radar and Camera 一种基于雷达与相机融合算法的目标检测方法
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660407
Sheng Zhuang, Lin Cao, Zongmin Zhao, Dongfeng Wang
The method based on the fusion of radar and video in this paper is oriented to detecting surrounding objects while driving. This is usually a method of improving robustness and accuracy by using several senses, which makes sensor fusion a key part of the perception system. We propose a new fusion method called CT-EPNP, which uses radar and camera data for fast detection. Adding a central fusion algorithm on the basis of EPNP, and use the truncated cone method to compensate the radar information on the associated image when mapping. CT-EPNP returns to the object attributes depth, rotation, speed and other attributes. Based on this, simulation verification and related derivation of mathematical formulas are proved. We combined the improved algorithm with the RetinaNet model to ensure that the model is satisfied with the normal driving of the vehicle while gaining a certain increase in the detection rate. We have also made a certain improvement in ensuring repeated detection without using any additional time information.
本文提出了一种基于雷达与视频融合的汽车驾驶过程中周围物体检测方法。这通常是一种通过使用多个感官来提高鲁棒性和准确性的方法,这使得传感器融合成为感知系统的关键部分。我们提出了一种新的融合方法,称为CT-EPNP,利用雷达和相机数据进行快速检测。在EPNP的基础上增加了中心融合算法,并在映射时使用截锥法对关联图像上的雷达信息进行补偿。CT-EPNP返回对象属性深度、旋转、速度等属性。在此基础上,进行了仿真验证和相关数学公式的推导。我们将改进后的算法与retanet模型相结合,在保证模型满足车辆正常行驶的同时,也获得了一定的检测率提高。在确保重复检测而不使用任何额外的时间信息方面,我们也做了一定的改进。
{"title":"A Target Detection Method Based on the Fusion Algorithm of Radar and Camera","authors":"Sheng Zhuang, Lin Cao, Zongmin Zhao, Dongfeng Wang","doi":"10.1109/IC-NIDC54101.2021.9660407","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660407","url":null,"abstract":"The method based on the fusion of radar and video in this paper is oriented to detecting surrounding objects while driving. This is usually a method of improving robustness and accuracy by using several senses, which makes sensor fusion a key part of the perception system. We propose a new fusion method called CT-EPNP, which uses radar and camera data for fast detection. Adding a central fusion algorithm on the basis of EPNP, and use the truncated cone method to compensate the radar information on the associated image when mapping. CT-EPNP returns to the object attributes depth, rotation, speed and other attributes. Based on this, simulation verification and related derivation of mathematical formulas are proved. We combined the improved algorithm with the RetinaNet model to ensure that the model is satisfied with the normal driving of the vehicle while gaining a certain increase in the detection rate. We have also made a certain improvement in ensuring repeated detection without using any additional time information.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129564232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-driven Differential Games for Affine Nonlinear Systems 仿射非线性系统的数据驱动微分对策
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660508
Conghui Ma, Bin Zhang, Lutao Yan, Haiyuan Li
This paper presents a data-driven optimal approach based on differential dynamic programming (DDP) for two-person differential game of nonlinear affine systems. Using test data, the Hamilton-Jacobi-Isaacs (HJI) equation is expanded into a set of high-order differential equations. Basis functions is adopted to approximate the unknown system function and value function. Based on the approximation, a data-driven optimal approach is proposed to obtain the unknown coefficients of the basis functions. A numerical example is proposed to demonstrate the effectiveness of this method.
针对非线性仿射系统的二人微分对策问题,提出了一种基于差分动态规划的数据驱动优化方法。利用试验数据,将Hamilton-Jacobi-Isaacs (HJI)方程展开为一组高阶微分方程。采用基函数逼近未知的系统函数和值函数。在此基础上,提出了一种数据驱动的优化方法来获取基函数的未知系数。通过数值算例验证了该方法的有效性。
{"title":"Data-driven Differential Games for Affine Nonlinear Systems","authors":"Conghui Ma, Bin Zhang, Lutao Yan, Haiyuan Li","doi":"10.1109/IC-NIDC54101.2021.9660508","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660508","url":null,"abstract":"This paper presents a data-driven optimal approach based on differential dynamic programming (DDP) for two-person differential game of nonlinear affine systems. Using test data, the Hamilton-Jacobi-Isaacs (HJI) equation is expanded into a set of high-order differential equations. Basis functions is adopted to approximate the unknown system function and value function. Based on the approximation, a data-driven optimal approach is proposed to obtain the unknown coefficients of the basis functions. A numerical example is proposed to demonstrate the effectiveness of this method.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131926997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Auto-Learning of Parameters for High Resolution Sparse Group Lasso SAR Imagery 高分辨率稀疏组Lasso SAR图像参数的自动学习
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660447
Wei Liu, Hanwen Xu, Cheng Fang, Lei Yang, Weidong Jiao
Aiming at the problem of adjusting the penalty term coefficient of feature enhancement in high-resolution synthetic aperture radar (SAR) imaging, a marginal estimation Bayes (MEB) algorithm is proposed, so that the prior features of the target can be fitted properly to improve the accuracy of image feature extraction. Firstly, the alternating direction method of multipliers (ADMM) convex optimization framework is modeled based on the echoed data, and least absolute shrinkage and selection operator (Lasso) model and sparse group Lasso (SG-Lasso) model are introduced, then the maximum marginal likelihood distribution of the regularization parameters is derived. Moreover, the Moreau Yoshida unadjusted Langevin algorithm (MYULA) is used to realize target posteriori sampling solution. Because the posterior distribution is difficult to solve, the gradient projection method is introduced to estimate the regularization parameters. Finally, auto-learning parameters are used to optimize the imaging. The proposed algorithm can not only estimate the parameters of a single regularization term, but also estimate the parameters of multiple regularization terms. Aiming at non-differentiable part in the prior, MYULA is adopted to calculate the subgradient of the non-differentiable posterior distribution. Therefore, the proposed algorithm is capable of auto-leaning parameters even regularization function is non-differentiable. In the experimental part, compared with the optimal value of manual debugging, the error between the proposed method and the optimal value is within 15%, and the effectiveness of the algorithm are verified by phase transition diagram (PTD).
针对高分辨率合成孔径雷达(SAR)成像中特征增强惩罚项系数的调整问题,提出了一种边缘估计贝叶斯(MEB)算法,使目标的先验特征得到适当拟合,提高图像特征提取的精度。首先,基于回波数据建立了交替方向乘子法(ADMM)凸优化框架模型,引入了最小绝对收缩和选择算子(Lasso)模型和稀疏群Lasso (SG-Lasso)模型,推导了正则化参数的最大边际似然分布;此外,采用Moreau - Yoshida unadjusted Langevin算法(MYULA)实现目标后验采样解。由于后验分布难以求解,引入梯度投影法估计正则化参数。最后,利用自学习参数对图像进行优化。该算法不仅可以估计单个正则化项的参数,还可以估计多个正则化项的参数。针对先验中不可微的部分,采用MYULA计算不可微后验分布的次梯度。因此,即使正则化函数不可微,该算法也能自动学习参数。在实验部分,与人工调试的最优值相比,所提方法与最优值的误差在15%以内,并通过相变图(PTD)验证了算法的有效性。
{"title":"Auto-Learning of Parameters for High Resolution Sparse Group Lasso SAR Imagery","authors":"Wei Liu, Hanwen Xu, Cheng Fang, Lei Yang, Weidong Jiao","doi":"10.1109/IC-NIDC54101.2021.9660447","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660447","url":null,"abstract":"Aiming at the problem of adjusting the penalty term coefficient of feature enhancement in high-resolution synthetic aperture radar (SAR) imaging, a marginal estimation Bayes (MEB) algorithm is proposed, so that the prior features of the target can be fitted properly to improve the accuracy of image feature extraction. Firstly, the alternating direction method of multipliers (ADMM) convex optimization framework is modeled based on the echoed data, and least absolute shrinkage and selection operator (Lasso) model and sparse group Lasso (SG-Lasso) model are introduced, then the maximum marginal likelihood distribution of the regularization parameters is derived. Moreover, the Moreau Yoshida unadjusted Langevin algorithm (MYULA) is used to realize target posteriori sampling solution. Because the posterior distribution is difficult to solve, the gradient projection method is introduced to estimate the regularization parameters. Finally, auto-learning parameters are used to optimize the imaging. The proposed algorithm can not only estimate the parameters of a single regularization term, but also estimate the parameters of multiple regularization terms. Aiming at non-differentiable part in the prior, MYULA is adopted to calculate the subgradient of the non-differentiable posterior distribution. Therefore, the proposed algorithm is capable of auto-leaning parameters even regularization function is non-differentiable. In the experimental part, compared with the optimal value of manual debugging, the error between the proposed method and the optimal value is within 15%, and the effectiveness of the algorithm are verified by phase transition diagram (PTD).","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"276 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134366693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Zero-Shot Voice Cloning Using Variational Embedding with Attention Mechanism 基于注意力机制的变分嵌入零采样语音克隆
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660599
Jaeuk Lee, Jiye G. Kim, Joon‐Hyuk Chang
Many voice cloning studies based on multi-speaker text-to-speech (TTS) have been conducted. Among the techniques of voice cloning, we focus on zero-shot voice cloning. The most important aspect of zero-shot voice cloning is which speaker embedding is used. In this study, two types of speaker embeddings are used. One is extracted from the mel spectrogram using a speaker encoder and the other is stored in an embedding dictionary, such as a vector quantized-variational autoencoder (VQ-VAE). To extract embedding from the embedding dictionary, an attention mechanism is applied, which we call attention- V AE (AT - V AE). By employing the embedding extracted by the speaker encoder as a query in the attention mechanism, the attention weights are calculated in the embedding dictionary. This mechanism allows the extraction of speaker embedding, which represents unseen speakers. In addition, training is applied to make our model robust to unseen speakers. Through the training stage, our system has developed further. The performance of the proposed method was validated in terms of various metrics, and it was demonstrated that the proposed method enables voice cloning without adaptation training.
许多基于多说话人文本到语音(TTS)的语音克隆研究已经开展。在语音克隆技术中,我们重点研究了零采样语音克隆技术。零射击语音克隆最重要的方面是使用哪一个扬声器嵌入。在本研究中,使用了两种类型的说话人嵌入。一种是使用扬声器编码器从mel频谱图中提取,另一种是存储在嵌入字典中,例如矢量量化变分自编码器(VQ-VAE)。为了从嵌入字典中提取嵌入,采用了一种注意机制,我们称之为注意- V AE (AT - V AE)。利用说话人编码器提取的嵌入作为注意机制的查询,在嵌入字典中计算注意权值。这种机制允许提取说话人嵌入,它代表看不见的说话人。此外,还进行了训练,使我们的模型对未见的说话者具有鲁棒性。经过培训阶段,我们的系统得到了进一步的发展。从多个指标对所提方法的性能进行了验证,结果表明所提方法无需自适应训练即可实现语音克隆。
{"title":"Zero-Shot Voice Cloning Using Variational Embedding with Attention Mechanism","authors":"Jaeuk Lee, Jiye G. Kim, Joon‐Hyuk Chang","doi":"10.1109/IC-NIDC54101.2021.9660599","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660599","url":null,"abstract":"Many voice cloning studies based on multi-speaker text-to-speech (TTS) have been conducted. Among the techniques of voice cloning, we focus on zero-shot voice cloning. The most important aspect of zero-shot voice cloning is which speaker embedding is used. In this study, two types of speaker embeddings are used. One is extracted from the mel spectrogram using a speaker encoder and the other is stored in an embedding dictionary, such as a vector quantized-variational autoencoder (VQ-VAE). To extract embedding from the embedding dictionary, an attention mechanism is applied, which we call attention- V AE (AT - V AE). By employing the embedding extracted by the speaker encoder as a query in the attention mechanism, the attention weights are calculated in the embedding dictionary. This mechanism allows the extraction of speaker embedding, which represents unseen speakers. In addition, training is applied to make our model robust to unseen speakers. Through the training stage, our system has developed further. The performance of the proposed method was validated in terms of various metrics, and it was demonstrated that the proposed method enables voice cloning without adaptation training.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114319105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Combined Coverage, Attention and Pointer Networks for Improving Slot Filling in Spoken Language Understanding 综合覆盖、注意和指针网络提高口语理解槽填充
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660465
Yaping Wang, Huiqin Shao, Zhen Li, Yan Zhu, Zhe Liu
Sequence to sequence (Seq2Seq) model together with pointer network (Ptr-Net) has recently show promising results in slot filling task, in the situation where only sentence-level annotations are available, while the model's prediction contains repetition of slot values. In this paper, we add a coverage mechanism to alleviate issues of repeating prediction in slot filling task. We use a coverage vector to record attention history, and then add to the computation of attention, which can force model to consider more about un-predicted slot values. Experiments show that the proposed model significantly improves slot value prediction F1 with 8.5% relative improvement compare to the baseline models on benchmark DSTC2 (Dialog State Tracking Challenge 2) datasets.
序列到序列(Seq2Seq)模型和指针网络(Ptr-Net)最近在只有句子级注释可用的情况下,在槽填充任务中显示出有希望的结果,而模型的预测包含槽值的重复。在本文中,我们增加了一种覆盖机制来缓解补槽任务中重复预测的问题。我们使用覆盖向量来记录注意力历史,然后添加到注意力的计算中,这可以迫使模型更多地考虑不可预测的槽值。实验表明,与基准DSTC2(对话状态跟踪挑战2)数据集上的基线模型相比,该模型显著提高了槽值预测F1,相对提高了8.5%。
{"title":"Combined Coverage, Attention and Pointer Networks for Improving Slot Filling in Spoken Language Understanding","authors":"Yaping Wang, Huiqin Shao, Zhen Li, Yan Zhu, Zhe Liu","doi":"10.1109/IC-NIDC54101.2021.9660465","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660465","url":null,"abstract":"Sequence to sequence (Seq2Seq) model together with pointer network (Ptr-Net) has recently show promising results in slot filling task, in the situation where only sentence-level annotations are available, while the model's prediction contains repetition of slot values. In this paper, we add a coverage mechanism to alleviate issues of repeating prediction in slot filling task. We use a coverage vector to record attention history, and then add to the computation of attention, which can force model to consider more about un-predicted slot values. Experiments show that the proposed model significantly improves slot value prediction F1 with 8.5% relative improvement compare to the baseline models on benchmark DSTC2 (Dialog State Tracking Challenge 2) datasets.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132321008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WCD: A New Chinese Online Social Media Dataset for Clickbait Analysis and Detection WCD:一个新的中国在线社交媒体数据集,用于标题党分析和检测
Pub Date : 2021-11-17 DOI: 10.1109/IC-NIDC54101.2021.9660453
Tong Liu, K. Yu, Lu Wang, Xuanyu Zhang, Xiaofei Wu
In online social medias, there is a large amount of clickbait using various tricks such as curious words and well-designed sentence structures, to attract users to click on hyperlinks for unknown benefits. Clickbait detection aims to detect these hyperlinks through automated algorithms. Most of the previous clickbait datasets are based on English online social media corpus. Detection models based on these datasets usually cannot be well generalized to Chinese social media scenarios. In this paper, we construct a WeChat based Chinese clickbait dataset, i.e., WCD. Based on the WCD, we conduct a detailed analysis of the clickbait features from three aspects: behavior features, headline text features, and content text features. Finally, we use popular methods for clickbait detection on our dataset. We also respectively propose a machine learning detection model using feature fusion and a deep learning detection model combining headline semantic and POS tag information, both of which achieve excellent detection performance. The results of clickbait analysis and detection show that the dataset we constructed is of high quality.
在网络社交媒体中,有大量的标题党使用各种各样的技巧,如奇怪的词语和精心设计的句子结构,吸引用户点击超链接,以获得未知的好处。标题党检测旨在通过自动算法检测这些超链接。之前的大多数标题党数据集都是基于英语在线社交媒体语料库的。基于这些数据集的检测模型通常不能很好地推广到中国的社交媒体场景。在本文中,我们构建了一个基于微信的中文标题党数据集,即WCD。基于WCD,我们从行为特征、标题文本特征和内容文本特征三个方面对标题党特征进行了详细的分析。最后,我们在我们的数据集上使用流行的方法来检测标题党。我们还分别提出了一种基于特征融合的机器学习检测模型和一种结合标题语义和POS标签信息的深度学习检测模型,两者都取得了优异的检测性能。标题党分析和检测的结果表明,我们构建的数据集是高质量的。
{"title":"WCD: A New Chinese Online Social Media Dataset for Clickbait Analysis and Detection","authors":"Tong Liu, K. Yu, Lu Wang, Xuanyu Zhang, Xiaofei Wu","doi":"10.1109/IC-NIDC54101.2021.9660453","DOIUrl":"https://doi.org/10.1109/IC-NIDC54101.2021.9660453","url":null,"abstract":"In online social medias, there is a large amount of clickbait using various tricks such as curious words and well-designed sentence structures, to attract users to click on hyperlinks for unknown benefits. Clickbait detection aims to detect these hyperlinks through automated algorithms. Most of the previous clickbait datasets are based on English online social media corpus. Detection models based on these datasets usually cannot be well generalized to Chinese social media scenarios. In this paper, we construct a WeChat based Chinese clickbait dataset, i.e., WCD. Based on the WCD, we conduct a detailed analysis of the clickbait features from three aspects: behavior features, headline text features, and content text features. Finally, we use popular methods for clickbait detection on our dataset. We also respectively propose a machine learning detection model using feature fusion and a deep learning detection model combining headline semantic and POS tag information, both of which achieve excellent detection performance. The results of clickbait analysis and detection show that the dataset we constructed is of high quality.","PeriodicalId":264468,"journal":{"name":"2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114417224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2021 7th IEEE International Conference on Network Intelligence and Digital Content (IC-NIDC)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1