首页 > 最新文献

2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)最新文献

英文 中文
EsiNet: Enhanced Network Representation via Further Learning the Semantic Information of Edges EsiNet:通过进一步学习边缘的语义信息来增强网络表示
Anqing Zheng, Chong Feng, Fang Yang, Huanhuan Zhang
Network representation learning (NRL) is a crucial method to learn low-dimensional vertex representations to capture network information. However, conventional NRL models only regard each edge as a binary or continuous value while neglecting the rich semantic information on edges. To enhance network representation for Social Relation Extraction (SRE) task, we present a novel deep neural network based model, EsiNet, by learning the structure and semantic information of edges simultaneously. Compared with previous work, EsiNet focuses on further learning the interactions between vertices and capturing the correlations between labels. By jointly optimizing the objective function of these two components, EsiNet can preserve both the semantic and structural information of edges. Extensive experiments on several public datasets demonstrate that EsiNet outperforms other baselines significantly, by around 3% to 5% on hits@10 absolutely.
网络表示学习(NRL)是学习低维顶点表示以获取网络信息的关键方法。然而,传统的NRL模型只将每条边视为二值或连续值,而忽略了边上丰富的语义信息。为了提高社会关系提取(SRE)任务的网络表示能力,我们提出了一种基于深度神经网络的EsiNet模型,该模型通过同时学习边缘的结构和语义信息来实现。与以前的工作相比,EsiNet侧重于进一步学习顶点之间的相互作用和捕获标签之间的相关性。EsiNet通过联合优化这两部分的目标函数,可以同时保留边缘的语义信息和结构信息。在几个公共数据集上进行的大量实验表明,EsiNet的性能明显优于其他基准,在hits@10上绝对高出约3%至5%。
{"title":"EsiNet: Enhanced Network Representation via Further Learning the Semantic Information of Edges","authors":"Anqing Zheng, Chong Feng, Fang Yang, Huanhuan Zhang","doi":"10.1109/ICTAI.2019.00142","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00142","url":null,"abstract":"Network representation learning (NRL) is a crucial method to learn low-dimensional vertex representations to capture network information. However, conventional NRL models only regard each edge as a binary or continuous value while neglecting the rich semantic information on edges. To enhance network representation for Social Relation Extraction (SRE) task, we present a novel deep neural network based model, EsiNet, by learning the structure and semantic information of edges simultaneously. Compared with previous work, EsiNet focuses on further learning the interactions between vertices and capturing the correlations between labels. By jointly optimizing the objective function of these two components, EsiNet can preserve both the semantic and structural information of edges. Extensive experiments on several public datasets demonstrate that EsiNet outperforms other baselines significantly, by around 3% to 5% on hits@10 absolutely.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116551448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Mechanism Design: Compact and Decomposition Linear Programming Models 自动化机构设计:紧凑和分解线性规划模型
B. Jaumard, Kia Babashahi Ashtiani, Nicolas Huin
In the context of multi-agent systems, Automated Mechanism Design (AMD) is the computer-based design of the rules of a mechanism, which reaches an equilibrium despite the fact that agents can be selfish and lie about their preferences. Although it has been shown that AMD can be modelled as a linear program, it is with an exponential number of variables and consequently, there is no known efficient algorithm. We revisit the latter linear program model proposed for the AMD problem and introduce a new one with a polynomial number of variables. We show that the latter model corresponds to a Dantzig-Wolfe decomposition of the second one and design efficient solution schemes in polynomial time for both two models. Numerical experiments compare the solution efficiency of both models and show that we can solve very significantly larger data instances than before, up to 2,000 agents or 2,000 resources in about 35 seconds.
在多智能体系统背景下,自动机制设计(Automated Mechanism Design, AMD)是一种基于计算机的机制规则设计,尽管智能体可能是自私的,并对自己的偏好撒谎,但该机制仍能达到平衡。虽然已经证明AMD可以建模为线性程序,但它具有指数数量的变量,因此,没有已知的有效算法。我们回顾了针对AMD问题提出的后一种线性规划模型,并引入了一种具有多项式变量数的新模型。我们证明了后一个模型对应于第二个模型的dantzigg - wolfe分解,并设计了两个模型在多项式时间内的有效解方案。数值实验比较了两种模型的求解效率,并表明我们可以在大约35秒内解决比以前大得多的数据实例,最多可解决2,000个代理或2,000个资源。
{"title":"Automated Mechanism Design: Compact and Decomposition Linear Programming Models","authors":"B. Jaumard, Kia Babashahi Ashtiani, Nicolas Huin","doi":"10.1109/ICTAI.2019.00031","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00031","url":null,"abstract":"In the context of multi-agent systems, Automated Mechanism Design (AMD) is the computer-based design of the rules of a mechanism, which reaches an equilibrium despite the fact that agents can be selfish and lie about their preferences. Although it has been shown that AMD can be modelled as a linear program, it is with an exponential number of variables and consequently, there is no known efficient algorithm. We revisit the latter linear program model proposed for the AMD problem and introduce a new one with a polynomial number of variables. We show that the latter model corresponds to a Dantzig-Wolfe decomposition of the second one and design efficient solution schemes in polynomial time for both two models. Numerical experiments compare the solution efficiency of both models and show that we can solve very significantly larger data instances than before, up to 2,000 agents or 2,000 resources in about 35 seconds.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114854129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Adaptive Cross-Layer Sampling-Based Node Embedding for Multiplex Networks 基于自适应跨层采样的多路网络节点嵌入
Nianwen Ning, Chenguang Song, Pengpeng Zhou, Yunlei Zhang, Bin Wu
Network embedding aims to learn a latent representation of each node which preserves the structure information. Many real-world networks have multiple dimensions of nodes and multiple types of relations. Therefore, it is more appropriate to represent such kind of networks as multiplex networks. A multiplex network is formed by a set of nodes connected in different layers by links indicating interactions of different types. However, existing random walk based multiplex networks embedding algorithms have problems with sampling bias and imbalanced relation types, thus leading the poor performance in the downstream tasks. In this paper, we propose a node embedding method based on adaptive cross-layer forest fire sampling (FFS) for multiplex networks (FFME). We first focus on the sampling strategies of FFS to address the bias issue of random walk. We utilize a fixed-length queue to record previously visited layers, which can balance the edge distribution over different layers in sampled node sequences. In addition, to adaptively sample node's context, we also propose a metric for node called Neighbors Partition Coefficient (N P C ). The generation process of node sequence is supervised by NPC for adaptive cross-layer sampling. Experiments on real-world networks in diverse fields show that our method outperforms the state-of-the-art methods in application tasks such as cross-domain link prediction and shared community structure detection.
网络嵌入的目的是学习保留结构信息的每个节点的潜在表示。许多现实世界的网络都有多个维度的节点和多种类型的关系。因此,用多路网络来表示这类网络更为合适。多路复用网络是由一组节点通过不同类型的链路连接在不同的层中形成的。然而,现有的基于随机行走的多路网络嵌入算法存在抽样偏差和关系类型不平衡的问题,导致其在下游任务中的性能较差。提出了一种基于自适应跨层森林火灾采样(FFS)的多路网络节点嵌入方法。我们首先关注FFS的抽样策略,以解决随机漫步的偏差问题。我们利用固定长度的队列来记录之前访问过的层,这可以平衡采样节点序列中不同层的边缘分布。此外,为了对节点的上下文进行自适应采样,我们还提出了一个节点的邻居划分系数(N P C)度量。节点序列的生成过程由NPC监督,用于自适应跨层采样。在不同领域的真实网络上进行的实验表明,我们的方法在跨域链接预测和共享社区结构检测等应用任务中优于最先进的方法。
{"title":"An Adaptive Cross-Layer Sampling-Based Node Embedding for Multiplex Networks","authors":"Nianwen Ning, Chenguang Song, Pengpeng Zhou, Yunlei Zhang, Bin Wu","doi":"10.1109/ICTAI.2019.00216","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00216","url":null,"abstract":"Network embedding aims to learn a latent representation of each node which preserves the structure information. Many real-world networks have multiple dimensions of nodes and multiple types of relations. Therefore, it is more appropriate to represent such kind of networks as multiplex networks. A multiplex network is formed by a set of nodes connected in different layers by links indicating interactions of different types. However, existing random walk based multiplex networks embedding algorithms have problems with sampling bias and imbalanced relation types, thus leading the poor performance in the downstream tasks. In this paper, we propose a node embedding method based on adaptive cross-layer forest fire sampling (FFS) for multiplex networks (FFME). We first focus on the sampling strategies of FFS to address the bias issue of random walk. We utilize a fixed-length queue to record previously visited layers, which can balance the edge distribution over different layers in sampled node sequences. In addition, to adaptively sample node's context, we also propose a metric for node called Neighbors Partition Coefficient (N P C ). The generation process of node sequence is supervised by NPC for adaptive cross-layer sampling. Experiments on real-world networks in diverse fields show that our method outperforms the state-of-the-art methods in application tasks such as cross-domain link prediction and shared community structure detection.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"83 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116408188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Optimizing Training using Information Theory-Based Curriculum Learning Factory 利用信息论课程学习工厂优化培训
Henok Ghebrechristos, G. Alaghband
We present a new system that can automatically generate input paths (syllabus) for a convolutional neural network to follow through a curriculum learning to improve training performance. Our system utilizes information-theoretic content measures of training samples to form syllabus at training time. We treat every sample as 2D random variable where a data point contained in the sample (such as a pixel) is modelled as an independent and identically distributed random variable (i.i.d) realization. We use several information theory methods to rank and determine when a sample is fed to a network by measuring its pixel composition and its relationship to other samples in the training set. Comparative evaluation of multiple state-of-the-art networks, including, GoogleNet, and VGG, on benchmark datasets demonstrate a syllabus that ranks samples using measures such as Joint Entropy between adjacent samples, can improve learning and significantly reduce the amount of training steps required to achieve desirable training accuracy. We present results that indicate our approach can reduce training loss by as much as a factor of 9 compared to conventional training.
我们提出了一个新的系统,它可以自动生成输入路径(教学大纲),让卷积神经网络通过课程学习来提高训练性能。本系统利用训练样本的信息论内容测度,在训练时形成教学大纲。我们将每个样本视为二维随机变量,其中样本中包含的数据点(例如像素)被建模为独立且同分布的随机变量(i.i.d)实现。我们使用几种信息论方法通过测量样本的像素组成及其与训练集中其他样本的关系来对样本进行排序和确定何时将样本馈送到网络中。在基准数据集上对多个最先进的网络(包括GoogleNet和VGG)进行比较评估,证明了使用相邻样本之间的联合熵等度量对样本进行排序的教学大纲,可以提高学习效果,并显著减少达到理想训练精度所需的训练步骤。我们提出的结果表明,与传统训练相比,我们的方法可以减少多达9倍的训练损失。
{"title":"Optimizing Training using Information Theory-Based Curriculum Learning Factory","authors":"Henok Ghebrechristos, G. Alaghband","doi":"10.1109/ICTAI.2019.00218","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00218","url":null,"abstract":"We present a new system that can automatically generate input paths (syllabus) for a convolutional neural network to follow through a curriculum learning to improve training performance. Our system utilizes information-theoretic content measures of training samples to form syllabus at training time. We treat every sample as 2D random variable where a data point contained in the sample (such as a pixel) is modelled as an independent and identically distributed random variable (i.i.d) realization. We use several information theory methods to rank and determine when a sample is fed to a network by measuring its pixel composition and its relationship to other samples in the training set. Comparative evaluation of multiple state-of-the-art networks, including, GoogleNet, and VGG, on benchmark datasets demonstrate a syllabus that ranks samples using measures such as Joint Entropy between adjacent samples, can improve learning and significantly reduce the amount of training steps required to achieve desirable training accuracy. We present results that indicate our approach can reduce training loss by as much as a factor of 9 compared to conventional training.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115389060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
ECPNet: An Efficient Attention-Based Convolution Network with Pseudo-3D Block for Human Action Recognition ECPNet:一种有效的基于注意力的伪三维块卷积网络,用于人体动作识别
Xiuping Bao, Jiabin Yuan, Bei Chen
Human action recognition has became an important task in computer vision and has received a significant amount of research interests in recent years. Convolutional Neural Network (CNN) has shown its power in image recognition task. While in the field of video recognition, it is still a challenge problem. In this paper, we introduce a high-efficient attention-based convolutional network named ECPNet for video understanding. ECPNet adopts the framework that is a consecutive connection of 2D CNN and pseudo-3D CNN. The pseudo-3D means we replace the traditional 3 × 3 × 3 kernel with two 3D convolutional filters shaped 1 × 3 × 3 and 3 × 1 × 1. Our ECPNet combines the advantages of both 2D and 3D CNNs: (1) ECPNet is an end-to-end network and can learn information of appearance from images and motion between frames. (2) ECPNet requires less computing resource and lower memory consumption than many state-of-art models. (3) ECPNet is easy to expand for different demands of runtime and classification accuracy. We evaluate the proposed model on three popular video benchmarks in human action recognition task: Kinetics-mini (split of full Kinetics), UCF101 and HMDB51. Our ECPNet achieves the excellent performance on above datasets with less time cost.
人体动作识别已成为计算机视觉领域的一项重要研究课题,近年来受到了广泛的关注。卷积神经网络(CNN)在图像识别任务中已经显示出其强大的能力。而在视频识别领域,这仍然是一个具有挑战性的问题。本文介绍了一种高效的基于注意力的卷积网络ECPNet,用于视频理解。ECPNet采用的框架是二维CNN和伪三维CNN的连续连接。伪3D意味着我们用两个形状为1 × 3 × 3和3 × 1 × 1的三维卷积滤波器取代传统的3 × 3 × 3核。我们的ECPNet结合了2D和3D cnn的优点:(1)ECPNet是一个端到端的网络,可以从图像和帧之间的运动中学习外观信息。(2)与许多最先进的模型相比,ECPNet需要更少的计算资源和更低的内存消耗。(3) ECPNet易于扩展以满足不同的运行时间和分类精度需求。我们在人类动作识别任务中三个流行的视频基准上对所提出的模型进行了评估:Kinetics-mini (full Kinetics的分裂),UCF101和HMDB51。我们的ECPNet以较少的时间成本在上述数据集上实现了优异的性能。
{"title":"ECPNet: An Efficient Attention-Based Convolution Network with Pseudo-3D Block for Human Action Recognition","authors":"Xiuping Bao, Jiabin Yuan, Bei Chen","doi":"10.1109/ICTAI.2019.00089","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00089","url":null,"abstract":"Human action recognition has became an important task in computer vision and has received a significant amount of research interests in recent years. Convolutional Neural Network (CNN) has shown its power in image recognition task. While in the field of video recognition, it is still a challenge problem. In this paper, we introduce a high-efficient attention-based convolutional network named ECPNet for video understanding. ECPNet adopts the framework that is a consecutive connection of 2D CNN and pseudo-3D CNN. The pseudo-3D means we replace the traditional 3 × 3 × 3 kernel with two 3D convolutional filters shaped 1 × 3 × 3 and 3 × 1 × 1. Our ECPNet combines the advantages of both 2D and 3D CNNs: (1) ECPNet is an end-to-end network and can learn information of appearance from images and motion between frames. (2) ECPNet requires less computing resource and lower memory consumption than many state-of-art models. (3) ECPNet is easy to expand for different demands of runtime and classification accuracy. We evaluate the proposed model on three popular video benchmarks in human action recognition task: Kinetics-mini (split of full Kinetics), UCF101 and HMDB51. Our ECPNet achieves the excellent performance on above datasets with less time cost.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121404601","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking Symbolic Execution Using Constraint Problems - Initial Results 使用约束问题对符号执行进行基准测试-初步结果
Sahil Verma, R. Yap
Symbolic execution is a powerful technique for bug finding and program testing. It is successful in finding bugs in real-world code. The core reasoning techniques use constraint solving, path exploration, and search, which are also the same techniques used in solving combinatorial problems, e.g., finite-domain constraint satisfaction problems (CSPs). We propose CSP instances as more challenging benchmarks to evaluate the effectiveness of the core techniques in symbolic execution. We transform CSP benchmarks into C programs suitable for testing the reasoning capabilities of symbolic execution tools. From a single CSP P, we transform P depending on transformation choice into different C programs. Preliminary testing with the KLEE, Tracer-X, and LLBMC tools show substantial runtime differences from transformation and solver choice. Our C benchmarks are effective in showing the limitations of existing symbolic execution tools. The motivation for this work is we believe that benchmarks of this form can spur the development and engineering of improved core reasoning in symbolic execution engines.
符号执行对于bug查找和程序测试来说是一种强大的技术。它在查找真实代码中的bug方面是成功的。核心推理技术使用约束求解、路径探索和搜索,这也是用于解决组合问题的相同技术,例如有限域约束满足问题(csp)。我们建议将CSP实例作为更具挑战性的基准来评估符号执行中核心技术的有效性。我们将CSP基准转换为适合于测试符号执行工具推理能力的C程序。从单个CSP P,根据变换选择将P变换成不同的C程序。使用KLEE、Tracer-X和LLBMC工具进行的初步测试显示,转换和求解器的选择在运行时存在很大差异。我们的C基准测试有效地展示了现有符号执行工具的局限性。这项工作的动机是我们相信这种形式的基准测试可以刺激符号执行引擎中改进的核心推理的开发和工程。
{"title":"Benchmarking Symbolic Execution Using Constraint Problems - Initial Results","authors":"Sahil Verma, R. Yap","doi":"10.1109/ICTAI.2019.00010","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00010","url":null,"abstract":"Symbolic execution is a powerful technique for bug finding and program testing. It is successful in finding bugs in real-world code. The core reasoning techniques use constraint solving, path exploration, and search, which are also the same techniques used in solving combinatorial problems, e.g., finite-domain constraint satisfaction problems (CSPs). We propose CSP instances as more challenging benchmarks to evaluate the effectiveness of the core techniques in symbolic execution. We transform CSP benchmarks into C programs suitable for testing the reasoning capabilities of symbolic execution tools. From a single CSP P, we transform P depending on transformation choice into different C programs. Preliminary testing with the KLEE, Tracer-X, and LLBMC tools show substantial runtime differences from transformation and solver choice. Our C benchmarks are effective in showing the limitations of existing symbolic execution tools. The motivation for this work is we believe that benchmarks of this form can spur the development and engineering of improved core reasoning in symbolic execution engines.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123310765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Rethink Gaussian Denoising Prior for Real-World Image Denoising 重新考虑高斯去噪先验在真实世界图像去噪中的应用
Tianyang Wang, Jun Huan, Bo Li, Kaoning Hu
Real-world image denoising is a challenging but significant problem in computer vision. Unlike Gaussian denoising on which most existing methods focus, the real-world noise is nonadditive, and the distributions are difficult to model. This leads to unsatisfactory performance when applying a Gaussian denoising approach on real-world denoising problems. In this paper, we propose a simple framework for effective real-world image denoising. Specifically, we investigate the intrinsic properties of the Gaussian denoising prior and demonstrate this prior can aid real-world image denoising. To leverage this prior, we fine-tune it for only one epoch on a recently proposed real-world image denoising dataset, and the learned model can enhance both visual and quantitative results (peak-signal-noise-ratio) for real-world image denoising tasks. Extensive experiments demonstrate the effectiveness of our approach, and indicate that the Gaussian denoising prior can be also transferred to real-world image denoising by exploiting appropriate training schemes.
在计算机视觉中,真实图像去噪是一个具有挑战性但又很重要的问题。与大多数现有方法关注的高斯去噪不同,现实世界的噪声是非加性的,分布难以建模。当将高斯去噪方法应用于现实世界的去噪问题时,这会导致不满意的性能。在本文中,我们提出了一个简单的框架,有效的现实世界的图像去噪。具体来说,我们研究了高斯去噪先验的内在特性,并证明了这种先验可以帮助现实世界的图像去噪。为了利用这一先验,我们在最近提出的真实世界图像去噪数据集上仅对其进行了一个epoch的微调,并且学习的模型可以增强真实世界图像去噪任务的视觉和定量结果(峰值信噪比)。大量的实验证明了该方法的有效性,并表明利用适当的训练方案也可以将高斯去噪先验转移到现实世界的图像去噪中。
{"title":"Rethink Gaussian Denoising Prior for Real-World Image Denoising","authors":"Tianyang Wang, Jun Huan, Bo Li, Kaoning Hu","doi":"10.1109/ICTAI.2019.00244","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00244","url":null,"abstract":"Real-world image denoising is a challenging but significant problem in computer vision. Unlike Gaussian denoising on which most existing methods focus, the real-world noise is nonadditive, and the distributions are difficult to model. This leads to unsatisfactory performance when applying a Gaussian denoising approach on real-world denoising problems. In this paper, we propose a simple framework for effective real-world image denoising. Specifically, we investigate the intrinsic properties of the Gaussian denoising prior and demonstrate this prior can aid real-world image denoising. To leverage this prior, we fine-tune it for only one epoch on a recently proposed real-world image denoising dataset, and the learned model can enhance both visual and quantitative results (peak-signal-noise-ratio) for real-world image denoising tasks. Extensive experiments demonstrate the effectiveness of our approach, and indicate that the Gaussian denoising prior can be also transferred to real-world image denoising by exploiting appropriate training schemes.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125176766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Sparse High-Level Attention Networks for Person Re-Identification 稀疏高层注意网络用于人物再识别
Sheng Xie, Canlong Zhang, Zhixin Li, Zhiwen Wang
When extracting convolutional features from person images with low resolution, a large amount of available information will be lost due to the pooling, which will lead to the reduction of the accuracy of person classification models. This paper proposes a new classification model, which can effectively to reduce the loss of important information about the convolutional neural works. Firstly, the SE module in the Squeeze-and-Excitation Networks (SENet) is extracted and normalized to generate the Normalized Squeeze-and-Excitation (NSE) module. Then, 4 NSE modules are applied to the convolutional layers of ResNet. Finally, a Sparse Normalized Squeeze-and-Excitation Network (SNSENet) is constructed by adding 4 shortcut connections between the convolutional layers. The experimental results of Market-1501 show that the rank-1 of SNSE-ResNet-50 is 3.7% and 4.2% higher than that of SE-ResNet-50 and ResNet-50 respectively, it has done well in other person re-identification datasets.
在从低分辨率的人物图像中提取卷积特征时,由于池化会丢失大量的可用信息,从而导致人物分类模型的精度降低。本文提出了一种新的分类模型,该模型可以有效地减少卷积神经网络中重要信息的丢失。首先,对压缩激励网络(SENet)中的SE模块进行提取和归一化,生成归一化压缩激励(NSE)模块。然后,将4个NSE模块应用于ResNet的卷积层。最后,通过在卷积层之间增加4个快捷连接,构造了稀疏归一化压缩激励网络(SNSENet)。Market-1501的实验结果表明,SNSE-ResNet-50的rank-1分别比SE-ResNet-50和ResNet-50高3.7%和4.2%,在其他人员再识别数据集中表现良好。
{"title":"Sparse High-Level Attention Networks for Person Re-Identification","authors":"Sheng Xie, Canlong Zhang, Zhixin Li, Zhiwen Wang","doi":"10.1109/ICTAI.2019.00213","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00213","url":null,"abstract":"When extracting convolutional features from person images with low resolution, a large amount of available information will be lost due to the pooling, which will lead to the reduction of the accuracy of person classification models. This paper proposes a new classification model, which can effectively to reduce the loss of important information about the convolutional neural works. Firstly, the SE module in the Squeeze-and-Excitation Networks (SENet) is extracted and normalized to generate the Normalized Squeeze-and-Excitation (NSE) module. Then, 4 NSE modules are applied to the convolutional layers of ResNet. Finally, a Sparse Normalized Squeeze-and-Excitation Network (SNSENet) is constructed by adding 4 shortcut connections between the convolutional layers. The experimental results of Market-1501 show that the rank-1 of SNSE-ResNet-50 is 3.7% and 4.2% higher than that of SE-ResNet-50 and ResNet-50 respectively, it has done well in other person re-identification datasets.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122775297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Recovering Extremely Degraded Faces by Joint Super-Resolution and Facial Composite 联合超分辨率和人脸合成技术恢复极度退化的人脸
Xiu Li, Guichun Duan, Zhouxia Wang, Jimmy S. J. Ren, Yongbing Zhang, Jiawei Zhang, Kaixiang Song
In the past a few years, we witnessed rapid advancement in face super-resolution from very low resolution(VLR) images. However, most of the previous studies focus on solving such problem without explicitly considering the impact of severe real-life image degradation (e.g. blur and noise). We can show that robustly recover details from VLR images is a task beyond the ability of current state-of-the-art method. In this paper, we borrow ideas from "facial composite" and propose an alternative approach to tackle this problem. We endow the degraded VLR images with additional cues by integrating existing face components from multiple reference images into a novel learning pipeline with both low level and high level semantic loss function as well as a specialized adversarial based training scheme. We show that our method is able to effectively and robustly restore relevant facial details from 16x16 images with extreme degradation. We also tested our approach against real-life images and our method performs favorably against previous methods.
在过去的几年中,我们见证了极低分辨率(VLR)图像在人脸超分辨率方面的快速发展。然而,以往的研究大多侧重于解决这一问题,而没有明确考虑现实生活中严重的图像退化(如模糊和噪声)的影响。我们可以证明,从VLR图像中稳健地恢复细节是一项超出当前最先进方法能力的任务。在本文中,我们借鉴了“面部复合材料”的思想,并提出了一种解决这一问题的替代方法。我们通过将来自多个参考图像的现有面部成分整合到一个具有低水平和高水平语义损失函数以及专门的基于对抗的训练方案的新型学习管道中,从而赋予退化的VLR图像额外的线索。我们的研究表明,我们的方法能够有效地、鲁棒地从极度退化的16x16图像中恢复相关的面部细节。我们还针对真实图像测试了我们的方法,与之前的方法相比,我们的方法表现得更好。
{"title":"Recovering Extremely Degraded Faces by Joint Super-Resolution and Facial Composite","authors":"Xiu Li, Guichun Duan, Zhouxia Wang, Jimmy S. J. Ren, Yongbing Zhang, Jiawei Zhang, Kaixiang Song","doi":"10.1109/ICTAI.2019.00079","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00079","url":null,"abstract":"In the past a few years, we witnessed rapid advancement in face super-resolution from very low resolution(VLR) images. However, most of the previous studies focus on solving such problem without explicitly considering the impact of severe real-life image degradation (e.g. blur and noise). We can show that robustly recover details from VLR images is a task beyond the ability of current state-of-the-art method. In this paper, we borrow ideas from \"facial composite\" and propose an alternative approach to tackle this problem. We endow the degraded VLR images with additional cues by integrating existing face components from multiple reference images into a novel learning pipeline with both low level and high level semantic loss function as well as a specialized adversarial based training scheme. We show that our method is able to effectively and robustly restore relevant facial details from 16x16 images with extreme degradation. We also tested our approach against real-life images and our method performs favorably against previous methods.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128787724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
ICTAI 2019 Conference Committees ICTAI 2019会议委员会
{"title":"ICTAI 2019 Conference Committees","authors":"","doi":"10.1109/ictai.2019.00007","DOIUrl":"https://doi.org/10.1109/ictai.2019.00007","url":null,"abstract":"","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129353693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1