首页 > 最新文献

ACM Transactions on Knowledge Discovery from Data (TKDD)最新文献

英文 中文
Lost in Transduction: Transductive Transfer Learning in Text Classification 迷失在转导中:文本分类中的转导迁移学习
Pub Date : 2021-07-03 DOI: 10.1145/3453146
A. Moreo, Andrea Esuli, F. Sebastiani
Obtaining high-quality labelled data for training a classifier in a new application domain is often costly. Transfer Learning (a.k.a. “Inductive Transfer”) tries to alleviate these costs by transferring, to the “target” domain of interest, knowledge available from a different “source” domain. In transfer learning the lack of labelled information from the target domain is compensated by the availability at training time of a set of unlabelled examples from the target distribution. Transductive Transfer Learning denotes the transfer learning setting in which the only set of target documents that we are interested in classifying is known and available at training time. Although this definition is indeed in line with Vapnik’s original definition of “transduction”, current terminology in the field is confused. In this article, we discuss how the term “transduction” has been misused in the transfer learning literature, and propose a clarification consistent with the original characterization of this term given by Vapnik. We go on to observe that the above terminology misuse has brought about misleading experimental comparisons, with inductive transfer learning methods that have been incorrectly compared with transductive transfer learning methods. We then, give empirical evidence that the difference in performance between the inductive version and the transductive version of a transfer learning method can indeed be statistically significant (i.e., that knowing at training time the only data one needs to classify indeed gives an advantage). Our clarification allows a reassessment of the field, and of the relative merits of the major, state-of-the-art algorithms for transfer learning in text classification.
获得高质量的标记数据来训练一个新的应用领域的分类器通常是昂贵的。迁移学习(又称“归纳迁移”)试图通过将来自不同“源”领域的知识转移到感兴趣的“目标”领域来减轻这些成本。在迁移学习中,来自目标域的标记信息的缺乏由训练时来自目标分布的一组未标记示例的可用性来补偿。转导迁移学习指的是一种迁移学习设置,在这种设置中,我们感兴趣分类的唯一目标文档集是已知的,并且在训练时是可用的。虽然这个定义确实符合Vapnik最初对“转导”的定义,但目前该领域的术语是混乱的。在本文中,我们讨论了术语“转导”在迁移学习文献中是如何被误用的,并提出了与Vapnik对该术语的原始描述一致的澄清。我们继续观察到,上述术语的误用带来了误导性的实验比较,归纳迁移学习方法被错误地与传导迁移学习方法进行了比较。然后,我们给出经验证据,证明迁移学习方法的归纳版本和转换版本之间的性能差异确实可以在统计上显着(即,在训练时知道唯一需要分类的数据确实具有优势)。我们的澄清允许对该领域进行重新评估,以及对文本分类中迁移学习的主要、最先进算法的相对优点进行重新评估。
{"title":"Lost in Transduction: Transductive Transfer Learning in Text Classification","authors":"A. Moreo, Andrea Esuli, F. Sebastiani","doi":"10.1145/3453146","DOIUrl":"https://doi.org/10.1145/3453146","url":null,"abstract":"Obtaining high-quality labelled data for training a classifier in a new application domain is often costly. Transfer Learning (a.k.a. “Inductive Transfer”) tries to alleviate these costs by transferring, to the “target” domain of interest, knowledge available from a different “source” domain. In transfer learning the lack of labelled information from the target domain is compensated by the availability at training time of a set of unlabelled examples from the target distribution. Transductive Transfer Learning denotes the transfer learning setting in which the only set of target documents that we are interested in classifying is known and available at training time. Although this definition is indeed in line with Vapnik’s original definition of “transduction”, current terminology in the field is confused. In this article, we discuss how the term “transduction” has been misused in the transfer learning literature, and propose a clarification consistent with the original characterization of this term given by Vapnik. We go on to observe that the above terminology misuse has brought about misleading experimental comparisons, with inductive transfer learning methods that have been incorrectly compared with transductive transfer learning methods. We then, give empirical evidence that the difference in performance between the inductive version and the transductive version of a transfer learning method can indeed be statistically significant (i.e., that knowing at training time the only data one needs to classify indeed gives an advantage). Our clarification allows a reassessment of the field, and of the relative merits of the major, state-of-the-art algorithms for transfer learning in text classification.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123999470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Cross-domain Recommendation with Bridge-Item Embeddings 桥项嵌入的跨域推荐
Pub Date : 2021-07-03 DOI: 10.1145/3447683
Chen Gao, Yong Li, Fuli Feng, Xiangning Chen, Kai Zhao, Xiangnan He, Depeng Jin
Web systems that provide the same functionality usually share a certain amount of items. This makes it possible to combine data from different websites to improve recommendation quality, known as the cross-domain recommendation task. Despite many research efforts on this task, the main drawback is that they largely assume the data of different systems can be fully shared. Such an assumption is unrealistic different systems are typically operated by different companies, and it may violate business privacy policy to directly share user behavior data since it is highly sensitive. In this work, we consider a more practical scenario to perform cross-domain recommendation. To avoid the leak of user privacy during the data sharing process, we consider sharing only the information of the item side, rather than user behavior data. Specifically, we transfer the item embeddings across domains, making it easier for two companies to reach a consensus (e.g., legal policy) on data sharing since the data to be shared is user-irrelevant and has no explicit semantics. To distill useful signals from transferred item embeddings, we rely on the strong representation power of neural networks and develop a new method named as NATR (short for Neural Attentive Transfer Recommendation). We perform extensive experiments on two real-world datasets, demonstrating that NATR achieves similar or even better performance than traditional cross-domain recommendation methods that directly share user-relevant data. Further insights are provided on the efficacy of NATR in using the transferred item embeddings to alleviate the data sparsity issue.
提供相同功能的Web系统通常共享一定数量的项。这使得整合来自不同网站的数据来提高推荐质量成为可能,被称为跨域推荐任务。尽管在这项任务上进行了许多研究,但主要的缺点是它们在很大程度上假设不同系统的数据可以完全共享。这样的假设是不现实的,不同的系统通常由不同的公司运营,直接共享用户行为数据可能会违反商业隐私政策,因为它是高度敏感的。在这项工作中,我们考虑了一个更实际的场景来执行跨领域推荐。为了避免在数据共享过程中泄露用户隐私,我们考虑只共享项目方信息,而不共享用户行为数据。具体来说,我们跨领域转移项目嵌入,使两家公司更容易就数据共享达成共识(例如,法律政策),因为要共享的数据与用户无关,没有明确的语义。为了从转移的项目嵌入中提取有用的信号,我们依靠神经网络的强大表征能力,开发了一种新的方法,称为NATR (neural attention Transfer Recommendation的缩写)。我们在两个真实世界的数据集上进行了大量的实验,证明NATR比直接共享用户相关数据的传统跨领域推荐方法实现了类似甚至更好的性能。在使用转移的项目嵌入来缓解数据稀疏性问题方面,提供了NATR的有效性的进一步见解。
{"title":"Cross-domain Recommendation with Bridge-Item Embeddings","authors":"Chen Gao, Yong Li, Fuli Feng, Xiangning Chen, Kai Zhao, Xiangnan He, Depeng Jin","doi":"10.1145/3447683","DOIUrl":"https://doi.org/10.1145/3447683","url":null,"abstract":"Web systems that provide the same functionality usually share a certain amount of items. This makes it possible to combine data from different websites to improve recommendation quality, known as the cross-domain recommendation task. Despite many research efforts on this task, the main drawback is that they largely assume the data of different systems can be fully shared. Such an assumption is unrealistic different systems are typically operated by different companies, and it may violate business privacy policy to directly share user behavior data since it is highly sensitive. In this work, we consider a more practical scenario to perform cross-domain recommendation. To avoid the leak of user privacy during the data sharing process, we consider sharing only the information of the item side, rather than user behavior data. Specifically, we transfer the item embeddings across domains, making it easier for two companies to reach a consensus (e.g., legal policy) on data sharing since the data to be shared is user-irrelevant and has no explicit semantics. To distill useful signals from transferred item embeddings, we rely on the strong representation power of neural networks and develop a new method named as NATR (short for Neural Attentive Transfer Recommendation). We perform extensive experiments on two real-world datasets, demonstrating that NATR achieves similar or even better performance than traditional cross-domain recommendation methods that directly share user-relevant data. Further insights are provided on the efficacy of NATR in using the transferred item embeddings to alleviate the data sparsity issue.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"170 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121432239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A Latent Variable Augmentation Method for Image Categorization with Insufficient Training Samples 一种训练样本不足的图像分类潜变量增强方法
Pub Date : 2021-07-03 DOI: 10.1145/3451165
Luyue Lin, Xin Zheng, Bo Liu, Wei Chen, Yanshan Xiao
Over the past few years, we have made great progress in image categorization based on convolutional neural networks (CNNs). These CNNs are always trained based on a large-scale image data set; however, people may only have limited training samples for training CNN in the real-world applications. To solve this problem, one intuition is augmenting training samples. In this article, we propose an algorithm called Lavagan (Latent Variables Augmentation Method based on Generative Adversarial Nets) to improve the performance of CNN with insufficient training samples. The proposed Lavagan method is mainly composed of two tasks. The first task is that we augment a number latent variables (LVs) from a set of adaptive and constrained LVs distributions. In the second task, we take the augmented LVs into the training procedure of the image classifier. By taking these two tasks into account, we propose a uniform objective function to incorporate the two tasks into the learning. We then put forward an alternative two-play minimization game to minimize this uniform loss function such that we can obtain the predictive classifier. Moreover, based on Hoeffding’s Inequality and Chernoff Bounding method, we analyze the feasibility and efficiency of the proposed Lavagan method, which manifests that the LV augmentation method is able to improve the performance of Lavagan with insufficient training samples. Finally, the experiment has shown that the proposed Lavagan method is able to deliver more accurate performance than the existing state-of-the-art methods.
在过去的几年里,我们在基于卷积神经网络(cnn)的图像分类方面取得了很大的进展。这些cnn总是基于大规模的图像数据集进行训练;然而,人们可能只有有限的训练样本来训练CNN在现实世界的应用。为了解决这个问题,一种直觉是增加训练样本。在本文中,我们提出了一种名为Lavagan (Latent Variables Augmentation Method based on Generative Adversarial Nets)的算法来改善训练样本不足的CNN的性能。提出的Lavagan方法主要由两个任务组成。第一个任务是我们从一组自适应和约束的潜在变量分布中增加一些潜在变量(lv)。在第二个任务中,我们将增强lv引入到图像分类器的训练过程中。考虑到这两个任务,我们提出了一个统一的目标函数,将这两个任务合并到学习中。然后,我们提出了一种备选的两局最小化对策,以最小化该均匀损失函数,从而获得预测分类器。此外,基于Hoeffding不等式和Chernoff边界法,我们分析了所提出的Lavagan方法的可行性和效率,表明LV增强方法能够在训练样本不足的情况下提高Lavagan的性能。最后,实验表明,所提出的Lavagan方法能够提供比现有的最先进的方法更准确的性能。
{"title":"A Latent Variable Augmentation Method for Image Categorization with Insufficient Training Samples","authors":"Luyue Lin, Xin Zheng, Bo Liu, Wei Chen, Yanshan Xiao","doi":"10.1145/3451165","DOIUrl":"https://doi.org/10.1145/3451165","url":null,"abstract":"Over the past few years, we have made great progress in image categorization based on convolutional neural networks (CNNs). These CNNs are always trained based on a large-scale image data set; however, people may only have limited training samples for training CNN in the real-world applications. To solve this problem, one intuition is augmenting training samples. In this article, we propose an algorithm called Lavagan (Latent Variables Augmentation Method based on Generative Adversarial Nets) to improve the performance of CNN with insufficient training samples. The proposed Lavagan method is mainly composed of two tasks. The first task is that we augment a number latent variables (LVs) from a set of adaptive and constrained LVs distributions. In the second task, we take the augmented LVs into the training procedure of the image classifier. By taking these two tasks into account, we propose a uniform objective function to incorporate the two tasks into the learning. We then put forward an alternative two-play minimization game to minimize this uniform loss function such that we can obtain the predictive classifier. Moreover, based on Hoeffding’s Inequality and Chernoff Bounding method, we analyze the feasibility and efficiency of the proposed Lavagan method, which manifests that the LV augmentation method is able to improve the performance of Lavagan with insufficient training samples. Finally, the experiment has shown that the proposed Lavagan method is able to deliver more accurate performance than the existing state-of-the-art methods.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133767866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DiMBERT: Learning Vision-Language Grounded Representations with Disentangled Multimodal-Attention 用解纠缠多模态注意学习基于视觉语言的表征
Pub Date : 2021-07-03 DOI: 10.1145/3447685
Fenglin Liu, Xuancheng Ren
Vision-and-language (V-L) tasks require the system to understand both vision content and natural language, thus learning fine-grained joint representations of vision and language (a.k.a. V-L representations) is of paramount importance. Recently, various pre-trained V-L models are proposed to learn V-L representations and achieve improved results in many tasks. However, the mainstream models process both vision and language inputs with the same set of attention matrices. As a result, the generated V-L representations are entangled in one common latent space. To tackle this problem, we propose DiMBERT (short for Disentangled Multimodal-Attention BERT), which is a novel framework that applies separated attention spaces for vision and language, and the representations of multi-modalities can thus be disentangled explicitly. To enhance the correlation between vision and language in disentangled spaces, we introduce the visual concepts to DiMBERT which represent visual information in textual format. In this manner, visual concepts help to bridge the gap between the two modalities. We pre-train DiMBERT on a large amount of image–sentence pairs on two tasks: bidirectional language modeling and sequence-to-sequence language modeling. After pre-train, DiMBERT is further fine-tuned for the downstream tasks. Experiments show that DiMBERT sets new state-of-the-art performance on three tasks (over four datasets), including both generation tasks (image captioning and visual storytelling) and classification tasks (referring expressions). The proposed DiM (short for Disentangled Multimodal-Attention) module can be easily incorporated into existing pre-trained V-L models to boost their performance, up to a 5% increase on the representative task. Finally, we conduct a systematic analysis and demonstrate the effectiveness of our DiM and the introduced visual concepts.
视觉和语言(V-L)任务要求系统同时理解视觉内容和自然语言,因此学习视觉和语言的细粒度联合表示(又称V-L表示)至关重要。最近,人们提出了各种预训练的V-L模型来学习V-L表示,并在许多任务中取得了改进的结果。然而,主流模型同时处理视觉和语言输入,使用同一组注意矩阵。因此,生成的V-L表示在一个公共潜在空间中纠缠。为了解决这个问题,我们提出了一个新的框架DiMBERT (Disentangled Multimodal-Attention BERT的缩写),它将视觉和语言的注意空间分开,从而可以明确地解开多模态的表征。为了增强视觉和语言在非纠缠空间中的相关性,我们在DiMBERT中引入了以文本形式表示视觉信息的视觉概念。通过这种方式,视觉概念有助于弥合两种模式之间的差距。我们在大量的图像-句子对上对DiMBERT进行了两个任务的预训练:双向语言建模和序列-序列语言建模。在预训练之后,DiMBERT将进一步针对下游任务进行微调。实验表明,DiMBERT在三个任务(超过四个数据集)上设置了新的最先进的性能,包括生成任务(图像字幕和视觉故事)和分类任务(引用表达式)。提出的DiM (Disentangled Multimodal-Attention的缩写)模块可以很容易地整合到现有的预训练V-L模型中,以提高它们的性能,在代表性任务上最多可提高5%。最后,我们进行了系统的分析,并证明了我们的DiM和引入的视觉概念的有效性。
{"title":"DiMBERT: Learning Vision-Language Grounded Representations with Disentangled Multimodal-Attention","authors":"Fenglin Liu, Xuancheng Ren","doi":"10.1145/3447685","DOIUrl":"https://doi.org/10.1145/3447685","url":null,"abstract":"Vision-and-language (V-L) tasks require the system to understand both vision content and natural language, thus learning fine-grained joint representations of vision and language (a.k.a. V-L representations) is of paramount importance. Recently, various pre-trained V-L models are proposed to learn V-L representations and achieve improved results in many tasks. However, the mainstream models process both vision and language inputs with the same set of attention matrices. As a result, the generated V-L representations are entangled in one common latent space. To tackle this problem, we propose DiMBERT (short for Disentangled Multimodal-Attention BERT), which is a novel framework that applies separated attention spaces for vision and language, and the representations of multi-modalities can thus be disentangled explicitly. To enhance the correlation between vision and language in disentangled spaces, we introduce the visual concepts to DiMBERT which represent visual information in textual format. In this manner, visual concepts help to bridge the gap between the two modalities. We pre-train DiMBERT on a large amount of image–sentence pairs on two tasks: bidirectional language modeling and sequence-to-sequence language modeling. After pre-train, DiMBERT is further fine-tuned for the downstream tasks. Experiments show that DiMBERT sets new state-of-the-art performance on three tasks (over four datasets), including both generation tasks (image captioning and visual storytelling) and classification tasks (referring expressions). The proposed DiM (short for Disentangled Multimodal-Attention) module can be easily incorporated into existing pre-trained V-L models to boost their performance, up to a 5% increase on the representative task. Finally, we conduct a systematic analysis and demonstrate the effectiveness of our DiM and the introduced visual concepts.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132378904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
MCS+: An Efficient Algorithm for Crawling the Community Structure in Multiplex Networks MCS+:一种高效的多路网络社区结构爬行算法
Pub Date : 2021-07-03 DOI: 10.1145/3451527
Ricky Laishram, Jeremy D. Wendt, S. Soundarajan
In this article, we consider the problem of crawling a multiplex network to identify the community structure of a layer-of-interest. A multiplex network is one where there are multiple types of relationships between the nodes. In many multiplex networks, some layers might be easier to explore (in terms of time, money etc.). We propose MCS+, an algorithm that can use the information from the easier to explore layers to help in the exploration of a layer-of-interest that is expensive to explore. We consider the goal of exploration to be generating a sample that is representative of the communities in the complete layer-of-interest. This work has practical applications in areas such as exploration of dark (e.g., criminal) networks, online social networks, biological networks, and so on. For example, in a terrorist network, relationships such as phone records, e-mail records, and so on are easier to collect; in contrast, data on the face-to-face communications are much harder to collect, but also potentially more valuable. We perform extensive experimental evaluations on real-world networks, and we observe that MCS+ consistently outperforms the best baseline—the similarity of the sample that MCS+ generates to the real network is up to three times that of the best baseline in some networks. We also perform theoretical and experimental evaluations on the scalability of MCS+ to network properties, and find that it scales well with the budget, number of layers in the multiplex network, and the average degree in the original network.
在本文中,我们考虑了爬行多路网络以识别兴趣层的社区结构的问题。多路网络是节点之间存在多种类型关系的网络。在许多多路网络中,某些层可能更容易探索(在时间、金钱等方面)。我们提出了MCS+算法,它可以使用来自更容易探索层的信息来帮助探索昂贵的感兴趣层。我们认为探索的目标是生成一个样本,在整个感兴趣的层中代表社区。这项工作在探索黑暗(如犯罪)网络、在线社交网络、生物网络等领域具有实际应用价值。例如,在恐怖分子网络中,诸如电话记录、电子邮件记录等关系更容易收集;相比之下,面对面交流的数据更难收集,但也可能更有价值。我们对现实世界的网络进行了广泛的实验评估,我们观察到MCS+始终优于最佳基线- MCS+生成的样本与真实网络的相似性在某些网络中高达最佳基线的三倍。我们还对MCS+对网络特性的可扩展性进行了理论和实验评估,发现它与预算、复用网络的层数和原始网络的平均度都有很好的扩展性。
{"title":"MCS+: An Efficient Algorithm for Crawling the Community Structure in Multiplex Networks","authors":"Ricky Laishram, Jeremy D. Wendt, S. Soundarajan","doi":"10.1145/3451527","DOIUrl":"https://doi.org/10.1145/3451527","url":null,"abstract":"In this article, we consider the problem of crawling a multiplex network to identify the community structure of a layer-of-interest. A multiplex network is one where there are multiple types of relationships between the nodes. In many multiplex networks, some layers might be easier to explore (in terms of time, money etc.). We propose MCS+, an algorithm that can use the information from the easier to explore layers to help in the exploration of a layer-of-interest that is expensive to explore. We consider the goal of exploration to be generating a sample that is representative of the communities in the complete layer-of-interest. This work has practical applications in areas such as exploration of dark (e.g., criminal) networks, online social networks, biological networks, and so on. For example, in a terrorist network, relationships such as phone records, e-mail records, and so on are easier to collect; in contrast, data on the face-to-face communications are much harder to collect, but also potentially more valuable. We perform extensive experimental evaluations on real-world networks, and we observe that MCS+ consistently outperforms the best baseline—the similarity of the sample that MCS+ generates to the real network is up to three times that of the best baseline in some networks. We also perform theoretical and experimental evaluations on the scalability of MCS+ to network properties, and find that it scales well with the budget, number of layers in the multiplex network, and the average degree in the original network.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124079447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed Latent Dirichlet Allocation on Streams 流上的分布式潜Dirichlet分配
Pub Date : 2021-07-03 DOI: 10.1145/3451528
Yunyan Guo, Jianzhong Li
Latent Dirichlet Allocation (LDA) has been widely used for topic modeling, with applications spanning various areas such as natural language processing and information retrieval. While LDA on small and static datasets has been extensively studied, several real-world challenges are posed in practical scenarios where datasets are often huge and are gathered in a streaming fashion. As the state-of-the-art LDA algorithm on streams, Streaming Variational Bayes (SVB) introduced Bayesian updating to provide a streaming procedure. However, the utility of SVB is limited in applications since it ignored three challenges of processing real-world streams: topic evolution, data turbulence, and real-time inference. In this article, we propose a novel distributed LDA algorithm—referred to as StreamFed-LDA—to deal with challenges on streams. For topic modeling of streaming data, the ability to capture evolving topics is essential for practical online inference. To achieve this goal, StreamFed-LDA is based on a specialized framework that supports lifelong (continual) learning of evolving topics. On the other hand, data turbulence is commonly present in streams due to real-life events. In that case, the design of StreamFed-LDA allows the model to learn new characteristics from the most recent data while maintaining the historical information. On massive streaming data, it is difficult and crucial to provide real-time inference results. To increase the throughput and reduce the latency, StreamFed-LDA introduces additional techniques that substantially reduce both computation and communication costs in distributed systems. Experiments on four real-world datasets show that the proposed framework achieves significantly better performance of online inference compared with the baselines. At the same time, StreamFed-LDA also reduces the latency by orders of magnitudes in real-world datasets.
潜在狄利克雷分配(Latent Dirichlet Allocation, LDA)在主题建模中得到了广泛的应用,在自然语言处理和信息检索等领域都有广泛的应用。虽然对小型静态数据集的LDA已经进行了广泛的研究,但在实际场景中,数据集通常是巨大的,并且以流方式收集,因此提出了几个现实世界的挑战。流变分贝叶斯(Streaming Variational Bayes, SVB)是目前流上最先进的LDA算法,它引入贝叶斯更新来提供一个流处理过程。然而,SVB的实用性在应用中受到限制,因为它忽略了处理现实世界流的三个挑战:主题演变、数据湍流和实时推理。在本文中,我们提出了一种新的分布式LDA算法——称为streamfed -LDA——来处理流上的挑战。对于流数据的主题建模,捕获不断变化的主题的能力对于实际的在线推理是必不可少的。为了实现这一目标,StreamFed-LDA基于一个专门的框架,该框架支持对不断发展的主题进行终身(持续)学习。另一方面,由于现实生活中的事件,数据湍流通常存在于流中。在这种情况下,StreamFed-LDA的设计允许模型从最新的数据中学习新的特征,同时保持历史信息。在海量流数据中,提供实时的推理结果是非常困难和关键的。为了提高吞吐量和减少延迟,StreamFed-LDA引入了额外的技术,这些技术大大降低了分布式系统中的计算和通信成本。在四个真实数据集上的实验表明,与基线相比,该框架的在线推理性能显著提高。与此同时,StreamFed-LDA还将现实世界数据集的延迟降低了几个数量级。
{"title":"Distributed Latent Dirichlet Allocation on Streams","authors":"Yunyan Guo, Jianzhong Li","doi":"10.1145/3451528","DOIUrl":"https://doi.org/10.1145/3451528","url":null,"abstract":"Latent Dirichlet Allocation (LDA) has been widely used for topic modeling, with applications spanning various areas such as natural language processing and information retrieval. While LDA on small and static datasets has been extensively studied, several real-world challenges are posed in practical scenarios where datasets are often huge and are gathered in a streaming fashion. As the state-of-the-art LDA algorithm on streams, Streaming Variational Bayes (SVB) introduced Bayesian updating to provide a streaming procedure. However, the utility of SVB is limited in applications since it ignored three challenges of processing real-world streams: topic evolution, data turbulence, and real-time inference. In this article, we propose a novel distributed LDA algorithm—referred to as StreamFed-LDA—to deal with challenges on streams. For topic modeling of streaming data, the ability to capture evolving topics is essential for practical online inference. To achieve this goal, StreamFed-LDA is based on a specialized framework that supports lifelong (continual) learning of evolving topics. On the other hand, data turbulence is commonly present in streams due to real-life events. In that case, the design of StreamFed-LDA allows the model to learn new characteristics from the most recent data while maintaining the historical information. On massive streaming data, it is difficult and crucial to provide real-time inference results. To increase the throughput and reduce the latency, StreamFed-LDA introduces additional techniques that substantially reduce both computation and communication costs in distributed systems. Experiments on four real-world datasets show that the proposed framework achieves significantly better performance of online inference compared with the baselines. At the same time, StreamFed-LDA also reduces the latency by orders of magnitudes in real-world datasets.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"06 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128926622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Modeling Temporal Patterns with Dilated Convolutions for Time-Series Forecasting 用扩展卷积建模时间序列预测的时间模式
Pub Date : 2021-07-03 DOI: 10.1145/3453724
Yangfan Li, Kenli Li, Cen Chen, Xu Zhou, Zeng Zeng, Kuan-Ching Li
Time-series forecasting is an important problem across a wide range of domains. Designing accurate and prompt forecasting algorithms is a non-trivial task, as temporal data that arise in real applications often involve both non-linear dynamics and linear dependencies, and always have some mixtures of sequential and periodic patterns, such as daily, weekly repetitions, and so on. At this point, however, most recent deep models often use Recurrent Neural Networks (RNNs) to capture these temporal patterns, which is hard to parallelize and not fast enough for real-world applications especially when a huge amount of user requests are coming. Recently, CNNs have demonstrated significant advantages for sequence modeling tasks over the de-facto RNNs, while providing high computational efficiency due to the inherent parallelism. In this work, we propose HyDCNN, a novel hybrid framework based on fully Dilated CNN for time-series forecasting tasks. The core component in HyDCNN is a proposed hybrid module, in which our proposed position-aware dilated CNNs are utilized to capture the sequential non-linear dynamics and an autoregressive model is leveraged to capture the sequential linear dependencies. To further capture the periodic temporal patterns, a novel hop scheme is introduced in the hybrid module. HyDCNN is then composed of multiple hybrid modules to capture the sequential and periodic patterns. Each of these hybrid modules targets on either the sequential pattern or one kind of periodic patterns. Extensive experiments on five real-world datasets have shown that the proposed HyDCNN is better compared with state-of-the-art baselines and is at least 200% better than RNN baselines. The datasets and source code will be published in Github to facilitate more future work.
时间序列预测是一个涉及广泛领域的重要问题。设计准确和及时的预测算法是一项重要的任务,因为在实际应用程序中出现的时间数据通常涉及非线性动态和线性依赖关系,并且总是有一些顺序和周期性模式的混合,例如每天、每周重复,等等。然而,在这一点上,大多数最新的深度模型通常使用循环神经网络(rnn)来捕获这些时间模式,这很难并行化,并且对于现实世界的应用程序来说不够快,特别是当大量用户请求到来时。近年来,cnn在序列建模任务中表现出明显优于事实rnn的优势,同时由于其固有的并行性提供了很高的计算效率。在这项工作中,我们提出了一种基于完全扩展CNN的新型混合框架HyDCNN,用于时间序列预测任务。HyDCNN的核心组件是一个混合模块,其中我们提出的位置感知扩展cnn用于捕获序列非线性动态,并利用自回归模型来捕获序列线性依赖关系。为了进一步捕获周期性时间模式,在混合模块中引入了一种新的跳码方案。然后,HyDCNN由多个混合模块组成,以捕获顺序和周期性模式。这些混合模块中的每一个都针对顺序模式或一种周期模式。在五个真实数据集上进行的大量实验表明,与最先进的基线相比,提出的HyDCNN更好,至少比RNN基线好200%。数据集和源代码将在Github上发布,以方便更多的未来工作。
{"title":"Modeling Temporal Patterns with Dilated Convolutions for Time-Series Forecasting","authors":"Yangfan Li, Kenli Li, Cen Chen, Xu Zhou, Zeng Zeng, Kuan-Ching Li","doi":"10.1145/3453724","DOIUrl":"https://doi.org/10.1145/3453724","url":null,"abstract":"Time-series forecasting is an important problem across a wide range of domains. Designing accurate and prompt forecasting algorithms is a non-trivial task, as temporal data that arise in real applications often involve both non-linear dynamics and linear dependencies, and always have some mixtures of sequential and periodic patterns, such as daily, weekly repetitions, and so on. At this point, however, most recent deep models often use Recurrent Neural Networks (RNNs) to capture these temporal patterns, which is hard to parallelize and not fast enough for real-world applications especially when a huge amount of user requests are coming. Recently, CNNs have demonstrated significant advantages for sequence modeling tasks over the de-facto RNNs, while providing high computational efficiency due to the inherent parallelism. In this work, we propose HyDCNN, a novel hybrid framework based on fully Dilated CNN for time-series forecasting tasks. The core component in HyDCNN is a proposed hybrid module, in which our proposed position-aware dilated CNNs are utilized to capture the sequential non-linear dynamics and an autoregressive model is leveraged to capture the sequential linear dependencies. To further capture the periodic temporal patterns, a novel hop scheme is introduced in the hybrid module. HyDCNN is then composed of multiple hybrid modules to capture the sequential and periodic patterns. Each of these hybrid modules targets on either the sequential pattern or one kind of periodic patterns. Extensive experiments on five real-world datasets have shown that the proposed HyDCNN is better compared with state-of-the-art baselines and is at least 200% better than RNN baselines. The datasets and source code will be published in Github to facilitate more future work.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124996552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Graph-Based Stock Recommendation by Time-Aware Relational Attention Network 基于时间感知关系注意网络的图型股票推荐
Pub Date : 2021-07-03 DOI: 10.1145/3451397
Jianliang Gao, Xiaoting Ying, Cong Xu, Jianxin Wang, Shichao Zhang, Zhao Li
The stock market investors aim at maximizing their investment returns. Stock recommendation task is to recommend stocks with higher return ratios for the investors. Most stock prediction methods study the historical sequence patterns to predict stock trend or price in the near future. In fact, the future price of a stock is correlated not only with its historical price, but also with other stocks. In this article, we take into account the relationships between stocks (corporations) by stock relation graph. Furthermore, we propose a Time-aware Relational Attention Network (TRAN) for graph-based stock recommendation according to return ratio ranking. In TRAN, the time-aware relational attention mechanism is designed to capture time-varying correlation strengths between stocks by the interaction of historical sequences and stock description documents. With the dynamic strengths, the nodes of the stock relation graph aggregate the features of neighbor stock nodes by graph convolution operation. For a given group of stocks, the proposed TRAN model can output the ranking results of stocks according to their return ratios. The experimental results on several real-world datasets demonstrate the effectiveness of our TRAN for stock recommendation.
股市投资者的目标是使他们的投资收益最大化。股票推荐任务是为投资者推荐回报率较高的股票。大多数股票预测方法通过研究历史序列模式来预测近期的股票走势或价格。事实上,一只股票的未来价格不仅与其历史价格相关,还与其他股票相关。在本文中,我们通过股票关系图来考虑股票(公司)之间的关系。在此基础上,我们提出了一种基于时间感知的关系注意网络(TRAN),用于基于收益率排序的基于图的股票推荐。在TRAN中,设计了时间感知的关系注意机制,通过历史序列和股票描述文件的相互作用来捕获股票之间时变的相关强度。股票关系图的节点利用动态优势,通过图卷积运算对相邻股票节点的特征进行聚合。对于给定的一组股票,提出的TRAN模型可以根据股票的收益率输出股票的排序结果。在几个真实数据集上的实验结果证明了我们的TRAN在股票推荐中的有效性。
{"title":"Graph-Based Stock Recommendation by Time-Aware Relational Attention Network","authors":"Jianliang Gao, Xiaoting Ying, Cong Xu, Jianxin Wang, Shichao Zhang, Zhao Li","doi":"10.1145/3451397","DOIUrl":"https://doi.org/10.1145/3451397","url":null,"abstract":"The stock market investors aim at maximizing their investment returns. Stock recommendation task is to recommend stocks with higher return ratios for the investors. Most stock prediction methods study the historical sequence patterns to predict stock trend or price in the near future. In fact, the future price of a stock is correlated not only with its historical price, but also with other stocks. In this article, we take into account the relationships between stocks (corporations) by stock relation graph. Furthermore, we propose a Time-aware Relational Attention Network (TRAN) for graph-based stock recommendation according to return ratio ranking. In TRAN, the time-aware relational attention mechanism is designed to capture time-varying correlation strengths between stocks by the interaction of historical sequences and stock description documents. With the dynamic strengths, the nodes of the stock relation graph aggregate the features of neighbor stock nodes by graph convolution operation. For a given group of stocks, the proposed TRAN model can output the ranking results of stocks according to their return ratios. The experimental results on several real-world datasets demonstrate the effectiveness of our TRAN for stock recommendation.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125607440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Page-Level Main Content Extraction From Heterogeneous Webpages 从异构网页中提取页面级主要内容
Pub Date : 2021-06-28 DOI: 10.1145/3451168
Julián Alarte, Josep Silva
The main content of a webpage is often surrounded by other boilerplate elements related to the template, such as menus, advertisements, copyright notices, and comments. For crawlers and indexers, isolating the main content from the template and other noisy information is an essential task, because processing and storing noisy information produce a waste of resources such as bandwidth, storage space, and computing time. Besides, the detection and extraction of the main content is useful in different areas, such as data mining, web summarization, and content adaptation to low resolutions. This work introduces a new technique for main content extraction. In contrast to most techniques, this technique not only extracts text, but also other types of content, such as images, and animations. It is a Document Object Model-based page-level technique, thus it only needs to load one single webpage to extract the main content. As a consequence, it is efficient enough as to be used online (in real-time). We have empirically evaluated the technique using a suite of real heterogeneous benchmarks producing very good results compared with other well-known content extraction techniques.
网页的主要内容通常被与模板相关的其他样板元素所包围,例如菜单、广告、版权声明和注释。对于爬虫和索引器来说,将主要内容与模板和其他噪声信息隔离开来是一项必不可少的任务,因为处理和存储噪声信息会浪费带宽、存储空间和计算时间等资源。此外,主要内容的检测和提取在数据挖掘、web摘要和低分辨率内容适应等不同领域都很有用。本文介绍了一种新的主内容提取技术。与大多数技术相比,这种技术不仅可以提取文本,还可以提取其他类型的内容,如图像和动画。它是一种基于文档对象模型的页面级技术,因此它只需要加载一个网页就可以提取主要内容。因此,它足够有效,可以在线(实时)使用。我们使用一套真正的异构基准测试对该技术进行了经验评估,与其他知名的内容提取技术相比,产生了非常好的结果。
{"title":"Page-Level Main Content Extraction From Heterogeneous Webpages","authors":"Julián Alarte, Josep Silva","doi":"10.1145/3451168","DOIUrl":"https://doi.org/10.1145/3451168","url":null,"abstract":"The main content of a webpage is often surrounded by other boilerplate elements related to the template, such as menus, advertisements, copyright notices, and comments. For crawlers and indexers, isolating the main content from the template and other noisy information is an essential task, because processing and storing noisy information produce a waste of resources such as bandwidth, storage space, and computing time. Besides, the detection and extraction of the main content is useful in different areas, such as data mining, web summarization, and content adaptation to low resolutions. This work introduces a new technique for main content extraction. In contrast to most techniques, this technique not only extracts text, but also other types of content, such as images, and animations. It is a Document Object Model-based page-level technique, thus it only needs to load one single webpage to extract the main content. As a consequence, it is efficient enough as to be used online (in real-time). We have empirically evaluated the technique using a suite of real heterogeneous benchmarks producing very good results compared with other well-known content extraction techniques.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114326070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
App2Vec: Context-Aware Application Usage Prediction App2Vec:上下文感知应用程序使用预测
Pub Date : 2021-06-28 DOI: 10.1145/3451396
Huandong Wang, Yong Li, Mu Du, Zhenhui Li, Depeng Jin
Both app developers and service providers have strong motivations to understand when and where certain apps are used by users. However, it has been a challenging problem due to the highly skewed and noisy app usage data. Moreover, apps are regarded as independent items in existing studies, which fail to capture the hidden semantics in app usage traces. In this article, we propose App2Vec, a powerful representation learning model to learn the semantic embedding of apps with the consideration of spatio-temporal context. Based on the obtained semantic embeddings, we develop a probabilistic model based on the Bayesian mixture model and Dirichlet process to capture when, where, and what semantics of apps are used to predict the future usage. We evaluate our model using two different app usage datasets, which involve over 1.7 million users and 2,000+ apps. Evaluation results show that our proposed App2Vec algorithm outperforms the state-of-the-art algorithms in app usage prediction with a performance gap of over 17.0%.
应用开发者和服务提供商都有强烈的动机去了解用户何时何地使用某些应用。然而,由于应用使用数据的高度扭曲和嘈杂,这是一个具有挑战性的问题。此外,在现有的研究中,应用程序被视为独立的项目,未能捕捉到应用程序使用痕迹中隐藏的语义。在本文中,我们提出了一个强大的表征学习模型App2Vec,该模型可以在考虑时空上下文的情况下学习应用的语义嵌入。基于获得的语义嵌入,我们开发了一个基于贝叶斯混合模型和狄利克雷过程的概率模型,以捕获应用程序的何时、何地以及使用什么语义来预测未来的使用情况。我们使用两个不同的应用使用数据集来评估我们的模型,这些数据集涉及超过170万用户和2000多个应用。评估结果表明,我们提出的App2Vec算法在应用使用预测方面优于目前最先进的算法,性能差距超过17.0%。
{"title":"App2Vec: Context-Aware Application Usage Prediction","authors":"Huandong Wang, Yong Li, Mu Du, Zhenhui Li, Depeng Jin","doi":"10.1145/3451396","DOIUrl":"https://doi.org/10.1145/3451396","url":null,"abstract":"Both app developers and service providers have strong motivations to understand when and where certain apps are used by users. However, it has been a challenging problem due to the highly skewed and noisy app usage data. Moreover, apps are regarded as independent items in existing studies, which fail to capture the hidden semantics in app usage traces. In this article, we propose App2Vec, a powerful representation learning model to learn the semantic embedding of apps with the consideration of spatio-temporal context. Based on the obtained semantic embeddings, we develop a probabilistic model based on the Bayesian mixture model and Dirichlet process to capture when, where, and what semantics of apps are used to predict the future usage. We evaluate our model using two different app usage datasets, which involve over 1.7 million users and 2,000+ apps. Evaluation results show that our proposed App2Vec algorithm outperforms the state-of-the-art algorithms in app usage prediction with a performance gap of over 17.0%.","PeriodicalId":435653,"journal":{"name":"ACM Transactions on Knowledge Discovery from Data (TKDD)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121402201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
ACM Transactions on Knowledge Discovery from Data (TKDD)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1