首页 > 最新文献

ACM Transactions on Information Systems最新文献

英文 中文
Understanding or Manipulation: Rethinking Online Performance Gains of Modern Recommender Systems 理解还是操纵?反思现代推荐系统的在线性能收益
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-12-15 DOI: 10.1145/3637869
Zhengbang Zhu, Rongjun Qin, Junjie Huang, Xinyi Dai, Yang Yu†, Yong Yu, Weinan Zhang†

Recommender systems are expected to be assistants that help human users find relevant information automatically without explicit queries. As recommender systems evolve, increasingly sophisticated learning techniques are applied and have achieved better performance in terms of user engagement metrics such as clicks and browsing time. The increase in the measured performance, however, can have two possible attributions: a better understanding of user preferences, and a more proactive ability to utilize human bounded rationality to seduce user over-consumption. A natural following question is whether current recommendation algorithms are manipulating user preferences. If so, can we measure the manipulation level? In this paper, we present a general framework for benchmarking the degree of manipulations of recommendation algorithms, in both slate recommendation and sequential recommendation scenarios. The framework consists of four stages, initial preference calculation, training data collection, algorithm training and interaction, and metrics calculation that involves two proposed metrics, Manipulation Score and Preference Shift. We benchmark some representative recommendation algorithms in both synthetic and real-world datasets under the proposed framework. We have observed that a high online click-through rate does not necessarily mean a better understanding of user initial preference, but ends in prompting users to choose more documents they initially did not favor. Moreover, we find that the training data have notable impacts on the manipulation degrees, and algorithms with more powerful modeling abilities are more sensitive to such impacts. The experiments also verified the usefulness of the proposed metrics for measuring the degree of manipulations. We advocate that future recommendation algorithm studies should be treated as an optimization problem with constrained user preference manipulations.

推荐系统有望成为帮助人类用户在没有明确查询的情况下自动查找相关信息的助手。随着推荐系统的发展,人们应用了越来越复杂的学习技术,并在用户参与度指标(如点击量和浏览时间)方面取得了更好的性能。然而,衡量性能的提高可能有两个原因:一是对用户偏好有了更好的理解,二是能够更主动地利用人类的有限理性来诱导用户过度消费。接下来的一个自然问题是,当前的推荐算法是否操纵了用户偏好。如果是,我们能否衡量操纵程度?在本文中,我们提出了一个通用框架,用于在板块推荐和顺序推荐两种情况下对推荐算法的操纵程度进行基准测试。该框架包括四个阶段:初始偏好计算、训练数据收集、算法训练和交互,以及指标计算,其中涉及两个建议的指标:操纵分数和偏好偏移。我们根据提出的框架,在合成数据集和真实数据集中对一些具有代表性的推荐算法进行了基准测试。我们发现,在线点击率高并不一定意味着能更好地了解用户的初始偏好,而是会促使用户选择更多他们最初并不喜欢的文档。此外,我们还发现训练数据对操作度有显著影响,而建模能力更强的算法对这种影响更为敏感。实验还验证了所提出的衡量操纵程度的指标的实用性。我们主张,未来的推荐算法研究应将用户偏好操作作为一个优化问题来处理。
{"title":"Understanding or Manipulation: Rethinking Online Performance Gains of Modern Recommender Systems","authors":"Zhengbang Zhu, Rongjun Qin, Junjie Huang, Xinyi Dai, Yang Yu†, Yong Yu, Weinan Zhang†","doi":"10.1145/3637869","DOIUrl":"https://doi.org/10.1145/3637869","url":null,"abstract":"<p>Recommender systems are expected to be assistants that help human users find relevant information automatically without explicit queries. As recommender systems evolve, increasingly sophisticated learning techniques are applied and have achieved better performance in terms of user engagement metrics such as clicks and browsing time. The increase in the measured performance, however, can have two possible attributions: a better understanding of user preferences, and a more proactive ability to utilize human bounded rationality to seduce user over-consumption. A natural following question is whether current recommendation algorithms are manipulating user preferences. If so, can we measure the manipulation level? In this paper, we present a general framework for benchmarking the degree of manipulations of recommendation algorithms, in both slate recommendation and sequential recommendation scenarios. The framework consists of four stages, initial preference calculation, training data collection, algorithm training and interaction, and metrics calculation that involves two proposed metrics, Manipulation Score and Preference Shift. We benchmark some representative recommendation algorithms in both synthetic and real-world datasets under the proposed framework. We have observed that a high online click-through rate does not necessarily mean a better understanding of user initial preference, but ends in prompting users to choose more documents they initially did not favor. Moreover, we find that the training data have notable impacts on the manipulation degrees, and algorithms with more powerful modeling abilities are more sensitive to such impacts. The experiments also verified the usefulness of the proposed metrics for measuring the degree of manipulations. We advocate that future recommendation algorithm studies should be treated as an optimization problem with constrained user preference manipulations.</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"26 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138691243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Effectiveness of Sampled Softmax Loss for Item Recommendation 论采样软最大损失在项目推荐中的有效性
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-12-13 DOI: 10.1145/3637061
Jiancan Wu, Xiang Wang, Xingyu Gao, Jiawei Chen, Hongcheng Fu, Tianyu Qiu, Xiangnan He

The learning objective plays a fundamental role to build a recommender system. Most methods routinely adopt either pointwise (e.g., binary cross-entropy) or pairwise (e.g., BPR) loss to train the model parameters, while rarely pay attention to softmax loss, which assumes the probabilities of all classes sum up to 1, due to its computational complexity when scaling up to large datasets or intractability for streaming data where the complete item space is not always available. The sampled softmax (SSM) loss emerges as an efficient substitute for softmax loss. Its special case, InfoNCE loss, has been widely used in self-supervised learning and exhibited remarkable performance for contrastive learning. Nonetheless, limited recommendation work uses the SSM loss as the learning objective. Worse still, none of them explores its properties thoroughly and answers “Does SSM loss suit for item recommendation?” and “What are the conceptual advantages of SSM loss, as compared with the prevalent losses?”, to the best of our knowledge.

In this work, we aim to offer a better understanding of SSM for item recommendation. Specifically, we first theoretically reveal three model-agnostic advantages: (1) mitigating popularity bias, which is beneficial to long-tail recommendation; (2) mining hard negative samples, which offers informative gradients to optimize model parameters; and (3) maximizing the ranking metric, which facilitates top-K performance. However, based on our empirical studies, we recognize that the default choice of cosine similarity function in SSM limits its ability in learning the magnitudes of representation vectors. As such, the combinations of SSM with the models that also fall short in adjusting magnitudes (e.g., matrix factorization) may result in poor representations. One step further, we provide mathematical proof that message passing schemes in graph convolution networks can adjust representation magnitude according to node degree, which naturally compensates for the shortcoming of SSM. Extensive experiments on four benchmark datasets justify our analyses, demonstrating the superiority of SSM for item recommendation. Our implementations are available in both TensorFlow and PyTorch.

学习目标对建立推荐系统起着至关重要的作用。大多数方法通常采用点状损失(如二元交叉熵)或成对损失(如 BPR)来训练模型参数,而很少关注软最大损失,因为软最大损失假定所有类别的概率总和为 1,这是因为它在扩展到大型数据集时计算复杂,或者对于并非总是有完整项目空间的流数据来说难以处理。采样软最大(SSM)损失作为软最大损失的有效替代品应运而生。它的特例--InfoNCE 损失已被广泛应用于自我监督学习中,并在对比学习中表现出卓越的性能。然而,使用 SSM 损失作为学习目标的推荐工作非常有限。更糟糕的是,据我们所知,没有一项研究深入探讨了 SSM 损失的特性,并回答了 "SSM 损失是否适合项目推荐?"以及 "与流行的损失相比,SSM 损失在概念上有哪些优势?"的问题。在这项工作中,我们旨在更好地理解用于项目推荐的 SSM。具体来说,我们首先从理论上揭示了与模型无关的三个优势:(1) 减少流行度偏差,这有利于长尾推荐;(2) 挖掘硬负样本,这为优化模型参数提供了信息梯度;(3) 最大化排名度量,这有利于 Top-K 性能。然而,根据我们的实证研究,我们认识到 SSM 默认选择的余弦相似度函数限制了其学习表示向量大小的能力。因此,将 SSM 与同样无法调整大小的模型(如矩阵因式分解)相结合,可能会导致较差的表示。我们进一步提供了数学证明,即图卷积网络中的消息传递方案可以根据节点度调整表示量,这自然弥补了 SSM 的不足。在四个基准数据集上进行的广泛实验证明了我们的分析,证明了 SSM 在项目推荐方面的优越性。我们的实现可在 TensorFlow 和 PyTorch 中使用。
{"title":"On the Effectiveness of Sampled Softmax Loss for Item Recommendation","authors":"Jiancan Wu, Xiang Wang, Xingyu Gao, Jiawei Chen, Hongcheng Fu, Tianyu Qiu, Xiangnan He","doi":"10.1145/3637061","DOIUrl":"https://doi.org/10.1145/3637061","url":null,"abstract":"<p>The learning objective plays a fundamental role to build a recommender system. Most methods routinely adopt either pointwise (<i>e.g.,</i> binary cross-entropy) or pairwise (<i>e.g.,</i> BPR) loss to train the model parameters, while rarely pay attention to softmax loss, which assumes the probabilities of all classes sum up to 1, due to its computational complexity when scaling up to large datasets or intractability for streaming data where the complete item space is not always available. The sampled softmax (SSM) loss emerges as an efficient substitute for softmax loss. Its special case, InfoNCE loss, has been widely used in self-supervised learning and exhibited remarkable performance for contrastive learning. Nonetheless, limited recommendation work uses the SSM loss as the learning objective. Worse still, none of them explores its properties thoroughly and answers “Does SSM loss suit for item recommendation?” and “What are the conceptual advantages of SSM loss, as compared with the prevalent losses?”, to the best of our knowledge. </p><p>In this work, we aim to offer a better understanding of SSM for item recommendation. Specifically, we first theoretically reveal three model-agnostic advantages: (1) mitigating popularity bias, which is beneficial to long-tail recommendation; (2) mining hard negative samples, which offers informative gradients to optimize model parameters; and (3) maximizing the ranking metric, which facilitates top-<i>K</i> performance. However, based on our empirical studies, we recognize that the default choice of cosine similarity function in SSM limits its ability in learning the magnitudes of representation vectors. As such, the combinations of SSM with the models that also fall short in adjusting magnitudes (<i>e.g.,</i> matrix factorization) may result in poor representations. One step further, we provide mathematical proof that message passing schemes in graph convolution networks can adjust representation magnitude according to node degree, which naturally compensates for the shortcoming of SSM. Extensive experiments on four benchmark datasets justify our analyses, demonstrating the superiority of SSM for item recommendation. Our implementations are available in both TensorFlow and PyTorch.</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"5 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138580490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy-Preserving Individual-Level COVID-19 Infection Prediction via Federated Graph Learning 通过联合图学习进行保护隐私的个人级 COVID-19 感染预测
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-12-07 DOI: 10.1145/3633202
Wenjie Fu, Huandong Wang, Chen Gao, Guanghua Liu, Yong Li, Tao Jiang

Accurately predicting individual-level infection state is of great value since its essential role in reducing the damage of the epidemic. However, there exists an inescapable risk of privacy leakage in the fine-grained user mobility trajectories required by individual-level infection prediction. In this paper, we focus on developing a framework of privacy-preserving individual-level infection prediction based on federated learning (FL) and graph neural networks (GNN). We propose Falcon, a Federated grAph Learning method for privacy-preserving individual-level infeCtion predictiON. It utilizes a novel hypergraph structure with spatio-temporal hyperedges to describe the complex interactions between individuals and locations in the contagion process. By organically combining the FL framework with hypergraph neural networks, the information propagation process of the graph machine learning is able to be divided into two stages distributed on the server and the clients, respectively, so as to effectively protect user privacy while transmitting high-level information. Furthermore, it elaborately designs a differential privacy perturbation mechanism as well as a plausible pseudo location generation approach to preserve user privacy in the graph structure. Besides, it introduces a cooperative coupling mechanism between the individual-level prediction model and an additional region-level model to mitigate the detrimental impacts caused by the injected obfuscation mechanisms. Extensive experimental results show that our methodology outperforms state-of-the-art algorithms and is able to protect user privacy against actual privacy attacks. Our code and datasets are available at the link: https://github.com/wjfu99/FL-epidemic.

准确预测个人层面的感染状态具有重要价值,因为它在减少疫情危害方面发挥着至关重要的作用。然而,个体级感染预测所需的细粒度用户移动轨迹存在不可避免的隐私泄露风险。在本文中,我们重点开发了一种基于联合学习(FL)和图神经网络(GNN)的保护隐私的个体级感染预测框架。我们提出了一种用于保护隐私的个体级推断预测的联合图神经网络学习方法 Falcon。它利用具有时空超边缘的新型超图结构来描述传染过程中个体与地点之间的复杂互动。通过将 FL 框架与超图神经网络有机结合,将图机器学习的信息传播过程分为两个阶段,分别分布在服务器和客户端,从而在传输高级信息的同时有效保护用户隐私。此外,它还精心设计了一种差分隐私扰动机制和一种可信的伪位置生成方法,以保护图结构中的用户隐私。此外,它还引入了个人级预测模型和附加区域级模型之间的合作耦合机制,以减轻注入式混淆机制造成的不利影响。广泛的实验结果表明,我们的方法优于最先进的算法,能够在实际隐私攻击中保护用户隐私。我们的代码和数据集可从以下链接获取:https://github.com/wjfu99/FL-epidemic。
{"title":"Privacy-Preserving Individual-Level COVID-19 Infection Prediction via Federated Graph Learning","authors":"Wenjie Fu, Huandong Wang, Chen Gao, Guanghua Liu, Yong Li, Tao Jiang","doi":"10.1145/3633202","DOIUrl":"https://doi.org/10.1145/3633202","url":null,"abstract":"<p>Accurately predicting individual-level infection state is of great value since its essential role in reducing the damage of the epidemic. However, there exists an inescapable risk of privacy leakage in the fine-grained user mobility trajectories required by individual-level infection prediction. In this paper, we focus on developing a framework of privacy-preserving individual-level infection prediction based on federated learning (FL) and graph neural networks (GNN). We propose <i>Falcon</i>, a <b>F</b>ederated gr<b>A</b>ph <b>L</b>earning method for privacy-preserving individual-level infe<b>C</b>tion predicti<b>ON</b>. It utilizes a novel hypergraph structure with spatio-temporal hyperedges to describe the complex interactions between individuals and locations in the contagion process. By organically combining the FL framework with hypergraph neural networks, the information propagation process of the graph machine learning is able to be divided into two stages distributed on the server and the clients, respectively, so as to effectively protect user privacy while transmitting high-level information. Furthermore, it elaborately designs a differential privacy perturbation mechanism as well as a plausible pseudo location generation approach to preserve user privacy in the graph structure. Besides, it introduces a cooperative coupling mechanism between the individual-level prediction model and an additional region-level model to mitigate the detrimental impacts caused by the injected obfuscation mechanisms. Extensive experimental results show that our methodology outperforms state-of-the-art algorithms and is able to protect user privacy against actual privacy attacks. Our code and datasets are available at the link: https://github.com/wjfu99/FL-epidemic.</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"21 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138546759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Effective and Efficient Sparse Neural Information Retrieval 实现有效、高效的稀疏神经信息检索
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-12-02 DOI: 10.1145/3634912
Thibault Formal, Carlos Lassance, Benjamin Piwowarski, Stéphane Clinchant

Sparse representation learning based on Pre-trained Language Models has seen a growing interest in Information Retrieval. Such approaches can take advantage of the proven efficiency of inverted indexes, and inherit desirable IR priors such as explicit lexical matching or some degree of interpretability. In this work, we thoroughly develop the framework of sparse representation learning in IR, which unifies term weighting and expansion in a supervised setting. We then build on SPLADE – a sparse expansion-based retriever – and show to which extent it is able to benefit from the same training improvements as dense bi-encoders, by studying the effect of distillation, hard negative mining as well as the Pre-trained Language Model’s initialization on its effectiveness – leading to state-of-the-art results in both in- and out-of-domain evaluation settings (SPLADE++). We furthermore propose efficiency improvements, allowing us to reach latency requirements on par with traditional keyword-based approaches (Efficient-SPLADE).

基于预训练语言模型的稀疏表示学习在信息检索领域越来越受到关注。这种方法可以利用倒排索引的公认效率,并继承理想的 IR 先验,如明确的词性匹配或一定程度的可解释性。在这项工作中,我们深入开发了 IR 中的稀疏表示学习框架,该框架将术语加权和扩展统一在一个有监督的环境中。然后,我们建立了基于稀疏扩展的检索器 SPLADE,并通过研究蒸馏、硬否定挖掘以及预训练语言模型的初始化对其有效性的影响,展示了 SPLADE 在多大程度上能够从与密集双编码器相同的训练改进中获益,从而在域内和域外评估设置(SPLADE++)中都取得了最先进的结果。此外,我们还提出了提高效率的建议,使我们能够达到与传统基于关键词的方法(Efficient-SPLADE)同等的延迟要求。
{"title":"Towards Effective and Efficient Sparse Neural Information Retrieval","authors":"Thibault Formal, Carlos Lassance, Benjamin Piwowarski, Stéphane Clinchant","doi":"10.1145/3634912","DOIUrl":"https://doi.org/10.1145/3634912","url":null,"abstract":"<p>Sparse representation learning based on Pre-trained Language Models has seen a growing interest in Information Retrieval. Such approaches can take advantage of the proven efficiency of inverted indexes, and inherit desirable IR priors such as explicit lexical matching or some degree of interpretability. In this work, we thoroughly develop the framework of sparse representation learning in IR, which unifies term weighting and expansion in a supervised setting. We then build on SPLADE – a sparse expansion-based retriever – and show to which extent it is able to benefit from the same training improvements as dense bi-encoders, by studying the effect of distillation, hard negative mining as well as the Pre-trained Language Model’s initialization on its <i>effectiveness</i> – leading to state-of-the-art results in both in- and out-of-domain evaluation settings (SPLADE++). We furthermore propose <i>efficiency</i> improvements, allowing us to reach latency requirements on par with traditional keyword-based approaches (Efficient-SPLADE).</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"17 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138559919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data Augmentation for Sample Efficient and Robust Document Ranking 基于样本高效鲁棒排序的数据增强
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-11-29 DOI: 10.1145/3634911
Abhijit Anand, Jurek Leonhardt, Jaspreet Singh, Koustav Rudra, Avishek Anand

Contextual ranking models have delivered impressive performance improvements over classical models in the document ranking task. However, these highly over-parameterized models tend to be data-hungry and require large amounts of data even for fine-tuning. In this paper, we propose data-augmentation methods for effective and robust ranking performance. One of the key benefits of using data augmentation is in achieving sample efficiency or learning effectively when we have only a small amount of training data. We propose supervised and unsupervised data augmentation schemes by creating training data using parts of the relevant documents in the query-document pairs. We then adapt a family of contrastive losses for the document ranking task that can exploit the augmented data to learn an effective ranking model. Our extensive experiments on subsets of the MS MARCO and TREC-DL test sets show that data augmentation, along with the ranking-adapted contrastive losses, results in performance improvements under most dataset sizes. Apart from sample efficiency, we conclusively show that data augmentation results in robust models when transferred to out-of-domain benchmarks. Our performance improvements in in-domain and more prominently in out-of-domain benchmarks show that augmentation regularizes the ranking model and improves its robustness and generalization capability.

在文档排序任务中,上下文排序模型比经典模型提供了令人印象深刻的性能改进。然而,这些高度过度参数化的模型往往需要大量数据,甚至需要大量数据进行微调。在本文中,我们提出了有效和稳健的排名性能的数据增强方法。使用数据增强的主要好处之一是在我们只有少量训练数据的情况下实现样本效率或有效学习。我们通过使用查询文档对中相关文档的部分创建训练数据,提出了有监督和无监督的数据增强方案。然后,我们为文档排序任务调整了一系列对比损失,可以利用增强的数据来学习有效的排序模型。我们对MS MARCO和TREC-DL测试集的子集进行了广泛的实验,结果表明,在大多数数据集大小下,数据增强以及与排名相适应的对比损失都能提高性能。除了样本效率之外,我们最后表明,当转移到域外基准测试时,数据增强会产生鲁棒模型。我们在域内和域外的性能改进表明,增强使排名模型规范化,提高了其鲁棒性和泛化能力。
{"title":"Data Augmentation for Sample Efficient and Robust Document Ranking","authors":"Abhijit Anand, Jurek Leonhardt, Jaspreet Singh, Koustav Rudra, Avishek Anand","doi":"10.1145/3634911","DOIUrl":"https://doi.org/10.1145/3634911","url":null,"abstract":"<p>Contextual ranking models have delivered impressive performance improvements over classical models in the document ranking task. However, these highly over-parameterized models tend to be data-hungry and require large amounts of data even for fine-tuning. In this paper, we propose data-augmentation methods for effective and robust ranking performance. One of the key benefits of using data augmentation is in achieving <i>sample efficiency</i> or learning effectively when we have only a small amount of training data. We propose supervised and unsupervised data augmentation schemes by creating training data using parts of the relevant documents in the query-document pairs. We then adapt a family of contrastive losses for the document ranking task that can exploit the augmented data to learn an effective ranking model. Our extensive experiments on subsets of the <span>MS MARCO</span> and <span>TREC-DL</span> test sets show that data augmentation, along with the ranking-adapted contrastive losses, results in performance improvements under most dataset sizes. Apart from sample efficiency, we conclusively show that data augmentation results in robust models when transferred to out-of-domain benchmarks. Our performance improvements in in-domain and more prominently in out-of-domain benchmarks show that augmentation regularizes the ranking model and improves its robustness and generalization capability.</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"9 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138537120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Less is More: Removing Redundancy of Graph Convolutional Networks for Recommendation 少即是多:去除图卷积网络的推荐冗余
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-11-20 DOI: 10.1145/3632751
Shaowen Peng, Kazunari Sugiyama, Tsunenori Mine

While Graph Convolutional Networks (GCNs) have shown great potential in recommender systems and collaborative filtering (CF), they suffer from expensive computational complexity and poor scalability. On top of that, recent works mostly combine GCNs with other advanced algorithms which further sacrifice model efficiency and scalability. In this work, we unveil the redundancy of existing GCN-based methods in three aspects: (1) Feature redundancy. By reviewing GCNs from a spectral perspective, we show that most spectral graph features are noisy for recommendation, while stacking graph convolution layers can suppress but cannot completely remove the noisy features, which we mostly summarize from our previous work; (2) Structure redundancy. By providing a deep insight into how user/item representations are generated, we show that what makes them distinctive lies in the spectral graph features, while the core idea of GCNs (i.e., neighborhood aggregation) is not the reason making GCNs effective; and (3) Distribution redundancy. Following observations from (1), we further show that the number of required spectral features is closely related to the spectral distribution, where important information tends to be concentrated in more (fewer) spectral features on a flatter (sharper) distribution. To make important information be concentrated in as few features as possible, we sharpen the spectral distribution by increasing the node similarity without changing the original data, thereby reducing the computational cost. To remove these three kinds of redundancies, we propose a Simplified Graph Denoising Encoder (SGDE) only exploiting the top-K singular vectors without explicitly aggregating neighborhood, which significantly reduces the complexity of GCN-based methods. We further propose a scalable contrastive learning framework to alleviate data sparsity and to boost model robustness and generalization, leading to significant improvement. Extensive experiments on three real-world datasets show that our proposed SGDE not only achieves state-of-the-art but also shows higher scalability and efficiency than our previously proposed GDE as well as traditional and GCN-based CF methods.

虽然图卷积网络(GCNs)在推荐系统和协同过滤(CF)中显示出巨大的潜力,但它们存在计算复杂度高和可扩展性差的问题。最重要的是,最近的研究大多将GCNs与其他高级算法结合起来,这进一步牺牲了模型的效率和可扩展性。在这项工作中,我们从三个方面揭示了现有的基于遗传神经网络的方法的冗余性:(1)特征冗余。通过从谱的角度回顾GCNs,我们发现大多数谱图特征是有噪声的,而叠加图卷积层可以抑制但不能完全去除有噪声的特征,这是我们从之前的工作中总结出来的;(2)结构冗余。通过深入了解用户/物品表征是如何生成的,我们发现它们的独特之处在于谱图特征,而GCNs的核心思想(即邻里聚集)并不是使GCNs有效的原因;(3)分布冗余。根据(1)的观测结果,我们进一步表明,所需光谱特征的数量与光谱分布密切相关,其中重要信息往往集中在更平坦(更锐利)分布的更多(更少)光谱特征中。为了使重要信息尽可能集中在较少的特征中,我们在不改变原始数据的情况下,通过增加节点相似度来锐化谱分布,从而降低计算成本。为了消除这三种冗余,我们提出了一种简化图去噪编码器(SGDE),它只利用top-K奇异向量,而不显式地聚集邻域,这大大降低了基于gcn方法的复杂性。我们进一步提出了一个可扩展的对比学习框架,以减轻数据稀疏性,并提高模型的鲁棒性和泛化,从而显著改善。在三个真实数据集上的大量实验表明,我们提出的SGDE不仅达到了最先进的水平,而且比我们之前提出的GDE以及传统的和基于gcn的CF方法具有更高的可扩展性和效率。
{"title":"Less is More: Removing Redundancy of Graph Convolutional Networks for Recommendation","authors":"Shaowen Peng, Kazunari Sugiyama, Tsunenori Mine","doi":"10.1145/3632751","DOIUrl":"https://doi.org/10.1145/3632751","url":null,"abstract":"<p>While Graph Convolutional Networks (GCNs) have shown great potential in recommender systems and collaborative filtering (CF), they suffer from expensive computational complexity and poor scalability. On top of that, recent works mostly combine GCNs with other advanced algorithms which further sacrifice model efficiency and scalability. In this work, we unveil the redundancy of existing GCN-based methods in three aspects: (1) <b>Feature redundancy</b>. By reviewing GCNs from a spectral perspective, we show that most spectral graph features are noisy for recommendation, while stacking graph convolution layers can suppress but cannot completely remove the noisy features, which we mostly summarize from our previous work; (2) <b>Structure redundancy</b>. By providing a deep insight into how user/item representations are generated, we show that what makes them distinctive lies in the spectral graph features, while the core idea of GCNs (<i>i.e.,</i> neighborhood aggregation) is not the reason making GCNs effective; and (3) <b>Distribution redundancy</b>. Following observations from (1), we further show that the number of required spectral features is closely related to the spectral distribution, where important information tends to be concentrated in more (fewer) spectral features on a flatter (sharper) distribution. To make important information be concentrated in as few features as possible, we sharpen the spectral distribution by increasing the node similarity without changing the original data, thereby reducing the computational cost. To remove these three kinds of redundancies, we propose a Simplified Graph Denoising Encoder (SGDE) only exploiting the top-<i>K</i> singular vectors without explicitly aggregating neighborhood, which significantly reduces the complexity of GCN-based methods. We further propose a scalable contrastive learning framework to alleviate data sparsity and to boost model robustness and generalization, leading to significant improvement. Extensive experiments on three real-world datasets show that our proposed SGDE not only achieves state-of-the-art but also shows higher scalability and efficiency than our previously proposed GDE as well as traditional and GCN-based CF methods.</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"1 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138537119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contextualizing and Expanding Conversational Queries without Supervision 上下文化和扩展会话查询没有监督
IF 5.6 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-11-17 DOI: 10.1145/3632622
Antonios Minas Krasakis, Andrew Yates, Evangelos Kanoulas

Most conversational passage retrieval systems try to resolve conversational dependencies by using an intermediate query resolution step. To do so, they synthesize conversational data or assume the availability of large-scale question rewritting datasets. To relax those conditions, we propose a zero-shot unified resolution–retrieval approach, that (i) contextualizes and (ii) expands query embeddings using the conversation history and without fine-tuning on conversational data. Contextualization biases the last user question embeddings towards the conversation. Query expansion is used in two ways: (i) abstractive expansion generates embeddings based on the current question and previous history, whereas (ii) extractive expansion tries to identify history term embeddings based on attention weights from the retriever. Our experiments demonstrate the effectiveness of both contextualization and unified expansion in improving conversational retrieval. Contextualization does so mostly by resolving anaphoras to the conversation and bringing their embeddings closer to the important resolution terms that were omitted. By adding embeddings to the query, expansion targets phenomena of ellipsis more explicitly, with our analysis verifying its effectiveness on identifying and adding important resolutions to the query. By combining contextualization and expansion, we find that our zero-shot unified resolution–retrieval methods are competitive and can even outperform supervised methods.

大多数会话通道检索系统试图通过使用中间查询解析步骤来解决会话依赖关系。为此,他们合成会话数据或假设大规模问题重写数据集的可用性。为了放松这些条件,我们提出了一种零采样统一分辨率检索方法,该方法(i)将会话历史上下文化,(ii)使用会话历史扩展查询嵌入,而不需要对会话数据进行微调。语境化使最后的用户问题嵌入偏向于对话。查询扩展以两种方式使用:(i)抽象扩展基于当前问题和以前的历史生成嵌入,而(ii)抽取扩展试图根据检索者的关注权重来识别历史术语嵌入。我们的实验证明了语境化和统一扩展在提高会话检索方面的有效性。语境化主要是通过解决对话中的回指,并使其嵌入更接近被省略的重要解决术语来实现的。通过在查询中添加嵌入,扩展更明确地针对省略号现象,我们的分析验证了它在识别和添加查询重要分辨率方面的有效性。通过将上下文化和扩展相结合,我们发现我们的零射击统一分辨率检索方法具有竞争力,甚至可以优于监督方法。
{"title":"Contextualizing and Expanding Conversational Queries without Supervision","authors":"Antonios Minas Krasakis, Andrew Yates, Evangelos Kanoulas","doi":"10.1145/3632622","DOIUrl":"https://doi.org/10.1145/3632622","url":null,"abstract":"<p>Most conversational passage retrieval systems try to resolve conversational dependencies by using an intermediate query resolution step. To do so, they synthesize conversational data or assume the availability of large-scale question rewritting datasets. To relax those conditions, we propose a zero-shot unified resolution–retrieval approach, that (i) contextualizes and (ii) expands query embeddings using the conversation history and without fine-tuning on conversational data. Contextualization biases the last user question embeddings towards the conversation. Query expansion is used in two ways: (i) abstractive expansion generates embeddings based on the current question and previous history, whereas (ii) extractive expansion tries to identify history term embeddings based on attention weights from the retriever. Our experiments demonstrate the effectiveness of both contextualization and unified expansion in improving conversational retrieval. Contextualization does so mostly by resolving anaphoras to the conversation and bringing their embeddings closer to the important resolution terms that were omitted. By adding embeddings to the query, expansion targets phenomena of ellipsis more explicitly, with our analysis verifying its effectiveness on identifying and adding important resolutions to the query. By combining contextualization and expansion, we find that our zero-shot unified resolution–retrieval methods are competitive and can even outperform supervised methods.</p>","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"26 1","pages":""},"PeriodicalIF":5.6,"publicationDate":"2023-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138537118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Dense Retrieval for Dialogue Response Selection 探索对话响应选择的密集检索
2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-11-14 DOI: 10.1145/3632750
Tian Lan, Deng Cai, Yan Wang, Yixuan Su, Heyan Huang, Xian-Ling Mao
Recent progress in deep learning has continuously improved the accuracy of dialogue response selection. However, in real-world scenarios, the high computation cost forces existing dialogue response selection models to rank only a small number of candidates, recalled by a coarse-grained model, precluding many high-quality candidates. To overcome this problem, we present a novel and efficient response selection model and a set of tailor-designed learning strategies to train it effectively. The proposed model consists of a dense retrieval module and an interaction layer, which could directly select the proper response from a large corpus. We conduct re-rank and full-rank evaluations on widely used benchmarks to evaluate our proposed model. Extensive experimental results demonstrate that our proposed model notably outperforms the state-of-the-art baselines on both re-rank and full-rank evaluations. Moreover, human evaluation results show that the response quality could be improved further by enlarging the candidate pool with nonparallel corpora. In addition, we also release high-quality benchmarks that are carefully annotated for more accurate dialogue response selection evaluation. All source codes, datasets, model parameters, and other related resources have been publicly available.
深度学习的最新进展不断提高了对话响应选择的准确性。然而,在现实场景中,高计算成本迫使现有的对话响应选择模型仅对少量候选对象进行排名,由粗粒度模型召回,从而排除了许多高质量的候选对象。为了克服这一问题,我们提出了一种新颖有效的响应选择模型和一套量身定制的学习策略来有效地训练它。该模型由密集检索模块和交互层组成,可以直接从大量语料库中选择合适的响应。我们对广泛使用的基准进行重新排序和全排序评估,以评估我们提出的模型。广泛的实验结果表明,我们提出的模型在重新排序和全排序评估上都明显优于最先进的基线。此外,人工评价结果表明,使用非并行语料库扩大候选语料库可以进一步提高响应质量。此外,我们还发布了经过仔细注释的高质量基准,以便更准确地评估对话响应选择。所有源代码、数据集、模型参数和其他相关资源都已公开。
{"title":"Exploring Dense Retrieval for Dialogue Response Selection","authors":"Tian Lan, Deng Cai, Yan Wang, Yixuan Su, Heyan Huang, Xian-Ling Mao","doi":"10.1145/3632750","DOIUrl":"https://doi.org/10.1145/3632750","url":null,"abstract":"Recent progress in deep learning has continuously improved the accuracy of dialogue response selection. However, in real-world scenarios, the high computation cost forces existing dialogue response selection models to rank only a small number of candidates, recalled by a coarse-grained model, precluding many high-quality candidates. To overcome this problem, we present a novel and efficient response selection model and a set of tailor-designed learning strategies to train it effectively. The proposed model consists of a dense retrieval module and an interaction layer, which could directly select the proper response from a large corpus. We conduct re-rank and full-rank evaluations on widely used benchmarks to evaluate our proposed model. Extensive experimental results demonstrate that our proposed model notably outperforms the state-of-the-art baselines on both re-rank and full-rank evaluations. Moreover, human evaluation results show that the response quality could be improved further by enlarging the candidate pool with nonparallel corpora. In addition, we also release high-quality benchmarks that are carefully annotated for more accurate dialogue response selection evaluation. All source codes, datasets, model parameters, and other related resources have been publicly available.","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"43 26","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134953620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
(Un)likelihood Training for Interpretable Embedding 可解释嵌入的(非)似然训练
2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-11-13 DOI: 10.1145/3632752
Jiaxin Wu, Chong-Wah Ngo, Wing-Kwong Chan, Zhijian Hou
Cross-modal representation learning has become a new normal for bridging the semantic gap between text and visual data. Learning modality agnostic representations in a continuous latent space, however, is often treated as a black-box data-driven training process. It is well-known that the effectiveness of representation learning depends heavily on the quality and scale of training data. For video representation learning, having a complete set of labels that annotate the full spectrum of video content for training is highly difficult, if not impossible. These issues, black-box training and dataset bias, make representation learning practically challenging to be deployed for video understanding due to unexplainable and unpredictable results. In this paper, we propose two novel training objectives, likelihood and unlikelihood functions, to unroll the semantics behind embeddings while addressing the label sparsity problem in training. The likelihood training aims to interpret semantics of embeddings beyond training labels, while the unlikelihood training leverages prior knowledge for regularization to ensure semantically coherent interpretation. With both training objectives, a new encoder-decoder network, which learns interpretable cross-modal representation, is proposed for ad-hoc video search. Extensive experiments on TRECVid and MSR-VTT datasets show that the proposed network outperforms several state-of-the-art retrieval models with a statistically significant performance margin.
跨模态表示学习已成为弥合文本和视觉数据之间语义差距的新常态。然而,在连续潜在空间中学习模态不可知表示通常被视为一个黑箱数据驱动的训练过程。众所周知,表示学习的有效性在很大程度上取决于训练数据的质量和规模。对于视频表示学习,拥有一套完整的标签来注释用于训练的全部视频内容是非常困难的,如果不是不可能的话。这些问题,黑箱训练和数据集偏差,由于无法解释和不可预测的结果,使得表示学习在视频理解中部署具有实际挑战性。在本文中,我们提出了两个新的训练目标,即似然函数和非似然函数,以揭示嵌入背后的语义,同时解决训练中的标签稀疏性问题。似然训练旨在解释超越训练标签的嵌入语义,而非似然训练利用先验知识进行正则化以确保语义连贯的解释。基于这两个训练目标,提出了一种新的编码器-解码器网络,该网络学习可解释的跨模态表示,用于自组织视频搜索。在TRECVid和MSR-VTT数据集上的大量实验表明,所提出的网络在统计上显著优于几种最先进的检索模型。
{"title":"(Un)likelihood Training for Interpretable Embedding","authors":"Jiaxin Wu, Chong-Wah Ngo, Wing-Kwong Chan, Zhijian Hou","doi":"10.1145/3632752","DOIUrl":"https://doi.org/10.1145/3632752","url":null,"abstract":"Cross-modal representation learning has become a new normal for bridging the semantic gap between text and visual data. Learning modality agnostic representations in a continuous latent space, however, is often treated as a black-box data-driven training process. It is well-known that the effectiveness of representation learning depends heavily on the quality and scale of training data. For video representation learning, having a complete set of labels that annotate the full spectrum of video content for training is highly difficult, if not impossible. These issues, black-box training and dataset bias, make representation learning practically challenging to be deployed for video understanding due to unexplainable and unpredictable results. In this paper, we propose two novel training objectives, likelihood and unlikelihood functions, to unroll the semantics behind embeddings while addressing the label sparsity problem in training. The likelihood training aims to interpret semantics of embeddings beyond training labels, while the unlikelihood training leverages prior knowledge for regularization to ensure semantically coherent interpretation. With both training objectives, a new encoder-decoder network, which learns interpretable cross-modal representation, is proposed for ad-hoc video search. Extensive experiments on TRECVid and MSR-VTT datasets show that the proposed network outperforms several state-of-the-art retrieval models with a statistically significant performance margin.","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"49 21","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136348319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cross-domain Recommendation via Dual Adversarial Adaptation 基于双对抗性适应的跨领域推荐
2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-11-11 DOI: 10.1145/3632524
Hongzu Su, Jingjing Li, Zhekai Du, Lei Zhu, Ke Lu, Heng Tao Shen
Data scarcity is a perpetual challenge of recommendation systems, and researchers have proposed a variety of cross-domain recommendation methods to alleviate the problem of data scarcity in target domains. However, in many real-world cross-domain recommendation systems, the source domain and the target domain are sampled from different data distributions, which obstructs the cross-domain knowledge transfer. In this paper, we propose to specifically align the data distributions between the source domain and the target domain to alleviate imbalanced sample distribution and thus challenge the data scarcity issue in the target domain. Technically, our proposed approach builds a dual adversarial adaptation (DAA) framework to adversarially train the target model together with a pre-trained source model. Two domain discriminators play the two-player minmax game with the target model and guide the target model to learn reliable domain-invariant features that can be transferred across domains. At the same time, the target model is calibrated to learn domain-specific information of the target domain. In addition, we formulate our approach as a plug-and-play module to boost existing recommendation systems. We apply the proposed method to address the issues of insufficient data and imbalanced sample distribution in real-world Click-Through Rate (CTR)/Conversion Rate (CVR) predictions on two large-scale industrial datasets. We evaluate the proposed method in scenarios with and without overlapping users/items, and extensive experiments verify that the proposed method is able to significantly improve the prediction performance on the target domain. For instance, our method can boost PLE with a performance improvement of 15.4% in terms of Area Under Curve (AUC) compared with single-domain PLE on our private game dataset. In addition, our method is able to surpass single-domain MMoE by 6.85% on the public datasets. Code: https://github.com/TL-UESTC/DAA.
数据稀缺性是推荐系统面临的一个永恒挑战,研究者们提出了多种跨领域推荐方法来缓解目标领域的数据稀缺性问题。然而,在现实世界的许多跨领域推荐系统中,源领域和目标领域的样本来自不同的数据分布,这阻碍了跨领域的知识转移。在本文中,我们提出针对性地对齐源域和目标域之间的数据分布,以缓解样本分布不平衡,从而挑战目标域的数据稀缺性问题。从技术上讲,我们提出的方法构建了一个双对抗性适应(DAA)框架,与预训练的源模型一起对抗性训练目标模型。两个域鉴别器与目标模型进行二人最小最大博弈,引导目标模型学习可跨域迁移的可靠域不变特征。同时,对目标模型进行标定,学习目标域的特定领域信息。此外,我们将我们的方法制定为一个即插即用模块,以促进现有的推荐系统。我们应用所提出的方法来解决两个大规模工业数据集上真实点击率(CTR)/转化率(CVR)预测中数据不足和样本分布不平衡的问题。我们在有和没有重叠用户/项目的场景下评估了所提出的方法,大量的实验验证了所提出的方法能够显著提高目标域的预测性能。例如,与我们的私人游戏数据集上的单域PLE相比,我们的方法可以在曲线下面积(AUC)方面提高15.4%的PLE性能。此外,在公共数据集上,我们的方法能够比单域MMoE高出6.85%。代码:https://github.com/TL-UESTC/DAA。
{"title":"Cross-domain Recommendation via Dual Adversarial Adaptation","authors":"Hongzu Su, Jingjing Li, Zhekai Du, Lei Zhu, Ke Lu, Heng Tao Shen","doi":"10.1145/3632524","DOIUrl":"https://doi.org/10.1145/3632524","url":null,"abstract":"Data scarcity is a perpetual challenge of recommendation systems, and researchers have proposed a variety of cross-domain recommendation methods to alleviate the problem of data scarcity in target domains. However, in many real-world cross-domain recommendation systems, the source domain and the target domain are sampled from different data distributions, which obstructs the cross-domain knowledge transfer. In this paper, we propose to specifically align the data distributions between the source domain and the target domain to alleviate imbalanced sample distribution and thus challenge the data scarcity issue in the target domain. Technically, our proposed approach builds a dual adversarial adaptation (DAA) framework to adversarially train the target model together with a pre-trained source model. Two domain discriminators play the two-player minmax game with the target model and guide the target model to learn reliable domain-invariant features that can be transferred across domains. At the same time, the target model is calibrated to learn domain-specific information of the target domain. In addition, we formulate our approach as a plug-and-play module to boost existing recommendation systems. We apply the proposed method to address the issues of insufficient data and imbalanced sample distribution in real-world Click-Through Rate (CTR)/Conversion Rate (CVR) predictions on two large-scale industrial datasets. We evaluate the proposed method in scenarios with and without overlapping users/items, and extensive experiments verify that the proposed method is able to significantly improve the prediction performance on the target domain. For instance, our method can boost PLE with a performance improvement of 15.4% in terms of Area Under Curve (AUC) compared with single-domain PLE on our private game dataset. In addition, our method is able to surpass single-domain MMoE by 6.85% on the public datasets. Code: https://github.com/TL-UESTC/DAA.","PeriodicalId":50936,"journal":{"name":"ACM Transactions on Information Systems","volume":"12 23","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135087072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ACM Transactions on Information Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1