{"title":"ECAT: A Entire space Continual and Adaptive Transfer Learning Framework for Cross-Domain Recommendation","authors":"Chaoqun Hou, Yuanhang Zhou, Yi Cao, Tong Liu","doi":"arxiv-2407.02542","DOIUrl":null,"url":null,"abstract":"In industrial recommendation systems, there are several mini-apps designed to\nmeet the diverse interests and needs of users. The sample space of them is\nmerely a small subset of the entire space, making it challenging to train an\nefficient model. In recent years, there have been many excellent studies\nrelated to cross-domain recommendation aimed at mitigating the problem of data\nsparsity. However, few of them have simultaneously considered the adaptability\nof both sample and representation continual transfer setting to the target\ntask. To overcome the above issue, we propose a Entire space Continual and\nAdaptive Transfer learning framework called ECAT which includes two core\ncomponents: First, as for sample transfer, we propose a two-stage method that\nrealizes a coarse-to-fine process. Specifically, we perform an initial\nselection through a graph-guided method, followed by a fine-grained selection\nusing domain adaptation method. Second, we propose an adaptive knowledge\ndistillation method for continually transferring the representations from a\nmodel that is well-trained on the entire space dataset. ECAT enables full\nutilization of the entire space samples and representations under the\nsupervision of the target task, while avoiding negative migration.\nComprehensive experiments on real-world industrial datasets from Taobao show\nthat ECAT advances state-of-the-art performance on offline metrics, and brings\n+13.6% CVR and +8.6% orders for Baiyibutie, a famous mini-app of Taobao.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":"2 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.02542","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In industrial recommendation systems, there are several mini-apps designed to
meet the diverse interests and needs of users. The sample space of them is
merely a small subset of the entire space, making it challenging to train an
efficient model. In recent years, there have been many excellent studies
related to cross-domain recommendation aimed at mitigating the problem of data
sparsity. However, few of them have simultaneously considered the adaptability
of both sample and representation continual transfer setting to the target
task. To overcome the above issue, we propose a Entire space Continual and
Adaptive Transfer learning framework called ECAT which includes two core
components: First, as for sample transfer, we propose a two-stage method that
realizes a coarse-to-fine process. Specifically, we perform an initial
selection through a graph-guided method, followed by a fine-grained selection
using domain adaptation method. Second, we propose an adaptive knowledge
distillation method for continually transferring the representations from a
model that is well-trained on the entire space dataset. ECAT enables full
utilization of the entire space samples and representations under the
supervision of the target task, while avoiding negative migration.
Comprehensive experiments on real-world industrial datasets from Taobao show
that ECAT advances state-of-the-art performance on offline metrics, and brings
+13.6% CVR and +8.6% orders for Baiyibutie, a famous mini-app of Taobao.