首页 > 最新文献

Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining最新文献

英文 中文
SMOILE SMOILE
A. Chenreddy, Parshan Pakiman, Selvaprabu Nadarajah, Ranganathan Chandrasekaran, Rick Abens
Product brands employ shopper marketing (SM) strategies to convert shoppers along the path to purchase. Traditional marketing mix models (MMMs), which leverage regression techniques and historical data, can be used to predict the component of sales lift due to SM tactics. The resulting predictive model is a critical input to plan future SM strategies. The implementation of traditional MMMs, however, requires significant ad-hoc manual intervention due to their limited flexibility in (i) explicitly capturing the temporal link between decisions; (ii) accounting for the interaction between business rules and past (sales and decision) data during the attribution of lift to SM; and (iii) ensuring that future decisions adhere to business rules. These issues necessitate MMMs with tailored structures for specific products and retailers, each requiring significant hand-engineering to achieve satisfactory performance -- a major implementation challenge. We propose an SM Optimization and Inverse Learning Engine (SMOILE) that combines optimization and inverse reinforcement learning to streamline implementation. SMOILE learns a model of lift by viewing SM tactic choice as a sequential process, leverages inverse reinforcement learning to explicitly couple sales and decision data, and employs an optimization approach to handle a wide-array of business rules. Using a unique dataset containing sales and SM spend information across retailers and products, we illustrate how SMOILE standardizes the use of data to prescribe future SM decisions. We also track an industry benchmark to showcase the importance of encoding SM lift and decision structures to mitigate spurious results when uncovering the impact of SM decisions.
{"title":"SMOILE","authors":"A. Chenreddy, Parshan Pakiman, Selvaprabu Nadarajah, Ranganathan Chandrasekaran, Rick Abens","doi":"10.1145/3292500.3330788","DOIUrl":"https://doi.org/10.1145/3292500.3330788","url":null,"abstract":"Product brands employ shopper marketing (SM) strategies to convert shoppers along the path to purchase. Traditional marketing mix models (MMMs), which leverage regression techniques and historical data, can be used to predict the component of sales lift due to SM tactics. The resulting predictive model is a critical input to plan future SM strategies. The implementation of traditional MMMs, however, requires significant ad-hoc manual intervention due to their limited flexibility in (i) explicitly capturing the temporal link between decisions; (ii) accounting for the interaction between business rules and past (sales and decision) data during the attribution of lift to SM; and (iii) ensuring that future decisions adhere to business rules. These issues necessitate MMMs with tailored structures for specific products and retailers, each requiring significant hand-engineering to achieve satisfactory performance -- a major implementation challenge. We propose an SM Optimization and Inverse Learning Engine (SMOILE) that combines optimization and inverse reinforcement learning to streamline implementation. SMOILE learns a model of lift by viewing SM tactic choice as a sequential process, leverages inverse reinforcement learning to explicitly couple sales and decision data, and employs an optimization approach to handle a wide-array of business rules. Using a unique dataset containing sales and SM spend information across retailers and products, we illustrate how SMOILE standardizes the use of data to prescribe future SM decisions. We also track an industry benchmark to showcase the importance of encoding SM lift and decision structures to mitigate spurious results when uncovering the impact of SM decisions.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"183 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127034336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Efficient and Effective Express via Contextual Cooperative Reinforcement Learning 基于上下文合作强化学习的高效表达
Yexin Li, Yu Zheng, Qiang Yang
Express systems are widely deployed in many major cities. Couriers in an express system load parcels at transit station and deliver them to customers. Meanwhile, they also try to serve the pick-up requests which come stochastically in real time during the delivery process. Having brought much convenience and promoted the development of e-commerce, express systems face challenges on courier management to complete the massive number of tasks per day. Considering this problem, we propose a reinforcement learning based framework to learn a courier management policy. Firstly, we divide the city into independent regions, in each of which a constant number of couriers deliver parcels and serve requests cooperatively. Secondly, we propose a soft-label clustering algorithm named Balanced Delivery-Service Burden (BDSB) to dispatch parcels to couriers in each region. BDSB guarantees that each courier has almost even delivery and expected request-service burden when departing from transit station, giving a reasonable initialization for online management later. As pick-up requests come in real time, a Contextual Cooperative Reinforcement Learning (CCRL) model is proposed to guide where should each courier deliver and serve in each short period. Being formulated in a multi-agent way, CCRL focuses on the cooperation among couriers while also considering the system context. Experiments on real-world data from Beijing are conducted to confirm the outperformance of our model.
许多大城市都广泛部署了快车系统。快递系统中的快递员在中转站装载包裹并将它们送到客户手中。与此同时,他们也试图在送货过程中实时处理随机出现的提货请求。快递系统在为电子商务的发展带来诸多便利的同时,也面临着快递管理方面的挑战。考虑到这个问题,我们提出了一个基于强化学习的框架来学习快递管理策略。首先,我们将城市划分为独立的区域,每个区域有一定数量的快递员合作投递包裹和服务请求。其次,我们提出了一种名为平衡配送服务负担(BDSB)的软标签聚类算法,将包裹分配给各个地区的快递员。BDSB保证每个快递员从中转站出发时的派送和预期的请求服务负担几乎相等,为后期的在线管理提供了合理的初始化。由于取件请求是实时的,提出了一种上下文合作强化学习(CCRL)模型来指导每个快递员在每个短时间内应该在哪里投递和服务。CCRL以多智能体的方式制定,在考虑系统环境的同时注重快递员之间的合作。对来自北京的真实数据进行了实验,以证实我们的模型的优异性能。
{"title":"Efficient and Effective Express via Contextual Cooperative Reinforcement Learning","authors":"Yexin Li, Yu Zheng, Qiang Yang","doi":"10.1145/3292500.3330968","DOIUrl":"https://doi.org/10.1145/3292500.3330968","url":null,"abstract":"Express systems are widely deployed in many major cities. Couriers in an express system load parcels at transit station and deliver them to customers. Meanwhile, they also try to serve the pick-up requests which come stochastically in real time during the delivery process. Having brought much convenience and promoted the development of e-commerce, express systems face challenges on courier management to complete the massive number of tasks per day. Considering this problem, we propose a reinforcement learning based framework to learn a courier management policy. Firstly, we divide the city into independent regions, in each of which a constant number of couriers deliver parcels and serve requests cooperatively. Secondly, we propose a soft-label clustering algorithm named Balanced Delivery-Service Burden (BDSB) to dispatch parcels to couriers in each region. BDSB guarantees that each courier has almost even delivery and expected request-service burden when departing from transit station, giving a reasonable initialization for online management later. As pick-up requests come in real time, a Contextual Cooperative Reinforcement Learning (CCRL) model is proposed to guide where should each courier deliver and serve in each short period. Being formulated in a multi-agent way, CCRL focuses on the cooperation among couriers while also considering the system context. Experiments on real-world data from Beijing are conducted to confirm the outperformance of our model.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127092311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Large-Scale Training Framework for Video Annotation 视频标注的大规模训练框架
Seong Jae Hwang, Joonseok Lee, Balakrishnan Varadarajan, A. Gordon, Zheng Xu, A. Natsev
Video is one of the richest sources of information available online but extracting deep insights from video content at internet scale is still an open problem, both in terms of depth and breadth of understanding, as well as scale. Over the last few years, the field of video understanding has made great strides due to the availability of large-scale video datasets and core advances in image, audio, and video modeling architectures. However, the state-of-the-art architectures on small scale datasets are frequently impractical to deploy at internet scale, both in terms of the ability to train such deep networks on hundreds of millions of videos, and to deploy them for inference on billions of videos. In this paper, we present a MapReduce-based training framework, which exploits both data parallelism and model parallelism to scale training of complex video models. The proposed framework uses alternating optimization and full-batch fine-tuning, and supports large Mixture-of-Experts classifiers with hundreds of thousands of mixtures, which enables a trade-off between model depth and breadth, and the ability to shift model capacity between shared (generalization) layers and per-class (specialization) layers. We demonstrate that the proposed framework is able to reach state-of-the-art performance on the largest public video datasets, YouTube-8M and Sports-1M, and can scale to 100 times larger datasets.
视频是网上最丰富的信息来源之一,但从互联网规模的视频内容中提取深刻的见解仍然是一个悬而未决的问题,无论是在理解的深度和广度方面,还是在规模方面。在过去的几年中,由于大规模视频数据集的可用性以及图像、音频和视频建模架构的核心进步,视频理解领域取得了巨大的进步。然而,小规模数据集上的最先进架构在互联网规模上部署通常是不切实际的,无论是在数亿个视频上训练这种深度网络的能力,还是在数十亿个视频上部署它们进行推理的能力。在本文中,我们提出了一个基于mapreduce的训练框架,该框架利用数据并行性和模型并行性来扩展复杂视频模型的训练。提出的框架使用交替优化和全批微调,并支持具有数十万个混合物的大型混合专家分类器,这使得模型深度和宽度之间的权衡,以及在共享(泛化)层和每个类(专门化)层之间转换模型容量的能力。我们证明了所提出的框架能够在最大的公共视频数据集(YouTube-8M和Sports-1M)上达到最先进的性能,并且可以扩展到100倍大的数据集。
{"title":"Large-Scale Training Framework for Video Annotation","authors":"Seong Jae Hwang, Joonseok Lee, Balakrishnan Varadarajan, A. Gordon, Zheng Xu, A. Natsev","doi":"10.1145/3292500.3330653","DOIUrl":"https://doi.org/10.1145/3292500.3330653","url":null,"abstract":"Video is one of the richest sources of information available online but extracting deep insights from video content at internet scale is still an open problem, both in terms of depth and breadth of understanding, as well as scale. Over the last few years, the field of video understanding has made great strides due to the availability of large-scale video datasets and core advances in image, audio, and video modeling architectures. However, the state-of-the-art architectures on small scale datasets are frequently impractical to deploy at internet scale, both in terms of the ability to train such deep networks on hundreds of millions of videos, and to deploy them for inference on billions of videos. In this paper, we present a MapReduce-based training framework, which exploits both data parallelism and model parallelism to scale training of complex video models. The proposed framework uses alternating optimization and full-batch fine-tuning, and supports large Mixture-of-Experts classifiers with hundreds of thousands of mixtures, which enables a trade-off between model depth and breadth, and the ability to shift model capacity between shared (generalization) layers and per-class (specialization) layers. We demonstrate that the proposed framework is able to reach state-of-the-art performance on the largest public video datasets, YouTube-8M and Sports-1M, and can scale to 100 times larger datasets.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"2175 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130079575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Origin-Destination Matrix Prediction via Graph Convolution: a New Perspective of Passenger Demand Modeling 基于图卷积的出发地矩阵预测:客运需求建模的新视角
Yuandong Wang, Hongzhi Yin, Hongxu Chen, Tianyu Wo, Jie Xu, Kai Zheng
Ride-hailing applications are becoming more and more popular for providing drivers and passengers with convenient ride services, especially in metropolises like Beijing or New York. To obtain the passengers' mobility patterns, the online platforms of ride services need to predict the number of passenger demands from one region to another in advance. We formulate this problem as an Origin-Destination Matrix Prediction (ODMP) problem. Though this problem is essential to large-scale providers of ride services for helping them make decisions and some providers have already put it forward in public, existing studies have not solved this problem well. One of the main reasons is that the ODMP problem is more challenging than the common demand prediction. Besides the number of demands in a region, it also requires the model to predict the destinations of them. In addition, data sparsity is a severe issue. To solve the problem effectively, we propose a unified model, Grid-Embedding based Multi-task Learning (GEML) which consists of two components focusing on spatial and temporal information respectively. The Grid-Embedding part is designed to model the spatial mobility patterns of passengers and neighboring relationships of different areas, the pre-weighted aggregator of which aims to sense the sparsity and range of data. The Multi-task Learning framework focuses on modeling temporal attributes and capturing several objectives of the ODMP problem. The evaluation of our model is conducted on real operational datasets from UCAR and Didi. The experimental results demonstrate the superiority of our GEML against the state-of-the-art approaches.
网约车应用越来越受欢迎,为司机和乘客提供方便的乘车服务,尤其是在北京或纽约等大都市。为了获得乘客的出行模式,网约车服务平台需要提前预测从一个地区到另一个地区的乘客需求数量。我们将这个问题表述为原点-目的地矩阵预测(ODMP)问题。虽然这个问题对于大规模的乘车服务提供商来说是必不可少的,可以帮助他们做出决策,并且一些提供商已经公开提出了这个问题,但是现有的研究并没有很好地解决这个问题。其中一个主要原因是ODMP问题比常见的需求预测更具挑战性。除了一个地区的需求数量外,还需要模型预测这些需求的目的地。此外,数据稀疏性也是一个严重的问题。为了有效地解决这一问题,我们提出了一种统一的基于网格嵌入的多任务学习(GEML)模型,该模型由两个部分组成,分别关注空间和时间信息。网格嵌入部分用于模拟乘客的空间移动模式和不同区域的相邻关系,其预加权聚合器旨在感知数据的稀疏性和范围。多任务学习框架侧重于建模时间属性和捕获ODMP问题的几个目标。我们的模型的评估是在UCAR和滴滴的真实运营数据集上进行的。实验结果证明了我们的GEML相对于最先进的方法的优越性。
{"title":"Origin-Destination Matrix Prediction via Graph Convolution: a New Perspective of Passenger Demand Modeling","authors":"Yuandong Wang, Hongzhi Yin, Hongxu Chen, Tianyu Wo, Jie Xu, Kai Zheng","doi":"10.1145/3292500.3330877","DOIUrl":"https://doi.org/10.1145/3292500.3330877","url":null,"abstract":"Ride-hailing applications are becoming more and more popular for providing drivers and passengers with convenient ride services, especially in metropolises like Beijing or New York. To obtain the passengers' mobility patterns, the online platforms of ride services need to predict the number of passenger demands from one region to another in advance. We formulate this problem as an Origin-Destination Matrix Prediction (ODMP) problem. Though this problem is essential to large-scale providers of ride services for helping them make decisions and some providers have already put it forward in public, existing studies have not solved this problem well. One of the main reasons is that the ODMP problem is more challenging than the common demand prediction. Besides the number of demands in a region, it also requires the model to predict the destinations of them. In addition, data sparsity is a severe issue. To solve the problem effectively, we propose a unified model, Grid-Embedding based Multi-task Learning (GEML) which consists of two components focusing on spatial and temporal information respectively. The Grid-Embedding part is designed to model the spatial mobility patterns of passengers and neighboring relationships of different areas, the pre-weighted aggregator of which aims to sense the sparsity and range of data. The Multi-task Learning framework focuses on modeling temporal attributes and capturing several objectives of the ODMP problem. The evaluation of our model is conducted on real operational datasets from UCAR and Didi. The experimental results demonstrate the superiority of our GEML against the state-of-the-art approaches.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130177727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 166
Focused Context Balancing for Robust Offline Policy Evaluation 聚焦上下文平衡的鲁棒离线策略评估
Hao Zou, Kun Kuang, Boqi Chen, Peixuan Chen, Peng Cui
Precisely evaluating the effect of new policies (e.g. ad-placement models, recommendation functions, ranking functions) is one of the most important problems for improving interactive systems. The conventional policy evaluation methods rely on online A/B tests, but they are usually extremely expensive and may have undesirable impacts. Recently, Inverse Propensity Score (IPS) estimators are proposed as alternatives to evaluate the effect of new policy with offline logged data that was collected from a different policy in the past. They tend to remove the distribution shift induced by past policy. However, they ignore the distribution shift that would be induced by the new policy, which results in imprecise evaluation. Moreover, their performances rely on accurate estimation of propensity score, which can not be guaranteed or validated in practice. In this paper, we propose a non-parametric method, named Focused Context Balancing (FCB) algorithm, to learn sample weights for context balancing, so that the distribution shift induced by the past policy and new policy can be eliminated respectively. To validate the effectiveness of our FCB algorithm, we conduct extensive experiments on both synthetic and real world datasets. The experimental results clearly demonstrate that our FCB algorithm outperforms existing estimators by achieving more precise and robust results for offline policy evaluation.
准确评估新策略(例如广告投放模型、推荐功能、排名功能)的效果是改进交互系统的最重要问题之一。传统的策略评估方法依赖于在线A/B测试,但它们通常非常昂贵,并且可能产生不良影响。最近,人们提出了逆倾向评分(IPS)估计器,作为评估新策略效果的替代方法,该策略使用过去从不同策略收集的离线日志数据。他们倾向于消除由过去政策引起的分配转移。然而,他们忽略了新政策可能引起的分布变化,这导致了不精确的评估。此外,它们的性能依赖于倾向得分的准确估计,这在实践中无法保证或验证。在本文中,我们提出了一种非参数的方法,即聚焦上下文平衡(FCB)算法,来学习上下文平衡的样本权重,从而分别消除由过去策略和新策略引起的分布偏移。为了验证我们的FCB算法的有效性,我们在合成和真实世界的数据集上进行了大量的实验。实验结果清楚地表明,我们的FCB算法通过获得更精确和鲁棒的离线策略评估结果,优于现有的估计器。
{"title":"Focused Context Balancing for Robust Offline Policy Evaluation","authors":"Hao Zou, Kun Kuang, Boqi Chen, Peixuan Chen, Peng Cui","doi":"10.1145/3292500.3330852","DOIUrl":"https://doi.org/10.1145/3292500.3330852","url":null,"abstract":"Precisely evaluating the effect of new policies (e.g. ad-placement models, recommendation functions, ranking functions) is one of the most important problems for improving interactive systems. The conventional policy evaluation methods rely on online A/B tests, but they are usually extremely expensive and may have undesirable impacts. Recently, Inverse Propensity Score (IPS) estimators are proposed as alternatives to evaluate the effect of new policy with offline logged data that was collected from a different policy in the past. They tend to remove the distribution shift induced by past policy. However, they ignore the distribution shift that would be induced by the new policy, which results in imprecise evaluation. Moreover, their performances rely on accurate estimation of propensity score, which can not be guaranteed or validated in practice. In this paper, we propose a non-parametric method, named Focused Context Balancing (FCB) algorithm, to learn sample weights for context balancing, so that the distribution shift induced by the past policy and new policy can be eliminated respectively. To validate the effectiveness of our FCB algorithm, we conduct extensive experiments on both synthetic and real world datasets. The experimental results clearly demonstrate that our FCB algorithm outperforms existing estimators by achieving more precise and robust results for offline policy evaluation.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132495902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Reserve Price Failure Rate Prediction with Header Bidding in Display Advertising 基于标题竞价的显示广告保留价失效率预测
Achir Kalra, Chong Wang, C. Borcea, Yi Chen
The revenue of online display advertising in the U.S. is projected to be 7.9 billion U.S. dollars by 2022. One main way of display advertising is through real-time bidding (RTB). In RTB, an ad exchange runs a second price auction among multiple advertisers to sell each ad impression. Publishers usually set up a reserve price, the lowest price acceptable for an ad impression. If there are bids higher than the reserve price, then the revenue is the higher price between the reserve price and the second highest bid; otherwise, the revenue is zero. Thus, a higher reserve price can potentially increase the revenue, but with higher risks associated. In this paper, we study the problem of estimating the failure rate of a reserve price, i.e., the probability that a reserve price fails to be outbid. The solution to this problem have managerial implications to publishers to set appropriate reserve prices in order to minimizes the risks and optimize the expected revenue. This problem is highly challenging since most publishers do not know the historical highest bidding prices offered by RTB advertisers. To address this problem, we develop a parametric survival model for reserve price failure rate prediction. The model is further improved by considering user and page interactions, and header bidding information. The experimental results demonstrate the effectiveness of the proposed approach.
到2022年,美国在线展示广告的收入预计将达到79亿美元。展示广告的一种主要方式是实时竞价(RTB)。在RTB中,广告交易所在多个广告商之间进行第二次价格拍卖,以出售每个广告印象。发行商通常会设定一个底价,即广告印象可接受的最低价格。如果有出价高于底价,则收入为底价与第二高出价之间的较高价格;否则,收入为零。因此,更高的底价可能会增加收入,但风险也会更高。在本文中,我们研究了保留价格失败率的估计问题,即保留价格没有被出价的概率。这一问题的解决方案对发行商来说具有管理意义,即设置适当的保留价格以最小化风险并优化预期收益。这个问题非常具有挑战性,因为大多数发行商并不知道RTB广告商提供的历史最高出价。为了解决这个问题,我们建立了一个参数生存模型来预测储备价格失效率。通过考虑用户和页面交互以及标头竞价信息,进一步改进了模型。实验结果证明了该方法的有效性。
{"title":"Reserve Price Failure Rate Prediction with Header Bidding in Display Advertising","authors":"Achir Kalra, Chong Wang, C. Borcea, Yi Chen","doi":"10.1145/3292500.3330729","DOIUrl":"https://doi.org/10.1145/3292500.3330729","url":null,"abstract":"The revenue of online display advertising in the U.S. is projected to be 7.9 billion U.S. dollars by 2022. One main way of display advertising is through real-time bidding (RTB). In RTB, an ad exchange runs a second price auction among multiple advertisers to sell each ad impression. Publishers usually set up a reserve price, the lowest price acceptable for an ad impression. If there are bids higher than the reserve price, then the revenue is the higher price between the reserve price and the second highest bid; otherwise, the revenue is zero. Thus, a higher reserve price can potentially increase the revenue, but with higher risks associated. In this paper, we study the problem of estimating the failure rate of a reserve price, i.e., the probability that a reserve price fails to be outbid. The solution to this problem have managerial implications to publishers to set appropriate reserve prices in order to minimizes the risks and optimize the expected revenue. This problem is highly challenging since most publishers do not know the historical highest bidding prices offered by RTB advertisers. To address this problem, we develop a parametric survival model for reserve price failure rate prediction. The model is further improved by considering user and page interactions, and header bidding information. The experimental results demonstrate the effectiveness of the proposed approach.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132981788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Disambiguation Enabled Linear Discriminant Analysis for Partial Label Dimensionality Reduction 局部标签降维的消歧线性判别分析
Jing-Han Wu, Min-Ling Zhang
Partial label learning is an emerging weakly-supervised learning framework where each training example is associated with multiple candidate labels among which only one is valid. Dimensionality reduction serves as an effective way to help improve the generalization ability of learning system, while the task of partial label dimensionality reduction is challenging due to the unknown ground-truth labeling information. In this paper, the first attempt towards partial label dimensionality reduction is investigated by endowing the popular linear discriminant analysis (LDA) techniques with the ability of dealing with partial label training examples. Specifically, a novel learning procedure named DELIN is proposed which alternates between LDA dimensionality reduction and candidate label disambiguation based on estimated labeling confidences over candidate labels. On one hand, the projection matrix of LDA is optimized by utilizing disambiguation-guided labeling confidences. On the other hand, the labeling confidences are disambiguated by resorting to kNN aggregation in the LDA-induced feature space. Extensive experiments on synthetic as well as real-world partial label data sets clearly validate the effectiveness of DELIN in improving the generalization ability of state-of-the-art partial label learning algorithms.
部分标签学习是一种新兴的弱监督学习框架,其中每个训练样例与多个候选标签相关联,其中只有一个是有效的。降维是提高学习系统泛化能力的有效途径,但由于未知的真值标记信息,部分标记的降维任务具有挑战性。本文通过赋予流行的线性判别分析(LDA)技术处理偏标签训练样例的能力,对偏标签降维进行了首次尝试。具体而言,提出了一种新的学习过程DELIN,该过程在LDA降维和候选标签消歧之间交替进行,该过程基于候选标签上估计的标签置信度。一方面,利用消歧引导的标注置信度对LDA的投影矩阵进行优化。另一方面,通过在lda诱导的特征空间中使用kNN聚合来消除标注置信度的歧义。在合成和现实世界的部分标签数据集上进行的大量实验清楚地验证了DELIN在提高最先进的部分标签学习算法的泛化能力方面的有效性。
{"title":"Disambiguation Enabled Linear Discriminant Analysis for Partial Label Dimensionality Reduction","authors":"Jing-Han Wu, Min-Ling Zhang","doi":"10.1145/3292500.3330901","DOIUrl":"https://doi.org/10.1145/3292500.3330901","url":null,"abstract":"Partial label learning is an emerging weakly-supervised learning framework where each training example is associated with multiple candidate labels among which only one is valid. Dimensionality reduction serves as an effective way to help improve the generalization ability of learning system, while the task of partial label dimensionality reduction is challenging due to the unknown ground-truth labeling information. In this paper, the first attempt towards partial label dimensionality reduction is investigated by endowing the popular linear discriminant analysis (LDA) techniques with the ability of dealing with partial label training examples. Specifically, a novel learning procedure named DELIN is proposed which alternates between LDA dimensionality reduction and candidate label disambiguation based on estimated labeling confidences over candidate labels. On one hand, the projection matrix of LDA is optimized by utilizing disambiguation-guided labeling confidences. On the other hand, the labeling confidences are disambiguated by resorting to kNN aggregation in the LDA-induced feature space. Extensive experiments on synthetic as well as real-world partial label data sets clearly validate the effectiveness of DELIN in improving the generalization ability of state-of-the-art partial label learning algorithms.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129249624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Diagnosing Sample Ratio Mismatch in Online Controlled Experiments: A Taxonomy and Rules of Thumb for Practitioners 在线控制实验诊断样本比例不匹配:从业者的分类和经验规则
Aleksander Fabijan, J. Gupchup, Somit Gupta, Jeff Omhover, Wen Qin, Lukas Vermeer, Pavel A. Dmitriev
Accurately learning what delivers value to customers is difficult. Online Controlled Experiments (OCEs), aka A/B tests, are becoming a standard operating procedure in software companies to address this challenge as they can detect small causal changes in user behavior due to product modifications (e.g. new features). However, like any data analysis method, OCEs are sensitive to trustworthiness and data quality issues which, if go unaddressed or unnoticed, may result in making wrong decisions. One of the most useful indicators of a variety of data quality issues is a Sample Ratio Mismatch (SRM) ? the situation when the observed sample ratio in the experiment is different from the expected. Just like fever is a symptom for multiple types of illness, an SRM is a symptom for a variety of data quality issues. While a simple statistical check is used to detect an SRM, correctly identifying the root cause and preventing it from happening in the future is often extremely challenging and time consuming. Ignoring the SRM without knowing the root cause may result in a bad product modification appearing to be good and getting shipped to users, or vice versa. The goal of this paper is to make diagnosing, fixing, and preventing SRMs easier. Based on our experience of running OCEs in four different software companies in over 25 different products used by hundreds of millions of users worldwide, we have derived a taxonomy for different types of SRMs. We share examples, detection guidelines, and best practices for preventing SRMs of each type. We hope that the lessons and practical tips we describe in this paper will speed up SRM investigations and prevent some of them. Ultimately, this should lead to improved decision making based on trustworthy experiment analysis.
准确地了解什么能为客户带来价值是很困难的。在线控制实验(OCEs),又名A/B测试,正在成为软件公司应对这一挑战的标准操作程序,因为它们可以检测到由于产品修改(例如新功能)而导致的用户行为的微小因果变化。然而,像任何数据分析方法一样,OCEs对可信度和数据质量问题很敏感,如果不加以解决或不被注意,可能会导致做出错误的决策。各种数据质量问题最有用的指标之一是样本比例不匹配(SRM) ?实验中观察到的样本比与期望的不一致的情况。就像发烧是多种疾病的症状一样,SRM是各种数据质量问题的症状。虽然可以使用简单的统计检查来检测SRM,但正确识别根本原因并防止其在将来发生通常是极具挑战性和耗时的。在不知道根本原因的情况下忽略SRM可能会导致一个糟糕的产品修改看起来很好,并被交付给用户,反之亦然。本文的目标是简化srm的诊断、修复和预防。根据我们在四家不同的软件公司中运行OCEs的经验,在全球数亿用户使用的超过25种不同的产品中,我们得出了不同类型srm的分类。我们将分享用于预防每种类型srm的示例、检测指南和最佳实践。我们希望我们在本文中描述的经验教训和实用技巧将加速SRM调查并防止其中的一些。最终,这将导致基于可信实验分析的改进决策。
{"title":"Diagnosing Sample Ratio Mismatch in Online Controlled Experiments: A Taxonomy and Rules of Thumb for Practitioners","authors":"Aleksander Fabijan, J. Gupchup, Somit Gupta, Jeff Omhover, Wen Qin, Lukas Vermeer, Pavel A. Dmitriev","doi":"10.1145/3292500.3330722","DOIUrl":"https://doi.org/10.1145/3292500.3330722","url":null,"abstract":"Accurately learning what delivers value to customers is difficult. Online Controlled Experiments (OCEs), aka A/B tests, are becoming a standard operating procedure in software companies to address this challenge as they can detect small causal changes in user behavior due to product modifications (e.g. new features). However, like any data analysis method, OCEs are sensitive to trustworthiness and data quality issues which, if go unaddressed or unnoticed, may result in making wrong decisions. One of the most useful indicators of a variety of data quality issues is a Sample Ratio Mismatch (SRM) ? the situation when the observed sample ratio in the experiment is different from the expected. Just like fever is a symptom for multiple types of illness, an SRM is a symptom for a variety of data quality issues. While a simple statistical check is used to detect an SRM, correctly identifying the root cause and preventing it from happening in the future is often extremely challenging and time consuming. Ignoring the SRM without knowing the root cause may result in a bad product modification appearing to be good and getting shipped to users, or vice versa. The goal of this paper is to make diagnosing, fixing, and preventing SRMs easier. Based on our experience of running OCEs in four different software companies in over 25 different products used by hundreds of millions of users worldwide, we have derived a taxonomy for different types of SRMs. We share examples, detection guidelines, and best practices for preventing SRMs of each type. We hope that the lessons and practical tips we describe in this paper will speed up SRM investigations and prevent some of them. Ultimately, this should lead to improved decision making based on trustworthy experiment analysis.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128791290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Transportation 运输
Jieping Ye
As decarbonisation is becoming increasingly important, many countries have placed an emphasis on decarbonising their transportation sector through electrification to support the transition to net zero. As such, research regarding the adoption of electric vehicles has drastically increased in recent years. Mathematical modelling plays an important role in optimising a transition to electric vehicles. This article describes a systematic literature review of existing works which perform mathematical modelling of the adoption of electric motor vehicles. In this study, 53 articles containing mathematical models of electric vehicle adoption are reviewed systematically to answer 6 research questions regarding the process of modelling transitions to electric vehicles. The mathematical modelling techniques observed in existing literature are discussed, along with the main barriers to electric vehicle adoption, and future research directions are suggested.
随着脱碳变得越来越重要,许多国家都强调通过电气化使交通运输部门脱碳,以支持向净零碳的过渡。因此,关于采用电动汽车的研究近年来急剧增加。数学建模在优化过渡到电动汽车方面发挥着重要作用。这篇文章描述了一个系统的文献综述,现有的工作,执行数学模型的采用电动汽车。在本研究中,系统地回顾了53篇包含电动汽车采用数学模型的文章,以回答有关建模过渡到电动汽车过程的6个研究问题。讨论了现有文献中观察到的数学建模技术,以及电动汽车采用的主要障碍,并提出了未来的研究方向。
{"title":"Transportation","authors":"Jieping Ye","doi":"10.1145/3292500.3340406","DOIUrl":"https://doi.org/10.1145/3292500.3340406","url":null,"abstract":"As decarbonisation is becoming increasingly important, many countries have placed an emphasis on decarbonising their transportation sector through electrification to support the transition to net zero. As such, research regarding the adoption of electric vehicles has drastically increased in recent years. Mathematical modelling plays an important role in optimising a transition to electric vehicles. This article describes a systematic literature review of existing works which perform mathematical modelling of the adoption of electric motor vehicles. In this study, 53 articles containing mathematical models of electric vehicle adoption are reviewed systematically to answer 6 research questions regarding the process of modelling transitions to electric vehicles. The mathematical modelling techniques observed in existing literature are discussed, along with the main barriers to electric vehicle adoption, and future research directions are suggested.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"280 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123307424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Hierarchical Multi-Task Word Embedding Learning for Synonym Prediction 面向同义词预测的分层多任务词嵌入学习
Hongliang Fei, Shulong Tan, Ping Li
Automatic synonym recognition is of great importance for entity-centric text mining and interpretation. Due to the high language use variability in real-life, manual construction of semantic resources to cover all synonyms is prohibitively expensive and may also result in limited coverage. Although there are public knowledge bases, they only have limited coverage for languages other than English. In this paper, we focus on medical domain and propose an automatic way to accelerate the process of medical synonymy resource development for Chinese, including both formal entities from healthcare professionals and noisy descriptions from end-users. Motivated by the success of distributed word representations, we design a multi-task model with hierarchical task relationship to learn more representative entity/term embeddings and apply them to synonym prediction. In our model, we extend the classical skip-gram word embedding model by introducing an auxiliary task "neighboring word semantic type prediction'' and hierarchically organize them based on the task complexity. Meanwhile, we incorporate existing medical term-term synonymous knowledge into our word embedding learning framework. We demonstrate that the embeddings trained from our proposed multi-task model yield significant improvement for entity semantic relatedness evaluation, neighboring word semantic type prediction and synonym prediction compared with baselines. Furthermore, we create a large medical text corpus in Chinese that includes annotations for entities, descriptions and synonymous pairs for future research in this direction.
自动同义词识别对于以实体为中心的文本挖掘和解释具有重要意义。由于现实生活中语言使用的高度可变性,手动构建语义资源以覆盖所有同义词是非常昂贵的,并且也可能导致有限的覆盖。虽然有公共知识库,但它们对英语以外语言的覆盖范围有限。本文以医学领域为研究对象,提出了一种自动加速中文医学同义词资源开发的方法,包括来自医疗专业人员的正式实体和来自最终用户的嘈杂描述。受分布式词表示成功的启发,我们设计了一个具有分层任务关系的多任务模型来学习更具代表性的实体/术语嵌入,并将其应用于同义词预测。在该模型中,我们通过引入一个辅助任务“邻词语义类型预测”来扩展经典的跳格词嵌入模型,并根据任务复杂度对其进行分层组织。同时,我们将已有的医学术语同义知识整合到我们的词嵌入学习框架中。我们证明,与基线相比,我们提出的多任务模型训练的嵌入在实体语义相关性评估、相邻词语义类型预测和同义词预测方面有显著改善。此外,我们还创建了一个大型中文医学文本语料库,其中包括对实体、描述和同义对的注释,以供未来的研究方向使用。
{"title":"Hierarchical Multi-Task Word Embedding Learning for Synonym Prediction","authors":"Hongliang Fei, Shulong Tan, Ping Li","doi":"10.1145/3292500.3330914","DOIUrl":"https://doi.org/10.1145/3292500.3330914","url":null,"abstract":"Automatic synonym recognition is of great importance for entity-centric text mining and interpretation. Due to the high language use variability in real-life, manual construction of semantic resources to cover all synonyms is prohibitively expensive and may also result in limited coverage. Although there are public knowledge bases, they only have limited coverage for languages other than English. In this paper, we focus on medical domain and propose an automatic way to accelerate the process of medical synonymy resource development for Chinese, including both formal entities from healthcare professionals and noisy descriptions from end-users. Motivated by the success of distributed word representations, we design a multi-task model with hierarchical task relationship to learn more representative entity/term embeddings and apply them to synonym prediction. In our model, we extend the classical skip-gram word embedding model by introducing an auxiliary task \"neighboring word semantic type prediction'' and hierarchically organize them based on the task complexity. Meanwhile, we incorporate existing medical term-term synonymous knowledge into our word embedding learning framework. We demonstrate that the embeddings trained from our proposed multi-task model yield significant improvement for entity semantic relatedness evaluation, neighboring word semantic type prediction and synonym prediction compared with baselines. Furthermore, we create a large medical text corpus in Chinese that includes annotations for entities, descriptions and synonymous pairs for future research in this direction.","PeriodicalId":186134,"journal":{"name":"Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126419338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
期刊
Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1