首页 > 最新文献

Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining最新文献

英文 中文
Peeking at A/B Tests: Why it matters, and what to do about it 窥探A/B测试:为什么它很重要,以及如何做
Ramesh Johari, P. Koomen, L. Pekelis, David Walsh
This paper reports on novel statistical methodology, which has been deployed by the commercial A/B testing platform Optimizely to communicate experimental results to their customers. Our methodology addresses the issue that traditional p-values and confidence intervals give unreliable inference. This is because users of A/B testing software are known to continuously monitor these measures as the experiment is running. We provide always valid p-values and confidence intervals that are provably robust to this effect. Not only does this make it safe for a user to continuously monitor, but it empowers her to detect true effects more efficiently. This paper provides simulations and numerical studies on Optimizely's data, demonstrating an improvement in detection performance over traditional methods.
本文报告了一种新的统计方法,该方法已被商业A/B测试平台optimely部署,用于向客户传达实验结果。我们的方法解决了传统p值和置信区间给出不可靠推断的问题。这是因为众所周知,A/B测试软件的用户会在实验运行过程中持续监控这些度量。我们提供了始终有效的p值和可证明对这种效应具有鲁棒性的置信区间。这不仅可以让用户安全地持续监控,还可以让用户更有效地检测到真实的效果。本文对optimely的数据进行了仿真和数值研究,证明了与传统方法相比,该方法的检测性能有所提高。
{"title":"Peeking at A/B Tests: Why it matters, and what to do about it","authors":"Ramesh Johari, P. Koomen, L. Pekelis, David Walsh","doi":"10.1145/3097983.3097992","DOIUrl":"https://doi.org/10.1145/3097983.3097992","url":null,"abstract":"This paper reports on novel statistical methodology, which has been deployed by the commercial A/B testing platform Optimizely to communicate experimental results to their customers. Our methodology addresses the issue that traditional p-values and confidence intervals give unreliable inference. This is because users of A/B testing software are known to continuously monitor these measures as the experiment is running. We provide always valid p-values and confidence intervals that are provably robust to this effect. Not only does this make it safe for a user to continuously monitor, but it empowers her to detect true effects more efficiently. This paper provides simulations and numerical studies on Optimizely's data, demonstrating an improvement in detection performance over traditional methods.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133932524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 138
Learning to Count Mosquitoes for the Sterile Insect Technique 学习为昆虫不育技术计算蚊子
Yaniv Ovadia, Yoni Halpern, Dilip Krishnan, Josh Livni, Daniel E. Newburger, R. Poplin, Tiantian Zha, D. Sculley
Mosquito-borne illnesses such as dengue, chikungunya, and Zika are major global health problems, which are not yet addressable with vaccines and must be countered by reducing mosquito populations. The Sterile Insect Technique (SIT) is a promising alternative to pesticides; however, effective SIT relies on minimal releases of female insects. This paper describes a multi-objective convolutional neural net to significantly streamline the process of counting male and female mosquitoes released from a SIT factory and provides a statistical basis for verifying strict contamination rate limits from these counts despite measurement noise. These results are a promising indication that such methods may dramatically reduce the cost of effective SIT methods in practice.
登革热、基孔肯雅热和寨卡等蚊媒疾病是全球主要的健康问题,目前还无法用疫苗解决,必须通过减少蚊子数量来应对。昆虫不育技术是一种很有前途的农药替代技术。然而,有效的SIT依赖于雌性昆虫的最小释放。本文描述了一个多目标卷积神经网络,以显着简化从SIT工厂释放的雄性和雌性蚊子的计数过程,并为验证这些计数的严格污染率限制提供了统计基础,尽管测量噪声。这些结果是一个有希望的迹象,表明这些方法可以在实践中显著降低有效SIT方法的成本。
{"title":"Learning to Count Mosquitoes for the Sterile Insect Technique","authors":"Yaniv Ovadia, Yoni Halpern, Dilip Krishnan, Josh Livni, Daniel E. Newburger, R. Poplin, Tiantian Zha, D. Sculley","doi":"10.1145/3097983.3098204","DOIUrl":"https://doi.org/10.1145/3097983.3098204","url":null,"abstract":"Mosquito-borne illnesses such as dengue, chikungunya, and Zika are major global health problems, which are not yet addressable with vaccines and must be countered by reducing mosquito populations. The Sterile Insect Technique (SIT) is a promising alternative to pesticides; however, effective SIT relies on minimal releases of female insects. This paper describes a multi-objective convolutional neural net to significantly streamline the process of counting male and female mosquitoes released from a SIT factory and provides a statistical basis for verifying strict contamination rate limits from these counts despite measurement noise. These results are a promising indication that such methods may dramatically reduce the cost of effective SIT methods in practice.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132771494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Data Mining Framework for Valuing Large Portfolios of Variable Annuities 大型可变年金投资组合估值的数据挖掘框架
Guojun Gan, Xiangji Huang
A variable annuity is a tax-deferred retirement vehicle created to address concerns that many people have about outliving their assets. In the past decade, the rapid growth of variable annuities has posed great challenges to insurance companies especially when it comes to valuing the complex guarantees embedded in these products. In this paper, we propose a novel data mining framework to address the computational issue associated with the valuation of large portfolios of variable annuity contracts. The data mining framework consists of two major components: a data clustering algorithm which is used to select representative variable annuity contracts, and a regression model which is used to predict quantities of interest for the whole portfolio based on the representative contracts. A series of numerical experiments are conducted on a portfolio of synthetic variable annuity contracts to demonstrate the performance of our proposed data mining framework in terms of accuracy and speed. The experimental results show that our proposed framework is able to produce accurate estimates of various quantities of interest and can reduce the runtime significantly.
可变年金是一种延税退休工具,旨在解决许多人对资产寿命的担忧。在过去的十年里,可变年金的快速增长给保险公司带来了巨大的挑战,尤其是在评估这些产品所包含的复杂担保时。在本文中,我们提出了一个新的数据挖掘框架来解决与大型可变年金合约组合估值相关的计算问题。数据挖掘框架由两个主要部分组成:用于选择具有代表性的可变年金合约的数据聚类算法,以及用于基于代表性合约预测整个投资组合的利息数量的回归模型。通过对一组合成可变年金合约组合进行数值实验,验证了本文提出的数据挖掘框架在准确性和速度方面的性能。实验结果表明,我们提出的框架能够准确地估计各种感兴趣的数量,并且可以显着减少运行时间。
{"title":"A Data Mining Framework for Valuing Large Portfolios of Variable Annuities","authors":"Guojun Gan, Xiangji Huang","doi":"10.1145/3097983.3098013","DOIUrl":"https://doi.org/10.1145/3097983.3098013","url":null,"abstract":"A variable annuity is a tax-deferred retirement vehicle created to address concerns that many people have about outliving their assets. In the past decade, the rapid growth of variable annuities has posed great challenges to insurance companies especially when it comes to valuing the complex guarantees embedded in these products. In this paper, we propose a novel data mining framework to address the computational issue associated with the valuation of large portfolios of variable annuity contracts. The data mining framework consists of two major components: a data clustering algorithm which is used to select representative variable annuity contracts, and a regression model which is used to predict quantities of interest for the whole portfolio based on the representative contracts. A series of numerical experiments are conducted on a portfolio of synthetic variable annuity contracts to demonstrate the performance of our proposed data mining framework in terms of accuracy and speed. The experimental results show that our proposed framework is able to produce accurate estimates of various quantities of interest and can reduce the runtime significantly.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133000180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Local Algorithm for User Action Prediction Towards Display Ads 面向展示广告的用户行为预测局部算法
Hongxia Yang, Yada Zhu, Jingrui He
User behavior modeling is essential in computational advertisement, which builds users' profiles by tracking their online behaviors and then delivers the relevant ads according to each user's interests and needs. Accurate models will lead to higher targeting accuracy and thus improved advertising performance. Intuitively, similar users tend to have similar behaviors towards the displayed ads (e.g., impression, click, conversion). However, to the best of our knowledge, there is not much previous work that explicitly investigates such similarities of various types of user behaviors, and incorporates them into ad response targeting and prediction, largely due to the prohibitive scale of the problem. To bridge this gap, in this paper, we use bipartite graphs to represent historical user behaviors, which consist of both user nodes and advertiser campaign nodes, as well as edges reflecting various types of user-campaign interactions in the past. Based on this representation, we study random-walk-based local algorithms for user behavior modeling and action prediction, whose computational complexity depends only on the size of the output cluster, rather than the entire graph. Our goal is to improve action prediction by leveraging historical user-user, campaign-campaign, and user-campaign interactions. In particular, we propose the bipartite graphs AdvUserGraph accompanied with the ADNI algorithm. ADNI extends the NIBBLE algorithm to AdvUserGraph, and it is able to find the local cluster consisting of interested users towards a specific advertiser campaign. We also propose two extensions of ADNI with improved efficiencies. The performance of the proposed algorithms is demonstrated on both synthetic data and a world leading Demand Side Platform (DSP), showing that they are able to discriminate extremely rare events in terms of their action propensity.
用户行为建模在计算广告中是必不可少的,它通过跟踪用户的在线行为来建立用户档案,然后根据每个用户的兴趣和需求提供相关的广告。准确的模型将导致更高的定位准确性,从而提高广告效果。直观地看,相似的用户倾向于对显示的广告有相似的行为(例如,印象、点击、转换)。然而,据我们所知,之前并没有多少研究明确调查各种类型用户行为的相似性,并将其纳入广告响应定位和预测中,这主要是由于问题的规模过大。为了弥补这一差距,在本文中,我们使用二部图来表示历史用户行为,它由用户节点和广告商活动节点组成,以及反映过去各种类型的用户活动交互的边。基于这种表示,我们研究了基于随机行走的用户行为建模和动作预测的局部算法,其计算复杂度仅取决于输出簇的大小,而不是整个图。我们的目标是通过利用历史用户-用户、活动-活动和用户-活动互动来改进行为预测。特别地,我们提出了与ADNI算法相结合的二部图AdvUserGraph。ADNI将NIBBLE算法扩展到AdvUserGraph,它能够找到由对特定广告活动感兴趣的用户组成的本地集群。我们还提出了两种提高效率的ADNI扩展。在合成数据和世界领先的需求侧平台(DSP)上证明了所提出算法的性能,表明它们能够根据其行为倾向区分极其罕见的事件。
{"title":"Local Algorithm for User Action Prediction Towards Display Ads","authors":"Hongxia Yang, Yada Zhu, Jingrui He","doi":"10.1145/3097983.3098089","DOIUrl":"https://doi.org/10.1145/3097983.3098089","url":null,"abstract":"User behavior modeling is essential in computational advertisement, which builds users' profiles by tracking their online behaviors and then delivers the relevant ads according to each user's interests and needs. Accurate models will lead to higher targeting accuracy and thus improved advertising performance. Intuitively, similar users tend to have similar behaviors towards the displayed ads (e.g., impression, click, conversion). However, to the best of our knowledge, there is not much previous work that explicitly investigates such similarities of various types of user behaviors, and incorporates them into ad response targeting and prediction, largely due to the prohibitive scale of the problem. To bridge this gap, in this paper, we use bipartite graphs to represent historical user behaviors, which consist of both user nodes and advertiser campaign nodes, as well as edges reflecting various types of user-campaign interactions in the past. Based on this representation, we study random-walk-based local algorithms for user behavior modeling and action prediction, whose computational complexity depends only on the size of the output cluster, rather than the entire graph. Our goal is to improve action prediction by leveraging historical user-user, campaign-campaign, and user-campaign interactions. In particular, we propose the bipartite graphs AdvUserGraph accompanied with the ADNI algorithm. ADNI extends the NIBBLE algorithm to AdvUserGraph, and it is able to find the local cluster consisting of interested users towards a specific advertiser campaign. We also propose two extensions of ADNI with improved efficiencies. The performance of the proposed algorithms is demonstrated on both synthetic data and a world leading Demand Side Platform (DSP), showing that they are able to discriminate extremely rare events in terms of their action propensity.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133233467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Sparse Compositional Local Metric Learning 稀疏组合局部度量学习
J. S. Amand, Jun Huan
Mahalanobis distance metric learning becomes an especially challenging problem as the dimension of the feature space p is scaled upwards. The number of parameters to optimize grows with space complexity of order O (p 2), making storage infeasible, interpretability poor, and causing the model to have a high tendency to overfit. Additionally, optimization while maintaining feasibility of the solution becomes prohibitively expensive, requiring a projection onto the positive semi-definite cone after every iteration. In addition to the obvious space and computational challenges, vanilla distance metric learning is unable to model complex and multi-modal trends in the data. Inspired by the recent resurgence of Frank-Wolfe style optimization, we propose a new method for sparse compositional local Mahalanobis distance metric learning. Our proposed technique learns a set of distance metrics which are composed of local and global components. We capture local interactions in the feature space, while ensuring that all metrics share a global component, which may act as a regularizer. We optimize our model using an alternating pairwise Frank-Wolfe style algorithm. This serves a dual purpose, we can control the sparsity of our solution, and altogether avoid any expensive projection operations. Finally, we conduct an empirical evaluation of our method with the current state of the art and present the results on five datasets from varying domains.
随着特征空间p维度的增大,马氏距离度量学习成为一个特别具有挑战性的问题。需要优化的参数数量随着空间复杂度O (p 2)阶的增长而增长,使得存储变得不可行,可解释性差,并且导致模型具有很高的过拟合倾向。此外,优化同时保持解决方案的可行性变得非常昂贵,每次迭代后都需要在正半定锥上进行投影。除了明显的空间和计算挑战外,传统的距离度量学习无法对数据中的复杂和多模态趋势进行建模。受Frank-Wolfe风格优化的启发,我们提出了一种稀疏组合局部Mahalanobis距离度量学习的新方法。我们提出的技术学习一组由局部和全局分量组成的距离度量。我们在特征空间中捕获局部交互,同时确保所有指标共享一个全局组件,该组件可以作为正则化器。我们使用交替的两两Frank-Wolfe算法来优化我们的模型。这有双重目的,我们可以控制我们的解决方案的稀疏性,并完全避免任何昂贵的投影操作。最后,我们对我们的方法进行了实证评估,并在来自不同领域的五个数据集上展示了结果。
{"title":"Sparse Compositional Local Metric Learning","authors":"J. S. Amand, Jun Huan","doi":"10.1145/3097983.3098153","DOIUrl":"https://doi.org/10.1145/3097983.3098153","url":null,"abstract":"Mahalanobis distance metric learning becomes an especially challenging problem as the dimension of the feature space p is scaled upwards. The number of parameters to optimize grows with space complexity of order O (p 2), making storage infeasible, interpretability poor, and causing the model to have a high tendency to overfit. Additionally, optimization while maintaining feasibility of the solution becomes prohibitively expensive, requiring a projection onto the positive semi-definite cone after every iteration. In addition to the obvious space and computational challenges, vanilla distance metric learning is unable to model complex and multi-modal trends in the data. Inspired by the recent resurgence of Frank-Wolfe style optimization, we propose a new method for sparse compositional local Mahalanobis distance metric learning. Our proposed technique learns a set of distance metrics which are composed of local and global components. We capture local interactions in the feature space, while ensuring that all metrics share a global component, which may act as a regularizer. We optimize our model using an alternating pairwise Frank-Wolfe style algorithm. This serves a dual purpose, we can control the sparsity of our solution, and altogether avoid any expensive projection operations. Finally, we conduct an empirical evaluation of our method with the current state of the art and present the results on five datasets from varying domains.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125801196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Semi-Supervised Techniques for Mining Learning Outcomes and Prerequisites 半监督技术的挖掘学习成果和先决条件
I. Labutov, Yun Huang, Peter Brusilovsky, Daqing He
Educational content of today no longer only resides in textbooks and classrooms; more and more learning material is found in a free, accessible form on the Internet. Our long-standing vision is to transform this web of educational content into an adaptive, web-scale "textbook", that can guide its readers to most relevant "pages" according to their learning goal and current knowledge. In this paper, we address one core, long-standing problem towards this goal: identifying outcome and prerequisite concepts within a piece of educational content (e.g., a tutorial). Specifically, we propose a novel approach that leverages textbooks as a source of distant supervision, but learns a model that can generalize to arbitrary documents (such as those on the web). As such, our model can take advantage of any existing textbook, without requiring expert annotation. At the task of predicting outcome and prerequisite concepts, we demonstrate improvements over a number of baselines on six textbooks, especially in the regime of little to no ground-truth labels available. Finally, we demonstrate the utility of a model learned using our approach at the task of identifying prerequisite documents for adaptive content recommendation --- an important step towards our vision of the "web as a textbook".
今天的教育内容不再只存在于教科书和教室;越来越多的学习材料以免费、可访问的形式出现在互联网上。我们长期以来的愿景是将这一教育内容网络转变为一种可适应的、网络规模的“教科书”,可以根据读者的学习目标和当前知识,引导他们进入最相关的“页面”。在本文中,我们解决了一个长期存在的核心问题:在一段教育内容(例如,教程)中确定结果和先决概念。具体来说,我们提出了一种新颖的方法,利用教科书作为远程监督的来源,但学习了一种可以推广到任意文档(如网络上的文档)的模型。因此,我们的模型可以利用任何现有的教科书,而不需要专家注释。在预测结果和前提概念的任务中,我们在六本教科书的一些基线上展示了改进,特别是在很少或没有基本事实标签的情况下。最后,我们展示了使用我们的方法在识别自适应内容推荐的先决条件文档的任务中学习到的模型的实用性——这是我们实现“网络作为教科书”愿景的重要一步。
{"title":"Semi-Supervised Techniques for Mining Learning Outcomes and Prerequisites","authors":"I. Labutov, Yun Huang, Peter Brusilovsky, Daqing He","doi":"10.1145/3097983.3098187","DOIUrl":"https://doi.org/10.1145/3097983.3098187","url":null,"abstract":"Educational content of today no longer only resides in textbooks and classrooms; more and more learning material is found in a free, accessible form on the Internet. Our long-standing vision is to transform this web of educational content into an adaptive, web-scale \"textbook\", that can guide its readers to most relevant \"pages\" according to their learning goal and current knowledge. In this paper, we address one core, long-standing problem towards this goal: identifying outcome and prerequisite concepts within a piece of educational content (e.g., a tutorial). Specifically, we propose a novel approach that leverages textbooks as a source of distant supervision, but learns a model that can generalize to arbitrary documents (such as those on the web). As such, our model can take advantage of any existing textbook, without requiring expert annotation. At the task of predicting outcome and prerequisite concepts, we demonstrate improvements over a number of baselines on six textbooks, especially in the regime of little to no ground-truth labels available. Finally, we demonstrate the utility of a model learned using our approach at the task of identifying prerequisite documents for adaptive content recommendation --- an important step towards our vision of the \"web as a textbook\".","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128380438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Ad Serving with Multiple KPIs 具有多个kpi的广告服务
B. Kitts, M. Krishnan, I. Yadav, Yongbo Zeng, Garrett Badeau, Andrew Potter, Sergey Tolkachov, Ethan Thornburg, Satyanarayana Reddy Janga
Ad-servers have to satisfy many different targeting criteria, and the combination can often result in no feasible solution. We hypothesize that advertisers may be defining these metrics to create a kind of "proxy target". We therefore reformulate the standard ad-serving problem to one where we attempt to get as close as possible to the advertiser's multi-dimensional target inclusive of delivery. We use a simple simulation to illustrate the behavior of this algorithm compared to Constraint and Pacing strategies. The system is then deployed in one of the largest video ad-servers in the United States and we show experimental results from live test ads, as well as 6 months of production performance across hundreds of ads. We find that the live ad-server tests match the simulation, and we report significant gains in multi-KPI performance from using the error minimization strategy.
广告服务器必须满足许多不同的目标定位标准,而这些标准的组合往往会导致没有可行的解决方案。我们假设,广告商可能正在定义这些指标,以创建一种“代理目标”。因此,我们重新制定了标准的广告服务问题,我们试图尽可能接近广告商的多维目标,包括交付。我们使用一个简单的模拟来说明与约束和步调策略相比,该算法的行为。然后将该系统部署在美国最大的视频广告服务器之一中,我们展示了实时测试广告的实验结果,以及6个月的数百个广告的生产性能。我们发现实时广告服务器测试与模拟相匹配,并且我们报告了使用误差最小化策略在多kpi性能方面的显着收益。
{"title":"Ad Serving with Multiple KPIs","authors":"B. Kitts, M. Krishnan, I. Yadav, Yongbo Zeng, Garrett Badeau, Andrew Potter, Sergey Tolkachov, Ethan Thornburg, Satyanarayana Reddy Janga","doi":"10.1145/3097983.3098085","DOIUrl":"https://doi.org/10.1145/3097983.3098085","url":null,"abstract":"Ad-servers have to satisfy many different targeting criteria, and the combination can often result in no feasible solution. We hypothesize that advertisers may be defining these metrics to create a kind of \"proxy target\". We therefore reformulate the standard ad-serving problem to one where we attempt to get as close as possible to the advertiser's multi-dimensional target inclusive of delivery. We use a simple simulation to illustrate the behavior of this algorithm compared to Constraint and Pacing strategies. The system is then deployed in one of the largest video ad-servers in the United States and we show experimental results from live test ads, as well as 6 months of production performance across hundreds of ads. We find that the live ad-server tests match the simulation, and we report significant gains in multi-KPI performance from using the error minimization strategy.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123275012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
BDT: Gradient Boosted Decision Tables for High Accuracy and Scoring Efficiency BDT:用于高精度和评分效率的梯度增强决策表
Yin Lou, M. Obukhov
In this paper we present gradient boosted decision tables (BDTs). A d-dimensional decision table is essentially a mapping from a sequence of d boolean tests to a real value in {R}. We propose novel algorithms to fit decision tables. Our thorough empirical study suggests that decision tables are better weak learners in the gradient boosting framework and can improve the accuracy of the boosted ensemble. In addition, we develop an efficient data structure to represent decision tables and propose a novel fast algorithm to improve the scoring efficiency for boosted ensemble of decision tables. Experiments on public classification and regression datasets demonstrate that our method is able to achieve 1.5x to 6x speedups over the boosted regression trees baseline. We complement our experimental evaluation with a bias-variance analysis that explains how different weak models influence the predictive power of the boosted ensemble. Our experiments suggest gradient boosting with randomly backfitted decision tables distinguishes itself as the most accurate method on a number of classification and regression problems. We have deployed a BDT model to LinkedIn news feed system and achieved significant lift on key metrics.
本文提出了梯度增强决策表(bdt)。d维决策表本质上是从d个布尔测试序列到{R}中的实值的映射。我们提出了新的算法来拟合决策表。我们的实证研究表明,决策表在梯度增强框架中是更好的弱学习器,可以提高增强集合的准确性。此外,我们开发了一种高效的数据结构来表示决策表,并提出了一种新的快速算法来提高决策表增强集合的评分效率。在公共分类和回归数据集上的实验表明,我们的方法能够在增强的回归树基线上实现1.5到6倍的加速。我们用偏方差分析来补充我们的实验评估,该分析解释了不同的弱模型如何影响增强集合的预测能力。我们的实验表明,随机反向拟合决策表的梯度增强在许多分类和回归问题上是最准确的方法。我们已经在LinkedIn新闻推送系统中部署了BDT模型,并在关键指标上取得了显著提升。
{"title":"BDT: Gradient Boosted Decision Tables for High Accuracy and Scoring Efficiency","authors":"Yin Lou, M. Obukhov","doi":"10.1145/3097983.3098175","DOIUrl":"https://doi.org/10.1145/3097983.3098175","url":null,"abstract":"In this paper we present gradient boosted decision tables (BDTs). A d-dimensional decision table is essentially a mapping from a sequence of d boolean tests to a real value in {R}. We propose novel algorithms to fit decision tables. Our thorough empirical study suggests that decision tables are better weak learners in the gradient boosting framework and can improve the accuracy of the boosted ensemble. In addition, we develop an efficient data structure to represent decision tables and propose a novel fast algorithm to improve the scoring efficiency for boosted ensemble of decision tables. Experiments on public classification and regression datasets demonstrate that our method is able to achieve 1.5x to 6x speedups over the boosted regression trees baseline. We complement our experimental evaluation with a bias-variance analysis that explains how different weak models influence the predictive power of the boosted ensemble. Our experiments suggest gradient boosting with randomly backfitted decision tables distinguishes itself as the most accurate method on a number of classification and regression problems. We have deployed a BDT model to LinkedIn news feed system and achieved significant lift on key metrics.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114912860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Detecting Network Effects: Randomizing Over Randomized Experiments 检测网络效应:随机化优于随机化实验
Martin Saveski, Jean Pouget-Abadie, Guillaume Saint-Jacques, Weitao Duan, Souvik Ghosh, Ya Xu, E. Airoldi
Randomized experiments, or A/B tests, are the standard approach for evaluating the causal effects of new product features, i.e., treatments. The validity of these tests rests on the "stable unit treatment value assumption" (SUTVA), which implies that the treatment only affects the behavior of treated users, and does not affect the behavior of their connections. Violations of SUTVA, common in features that exhibit network effects, result in inaccurate estimates of the causal effect of treatment. In this paper, we leverage a new experimental design for testing whether SUTVA holds, without making any assumptions on how treatment effects may spill over between the treatment and the control group. To achieve this, we simultaneously run both a completely randomized and a cluster-based randomized experiment, and then we compare the difference of the resulting estimates. We present a statistical test for measuring the significance of this difference and offer theoretical bounds on the Type I error rate. We provide practical guidelines for implementing our methodology on large-scale experimentation platforms. Importantly, the proposed methodology can be applied to settings in which a network is not necessarily observed but, if available, can be used in the analysis. Finally, we deploy this design to LinkedIn's experimentation platform and apply it to two online experiments, highlighting the presence of network effects and bias in standard A/B testing approaches in a real-world setting.
随机实验或A/B测试是评估新产品特性(即治疗方法)的因果效应的标准方法。这些测试的有效性取决于“稳定单位处理值假设”(SUTVA),这意味着处理只影响被处理用户的行为,而不影响其连接的行为。违反SUTVA在表现出网络效应的特征中很常见,导致对治疗因果效应的估计不准确。在本文中,我们利用一种新的实验设计来测试SUTVA是否成立,而没有对治疗效果如何在治疗组和对照组之间溢出做出任何假设。为了达到这个目的,我们同时运行一个完全随机和一个基于集群的随机实验,然后我们比较结果估计的差异。我们提出了一个统计检验来衡量这种差异的显著性,并提供了第一类错误率的理论界限。我们为在大规模实验平台上实施我们的方法提供了实用指南。重要的是,所提出的方法可以应用于不一定观察到网络的设置,但如果可用,可以用于分析。最后,我们将此设计部署到LinkedIn的实验平台上,并将其应用于两个在线实验,突出了现实环境中标准A/B测试方法中网络效应和偏见的存在。
{"title":"Detecting Network Effects: Randomizing Over Randomized Experiments","authors":"Martin Saveski, Jean Pouget-Abadie, Guillaume Saint-Jacques, Weitao Duan, Souvik Ghosh, Ya Xu, E. Airoldi","doi":"10.1145/3097983.3098192","DOIUrl":"https://doi.org/10.1145/3097983.3098192","url":null,"abstract":"Randomized experiments, or A/B tests, are the standard approach for evaluating the causal effects of new product features, i.e., treatments. The validity of these tests rests on the \"stable unit treatment value assumption\" (SUTVA), which implies that the treatment only affects the behavior of treated users, and does not affect the behavior of their connections. Violations of SUTVA, common in features that exhibit network effects, result in inaccurate estimates of the causal effect of treatment. In this paper, we leverage a new experimental design for testing whether SUTVA holds, without making any assumptions on how treatment effects may spill over between the treatment and the control group. To achieve this, we simultaneously run both a completely randomized and a cluster-based randomized experiment, and then we compare the difference of the resulting estimates. We present a statistical test for measuring the significance of this difference and offer theoretical bounds on the Type I error rate. We provide practical guidelines for implementing our methodology on large-scale experimentation platforms. Importantly, the proposed methodology can be applied to settings in which a network is not necessarily observed but, if available, can be used in the analysis. Finally, we deploy this design to LinkedIn's experimentation platform and apply it to two online experiments, highlighting the presence of network effects and bias in standard A/B testing approaches in a real-world setting.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124318782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
GELL: Automatic Extraction of Epidemiological Line Lists from Open Sources GELL:从开放资源中自动提取流行病学线列表
Saurav Ghosh, Prithwish Chakraborty, B. Lewis, M. Majumder, E. Cohn, J. Brownstein, M. Marathe, Naren Ramakrishnan
Real-time monitoring and responses to emerging public health threats rely on the availability of timely surveillance data. During the early stages of an epidemic, the ready availability of line lists with detailed tabular information about laboratory-confirmed cases can assist epidemiologists in making reliable inferences and forecasts. Such inferences are crucial to understand the epidemiology of a specific disease early enough to stop or control the outbreak. However, construction of such line lists requires considerable human supervision and therefore, difficult to generate in real-time. In this paper, we motivate Guided Epidemiological Line List (GELL), the first tool for building automated line lists (in near real-time) from open source reports of emerging disease outbreaks. Specifically, we focus on deriving epidemiological characteristics of an emerging disease and the affected population from reports of illness. GELL uses distributed vector representations (ala word2vec) to discover a set of indicators for each line list feature. This discovery of indicators is followed by the use of dependency parsing based techniques for final extraction in tabular form. We evaluate the performance of GELL against a human annotated line list provided by HealthMap corresponding to MERS outbreaks in Saudi Arabia. We demonstrate that GELL extracts line list features with increased accuracy compared to a baseline method. We further show how these automatically extracted line list features can be used for making epidemiological inferences, such as inferring demographics and symptoms-to-hospitalization period of affected individuals.
实时监测和应对新出现的公共卫生威胁取决于能否获得及时的监测数据。在流行病的早期阶段,现成的带有实验室确诊病例详细表格信息的清单可帮助流行病学家作出可靠的推断和预测。这样的推断对于及早了解特定疾病的流行病学以阻止或控制疫情至关重要。然而,这种行列表的构建需要大量的人工监督,因此难以实时生成。在本文中,我们激发了引导流行病学线列表(GELL),这是第一个从新出现的疾病暴发的开源报告中构建自动化线列表(近乎实时)的工具。具体而言,我们侧重于从疾病报告中得出新出现疾病和受影响人群的流行病学特征。GELL使用分布式向量表示(类似于word2vec)为每个行列表特征发现一组指标。在发现指示器之后,使用基于依赖项解析的技术以表格形式进行最终提取。我们根据HealthMap提供的与沙特阿拉伯中东呼吸综合征爆发相对应的人类注释线列表评估了GELL的性能。我们证明,与基线方法相比,GELL提取行列表特征的准确性更高。我们进一步展示了如何使用这些自动提取的线列表特征进行流行病学推断,例如推断受影响个体的人口统计学和症状到住院时间。
{"title":"GELL: Automatic Extraction of Epidemiological Line Lists from Open Sources","authors":"Saurav Ghosh, Prithwish Chakraborty, B. Lewis, M. Majumder, E. Cohn, J. Brownstein, M. Marathe, Naren Ramakrishnan","doi":"10.1145/3097983.3098073","DOIUrl":"https://doi.org/10.1145/3097983.3098073","url":null,"abstract":"Real-time monitoring and responses to emerging public health threats rely on the availability of timely surveillance data. During the early stages of an epidemic, the ready availability of line lists with detailed tabular information about laboratory-confirmed cases can assist epidemiologists in making reliable inferences and forecasts. Such inferences are crucial to understand the epidemiology of a specific disease early enough to stop or control the outbreak. However, construction of such line lists requires considerable human supervision and therefore, difficult to generate in real-time. In this paper, we motivate Guided Epidemiological Line List (GELL), the first tool for building automated line lists (in near real-time) from open source reports of emerging disease outbreaks. Specifically, we focus on deriving epidemiological characteristics of an emerging disease and the affected population from reports of illness. GELL uses distributed vector representations (ala word2vec) to discover a set of indicators for each line list feature. This discovery of indicators is followed by the use of dependency parsing based techniques for final extraction in tabular form. We evaluate the performance of GELL against a human annotated line list provided by HealthMap corresponding to MERS outbreaks in Saudi Arabia. We demonstrate that GELL extracts line list features with increased accuracy compared to a baseline method. We further show how these automatically extracted line list features can be used for making epidemiological inferences, such as inferring demographics and symptoms-to-hospitalization period of affected individuals.","PeriodicalId":314049,"journal":{"name":"Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116838385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1