首页 > 最新文献

Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval最新文献

英文 中文
Graph-based text classification: learn from your neighbors 基于图的文本分类:向你的邻居学习
Ralitsa Angelova, G. Weikum
Automatic classification of data items, based on training samples, can be boosted by considering the neighborhood of data items in a graph structure (e.g., neighboring documents in a hyperlink environment or co-authors and their publications for bibliographic data entries). This paper presents a new method for graph-based classification, with particular emphasis on hyperlinked text documents but broader applicability. Our approach is based on iterative relaxation labeling and can be combined with either Bayesian or SVM classifiers on the feature spaces of the given data items. The graph neighborhood is taken into consideration to exploit locality patterns while at the same time avoiding overfitting. In contrast to prior work along these lines, our approach employs a number of novel techniques: dynamically inferring the link/class pattern in the graph in the run of the iterative relaxation labeling, judicious pruning of edges from the neighborhood graph based on node dissimilarities and node degrees, weighting the influence of edges based on a distance metric between the classification labels of interest and weighting edges by content similarity measures. Our techniques considerably improve the robustness and accuracy of the classification outcome, as shown in systematic experimental comparisons with previously published methods on three different real-world datasets.
基于训练样本的数据项自动分类可以通过考虑图结构中数据项的邻域来提高(例如,超链接环境中的相邻文档或书目数据条目的合著者及其出版物)。本文提出了一种新的基于图的分类方法,特别强调了超链接文本文档,但具有更广泛的适用性。我们的方法基于迭代松弛标记,可以与给定数据项的特征空间上的贝叶斯或支持向量机分类器相结合。在利用局部性模式的同时,考虑了图邻域,避免了过拟合。与之前沿着这些方向的工作相反,我们的方法采用了许多新技术:在迭代松弛标记的运行中动态推断图中的链接/类模式,基于节点不相似度和节点度从邻域图中明智地修剪边缘,基于感兴趣的分类标签之间的距离度量加权边缘的影响,以及通过内容相似性度量加权边缘。我们的技术大大提高了分类结果的稳健性和准确性,与之前发表的方法在三个不同的现实世界数据集上的系统实验比较表明。
{"title":"Graph-based text classification: learn from your neighbors","authors":"Ralitsa Angelova, G. Weikum","doi":"10.1145/1148170.1148254","DOIUrl":"https://doi.org/10.1145/1148170.1148254","url":null,"abstract":"Automatic classification of data items, based on training samples, can be boosted by considering the neighborhood of data items in a graph structure (e.g., neighboring documents in a hyperlink environment or co-authors and their publications for bibliographic data entries). This paper presents a new method for graph-based classification, with particular emphasis on hyperlinked text documents but broader applicability. Our approach is based on iterative relaxation labeling and can be combined with either Bayesian or SVM classifiers on the feature spaces of the given data items. The graph neighborhood is taken into consideration to exploit locality patterns while at the same time avoiding overfitting. In contrast to prior work along these lines, our approach employs a number of novel techniques: dynamically inferring the link/class pattern in the graph in the run of the iterative relaxation labeling, judicious pruning of edges from the neighborhood graph based on node dissimilarities and node degrees, weighting the influence of edges based on a distance metric between the classification labels of interest and weighting edges by content similarity measures. Our techniques considerably improve the robustness and accuracy of the classification outcome, as shown in systematic experimental comparisons with previously published methods on three different real-world datasets.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"181 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131883780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 188
Load balancing for term-distributed parallel retrieval 项分布并行检索的负载平衡
Alistair Moffat, William Webber, J. Zobel
Large-scale web and text retrieval systems deal with amounts of data that greatly exceed the capacity of any single machine. To handle the necessary data volumes and query throughput rates, parallel systems are used, in which the document and index data are split across tightly-clustered distributed computing systems. The index data can be distributed either by document or by term. In this paper we examine methods for load balancing in term-distributed parallel architectures, and propose a suite of techniques for reducing net querying costs. In combination, the techniques we describe allow a 30% improvement in query throughput when tested on an eight-node parallel computer system.
大规模的网络和文本检索系统处理的数据量远远超过任何一台机器的能力。为了处理必要的数据量和查询吞吐率,需要使用并行系统,其中文档和索引数据在紧密集群的分布式计算系统中分开。索引数据既可以按文档分发,也可以按术语分发。在本文中,我们研究了术语分布式并行架构中的负载平衡方法,并提出了一套降低净查询成本的技术。在一个八节点并行计算机系统上进行测试时,我们所描述的技术可以将查询吞吐量提高30%。
{"title":"Load balancing for term-distributed parallel retrieval","authors":"Alistair Moffat, William Webber, J. Zobel","doi":"10.1145/1148170.1148232","DOIUrl":"https://doi.org/10.1145/1148170.1148232","url":null,"abstract":"Large-scale web and text retrieval systems deal with amounts of data that greatly exceed the capacity of any single machine. To handle the necessary data volumes and query throughput rates, parallel systems are used, in which the document and index data are split across tightly-clustered distributed computing systems. The index data can be distributed either by document or by term. In this paper we examine methods for load balancing in term-distributed parallel architectures, and propose a suite of techniques for reducing net querying costs. In combination, the techniques we describe allow a 30% improvement in query throughput when tested on an eight-node parallel computer system.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132239434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 98
Learning to advertise 学习做广告
A. Lacerda, Marco Cristo, Marcos André Gonçalves, Weiguo Fan, N. Ziviani, B. Ribeiro-Neto
Content-targeted advertising, the task of automatically associating ads to a Web page, constitutes a key Web monetization strategy nowadays. Further, it introduces new challenging technical problems and raises interesting questions. For instance, how to design ranking functions able to satisfy conflicting goals such as selecting advertisements (ads) that are relevant to the users and suitable and profitable to the publishers and advertisers? In this paper we propose a new framework for associating ads with web pages based on Genetic Programming (GP). Our GP method aims at learning functions that select the most appropriate ads, given the contents of a Web page. These ranking functions are designed to optimize overall precision and minimize the number of misplacements. By using a real ad collection and web pages from a newspaper, we obtained a gain over a state-of-the-art baseline method of 61.7% in average precision. Further, by evolving individuals to provide good ranking estimations, GP was able to discover ranking functions that are very effective in placing ads in web pages while avoiding irrelevant ones.
内容定向广告,即自动将广告与网页关联起来的任务,构成了当今关键的网络货币化策略。此外,它还引入了新的具有挑战性的技术问题,并提出了有趣的问题。例如,如何设计排名功能,以满足相互冲突的目标,如选择与用户相关的广告(广告),并对出版商和广告商来说是合适的和有利可图的?本文提出了一种基于遗传规划的广告与网页关联框架。我们的GP方法旨在学习在给定网页内容的情况下选择最合适广告的函数。这些排序功能的设计是为了优化整体精度,并尽量减少错位的数量。通过使用真实的广告集和报纸上的网页,我们获得了比最先进的基线方法平均精度61.7%的增益。此外,通过进化个体来提供良好的排名估计,GP能够发现在网页上放置广告时非常有效的排名功能,同时避免不相关的广告。
{"title":"Learning to advertise","authors":"A. Lacerda, Marco Cristo, Marcos André Gonçalves, Weiguo Fan, N. Ziviani, B. Ribeiro-Neto","doi":"10.1145/1148170.1148265","DOIUrl":"https://doi.org/10.1145/1148170.1148265","url":null,"abstract":"Content-targeted advertising, the task of automatically associating ads to a Web page, constitutes a key Web monetization strategy nowadays. Further, it introduces new challenging technical problems and raises interesting questions. For instance, how to design ranking functions able to satisfy conflicting goals such as selecting advertisements (ads) that are relevant to the users and suitable and profitable to the publishers and advertisers? In this paper we propose a new framework for associating ads with web pages based on Genetic Programming (GP). Our GP method aims at learning functions that select the most appropriate ads, given the contents of a Web page. These ranking functions are designed to optimize overall precision and minimize the number of misplacements. By using a real ad collection and web pages from a newspaper, we obtained a gain over a state-of-the-art baseline method of 61.7% in average precision. Further, by evolving individuals to provide good ranking estimations, GP was able to discover ranking functions that are very effective in placing ads in web pages while avoiding irrelevant ones.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"223 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134455952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 167
A method of rating the credibility of news documents on the web 一种评价网络上新闻文件可信度的方法
Ryosuke Nagura, Yohei Seki, N. Kando, Masaki Aono
We propose a method to rate the credibility of news articles using three clues: (1) commonality of the contents of articles among different news publishers; (2) numerical agreement versus contradiction of numerical values reported in the articles; and (3) objectivity based on subjective speculative phrases and news sources. We tested this method on news stories taken from seven different news sites on the Web. The average agreement between the system-produced "credibility" and the manual judgments of three human assessors on the 52 sample articles was 69.1%. The limitations of the current approach and future directions are discussed.
我们提出了一种利用三个线索来评价新闻文章可信度的方法:(1)不同新闻发布者文章内容的共性;(二)文章中数值的一致与矛盾;(3)基于主观猜测短语和新闻来源的客观性。我们在七个不同的新闻网站上对这种方法进行了测试。系统产生的“可信度”与三名人工评估员对52篇样本文章的人工判断之间的平均一致性为69.1%。讨论了当前方法的局限性和未来的发展方向。
{"title":"A method of rating the credibility of news documents on the web","authors":"Ryosuke Nagura, Yohei Seki, N. Kando, Masaki Aono","doi":"10.1145/1148170.1148316","DOIUrl":"https://doi.org/10.1145/1148170.1148316","url":null,"abstract":"We propose a method to rate the credibility of news articles using three clues: (1) commonality of the contents of articles among different news publishers; (2) numerical agreement versus contradiction of numerical values reported in the articles; and (3) objectivity based on subjective speculative phrases and news sources. We tested this method on news stories taken from seven different news sites on the Web. The average agreement between the system-produced \"credibility\" and the manual judgments of three human assessors on the 52 sample articles was 69.1%. The limitations of the current approach and future directions are discussed.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115031089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
One-sided measures for evaluating ranked retrieval effectiveness with spontaneous conversational speech 评价自发性会话语音排序检索效果的单侧测量方法
Baolong Liu, Douglas W. Oard
Early speech retrieval experiments focused on news broadcasts, for which adequate Automatic Speech Recognition (ASR) accuracy could be obtained. Like newspapers, news broadcasts are a manually selected and arranged set of stories. Evaluation designs reflected that, using known story boundaries as a basis for evaluation. Substantial advances in ASR accuracy now make it possible to build search systems for some types of spontaneous conversational speech, but present evaluation designs continue to rely on known topic boundaries that are no longer well matched to the nature of the materials. We propose a new class of measures for speech retrieval based on manual annotation of points at which a user with specific topical interests would wish replay to begin.
早期的语音检索实验主要集中在新闻广播中,可以获得足够的自动语音识别(ASR)精度。像报纸一样,新闻广播是人工选择和安排的一组故事。评估设计反映了这一点,使用已知的故事边界作为评估的基础。在ASR准确度方面取得的实质性进展现在使得为某些类型的自发会话语音建立搜索系统成为可能,但是目前的评估设计仍然依赖于已知的主题边界,这些主题边界不再与材料的性质很好地匹配。我们提出了一种新的语音检索方法,该方法基于对具有特定主题兴趣的用户希望重播开始的点进行手动注释。
{"title":"One-sided measures for evaluating ranked retrieval effectiveness with spontaneous conversational speech","authors":"Baolong Liu, Douglas W. Oard","doi":"10.1145/1148170.1148311","DOIUrl":"https://doi.org/10.1145/1148170.1148311","url":null,"abstract":"Early speech retrieval experiments focused on news broadcasts, for which adequate Automatic Speech Recognition (ASR) accuracy could be obtained. Like newspapers, news broadcasts are a manually selected and arranged set of stories. Evaluation designs reflected that, using known story boundaries as a basis for evaluation. Substantial advances in ASR accuracy now make it possible to build search systems for some types of spontaneous conversational speech, but present evaluation designs continue to rely on known topic boundaries that are no longer well matched to the nature of the materials. We propose a new class of measures for speech retrieval based on manual annotation of points at which a user with specific topical interests would wish replay to begin.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115513761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Unity: relevance feedback using user query logs 统一:通过用户查询日志进行相关性反馈
J. Parikh, S. Kapur
The exponential growth of the Web and the increasing ability of web search engines to index data have led to a problem of plenty. The number of results returned per query is typically in the order of millions of documents for many common queries. Although there is the benefit of added coverage for every query, the problem of ranking these documents and giving the best results gets worse. The problem is even more difficult in case of temporal and ambiguous queries. We try to address this problem using feedback from user query logs. We leverage a technology called Units for generating query refinements which are shown as Also try queries on Yahoo! Search. We consider these refinements as sub-concepts which help define user intent and use them to improve search relevance. The results obtained via live testing on Yahoo! Search are encouraging.
网络的指数级增长和网络搜索引擎索引数据的能力不断增强,导致了数据过剩的问题。对于许多常见查询,每个查询返回的结果数量通常为数百万个文档。尽管为每个查询增加覆盖范围是有好处的,但是对这些文档进行排序并给出最佳结果的问题变得更糟了。在时态查询和模糊查询的情况下,这个问题甚至更加困难。我们尝试使用用户查询日志的反馈来解决这个问题。我们利用一种称为Units的技术来生成查询改进,如下所示:Also try queries on Yahoo!搜索。我们将这些改进视为有助于定义用户意图并使用它们来提高搜索相关性的子概念。通过在Yahoo!搜索结果令人鼓舞。
{"title":"Unity: relevance feedback using user query logs","authors":"J. Parikh, S. Kapur","doi":"10.1145/1148170.1148319","DOIUrl":"https://doi.org/10.1145/1148170.1148319","url":null,"abstract":"The exponential growth of the Web and the increasing ability of web search engines to index data have led to a problem of plenty. The number of results returned per query is typically in the order of millions of documents for many common queries. Although there is the benefit of added coverage for every query, the problem of ranking these documents and giving the best results gets worse. The problem is even more difficult in case of temporal and ambiguous queries. We try to address this problem using feedback from user query logs. We leverage a technology called Units for generating query refinements which are shown as Also try queries on Yahoo! Search. We consider these refinements as sub-concepts which help define user intent and use them to improve search relevance. The results obtained via live testing on Yahoo! Search are encouraging.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124490481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Concept-based biomedical text retrieval 基于概念的生物医学文本检索
Ming Zhong, Xiangji Huang
One challenging problem for biomedical text retrieval is to find accurate synonyms or name variants for biomedical entities. In this paper, we propose a new concept-based approach to tackle this problem. In this approach, a set of concepts instead of keywords will be extracted from a query first. Then these concepts will be used for retrieval purpose. The experiment results show that the proposed approach can boost the retrieval performance and it generates very good results on 2005 TREC Genomics data sets.
生物医学文本检索中一个具有挑战性的问题是如何找到准确的生物医学实体的同义词或名称变体。在本文中,我们提出了一种新的基于概念的方法来解决这个问题。在这种方法中,首先从查询中提取一组概念而不是关键字。然后这些概念将用于检索目的。实验结果表明,该方法可以提高检索性能,并在2005年TREC Genomics数据集上取得了很好的检索结果。
{"title":"Concept-based biomedical text retrieval","authors":"Ming Zhong, Xiangji Huang","doi":"10.1145/1148170.1148336","DOIUrl":"https://doi.org/10.1145/1148170.1148336","url":null,"abstract":"One challenging problem for biomedical text retrieval is to find accurate synonyms or name variants for biomedical entities. In this paper, we propose a new concept-based approach to tackle this problem. In this approach, a set of concepts instead of keywords will be extracted from a query first. Then these concepts will be used for retrieval purpose. The experiment results show that the proposed approach can boost the retrieval performance and it generates very good results on 2005 TREC Genomics data sets.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123930759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
An analysis of the coupling between training set and neighborhood sizes for the kNN classifier kNN分类器训练集与邻域大小耦合分析
J. S. Olsson
We consider the relationship between training set size and the parameter k for the k-Nearest Neighbors (kNN) classifier. When few examples are available, we observe that accuracy is sensitive to k and that best k tends to increase with training size. We explore the subsequent risk that k tuned on partitions will be suboptimal after aggregation and re-training. This risk is found to be most severe when little data is available. For larger training sizes, accuracy becomes increasingly stable with respect to k and the risk decreases.
我们考虑k-最近邻(kNN)分类器的训练集大小和参数k之间的关系。当可用的示例很少时,我们观察到准确率对k很敏感,并且最佳k倾向于随着训练规模的增加而增加。我们探讨了在聚合和重新训练之后,在分区上调优的k将是次优的风险。当可用数据很少时,发现这种风险最为严重。对于较大的训练规模,准确率相对于k变得越来越稳定,风险降低。
{"title":"An analysis of the coupling between training set and neighborhood sizes for the kNN classifier","authors":"J. S. Olsson","doi":"10.1145/1148170.1148317","DOIUrl":"https://doi.org/10.1145/1148170.1148317","url":null,"abstract":"We consider the relationship between training set size and the parameter k for the k-Nearest Neighbors (kNN) classifier. When few examples are available, we observe that accuracy is sensitive to k and that best k tends to increase with training size. We explore the subsequent risk that k tuned on partitions will be suboptimal after aggregation and re-training. This risk is found to be most severe when little data is available. For larger training sizes, accuracy becomes increasingly stable with respect to k and the risk decreases.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125112575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Information retrieval at Boeing: plans and successes 波音公司的信息检索:计划和成功
R. Radhakrishnan
Background Many information technology products in the marketplace are designed as “Enterprise” solutions and systems. Boeing is an enterprise composed of multiple enterprises: Boeing Commercial Airplanes, Integrated Defense Systems, Boeing Capital Corporation and Connexion by Boeing. In the necessary globalization of the 21 Century, each of these major business units are really extended enterprises due to the partnerships and business arrangements we have made, involving worldwide engineering design and manufacturing companies, suppliers and subcontractors, air carriers and leasing operators, and military & government agencies. It is difficult to scale many IT design approaches and products to the Boeing operating environment.
市场上的许多信息技术产品都被设计为“企业”解决方案和系统。波音公司是由波音商用飞机公司、波音综合防御系统公司、波音资本公司和波音公司旗下的Connexion等多家企业组成的企业。在21世纪必要的全球化中,由于我们的合作伙伴关系和业务安排,这些主要业务单位中的每一个都是真正的扩展企业,涉及全球工程设计和制造公司,供应商和分包商,航空公司和租赁运营商,以及军事和政府机构。很难将许多It设计方法和产品扩展到波音的操作环境中。
{"title":"Information retrieval at Boeing: plans and successes","authors":"R. Radhakrishnan","doi":"10.1145/1148170.1148173","DOIUrl":"https://doi.org/10.1145/1148170.1148173","url":null,"abstract":"Background Many information technology products in the marketplace are designed as “Enterprise” solutions and systems. Boeing is an enterprise composed of multiple enterprises: Boeing Commercial Airplanes, Integrated Defense Systems, Boeing Capital Corporation and Connexion by Boeing. In the necessary globalization of the 21 Century, each of these major business units are really extended enterprises due to the partnerships and business arrangements we have made, involving worldwide engineering design and manufacturing companies, suppliers and subcontractors, air carriers and leasing operators, and military & government agencies. It is difficult to scale many IT design approaches and products to the Boeing operating environment.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126275028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Evaluating evaluation metrics based on the bootstrap 评估基于自举的评估指标
T. Sakai
This paper describes how the Bootstrap approach to statistics can be applied to the evaluation of IR effectiveness metrics. First, we argue that Bootstrap Hypothesis Tests deserve more attention from the IR community, as they are based on fewer assumptions than traditional statistical significance tests. We then describe straightforward methods for comparing the sensitivity of IR metrics based on Bootstrap Hypothesis Tests. Unlike the heuristics-based "swap" method proposed by Voorhees and Buckley, our method estimates the performance difference required to achieve a given significance level directly from Bootstrap Hypothesis Test results. In addition, we describe a simple way of examining the accuracy of rank correlation between two metrics based on the Bootstrap Estimate of Standard Error. We demonstrate the usefulness of our methods using test collections and runs from the NTCIR CLIR track for comparing seven IR metrics, including those that can handle graded relevance and those based on the Geometric Mean.
本文描述了如何将统计方法应用于IR有效性指标的评估。首先,我们认为自举假设检验值得IR社区更多的关注,因为它们比传统的统计显著性检验基于更少的假设。然后,我们描述了基于Bootstrap假设检验比较IR指标灵敏度的直接方法。与Voorhees和Buckley提出的基于启发式的“交换”方法不同,我们的方法直接从Bootstrap假设检验结果中估计达到给定显著性水平所需的性能差异。此外,我们描述了一种简单的方法来检查两个指标之间的等级相关的准确性基于标准误差的Bootstrap估计。我们使用测试集合和NTCIR CLIR轨道上的运行来比较七个IR指标,包括那些可以处理分级相关性的指标和那些基于几何平均值的指标,从而证明了我们的方法的实用性。
{"title":"Evaluating evaluation metrics based on the bootstrap","authors":"T. Sakai","doi":"10.1145/1148170.1148261","DOIUrl":"https://doi.org/10.1145/1148170.1148261","url":null,"abstract":"This paper describes how the Bootstrap approach to statistics can be applied to the evaluation of IR effectiveness metrics. First, we argue that Bootstrap Hypothesis Tests deserve more attention from the IR community, as they are based on fewer assumptions than traditional statistical significance tests. We then describe straightforward methods for comparing the sensitivity of IR metrics based on Bootstrap Hypothesis Tests. Unlike the heuristics-based \"swap\" method proposed by Voorhees and Buckley, our method estimates the performance difference required to achieve a given significance level directly from Bootstrap Hypothesis Test results. In addition, we describe a simple way of examining the accuracy of rank correlation between two metrics based on the Bootstrap Estimate of Standard Error. We demonstrate the usefulness of our methods using test collections and runs from the NTCIR CLIR track for comparing seven IR metrics, including those that can handle graded relevance and those based on the Geometric Mean.","PeriodicalId":433366,"journal":{"name":"Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130230063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 227
期刊
Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1