该理论具有预测性,但它是完整的吗?:在人类随机性感知中的应用

J. Kleinberg, Annie Liang, S. Mullainathan
{"title":"该理论具有预测性,但它是完整的吗?:在人类随机性感知中的应用","authors":"J. Kleinberg, Annie Liang, S. Mullainathan","doi":"10.1145/3033274.3084094","DOIUrl":null,"url":null,"abstract":"When we test a theory using data, it is common to focus on correctness: do the predictions of the theory match what we see in the data? But we also care about completeness: how much of the predictable variation in the data is captured by the theory? This question is difficult to answer, because in general we do not know how much \"predictable variation\" there is in the problem. In this paper, we consider approaches motivated by machine learning algorithms as a means of constructing a benchmark for the best attainable level of prediction. We illustrate our methods on the task of prediction of human-generated random sequences. Relative to an atheoretical machine learning algorithm benchmark, we find that existing behavioral models explain roughly 10 to 12 percent of the predictable variation in this problem. This fraction is robust across several variations on the problem. We also consider a version of this approach for analyzing field data from domains in which human perception and generation of randomness has been used as a conceptual framework; these include sequential decision-making and repeated zero-sum games. In these domains, our framework for testing the completeness of theories suggest that existing theoretical models may be more complete in their predictions for some domains than for others, suggesting that our methods can offer a comparative perspective across settings. Overall, our results indicate that (i) there is a significant amount of structure in this problem that existing models have yet to capture and (ii) there are rich domains in which machine learning may provide a viable approach to testing completeness.","PeriodicalId":287551,"journal":{"name":"Proceedings of the 2017 ACM Conference on Economics and Computation","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"15","resultStr":"{\"title\":\"The Theory is Predictive, but is it Complete?: An Application to Human Perception of Randomness\",\"authors\":\"J. Kleinberg, Annie Liang, S. Mullainathan\",\"doi\":\"10.1145/3033274.3084094\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"When we test a theory using data, it is common to focus on correctness: do the predictions of the theory match what we see in the data? But we also care about completeness: how much of the predictable variation in the data is captured by the theory? This question is difficult to answer, because in general we do not know how much \\\"predictable variation\\\" there is in the problem. In this paper, we consider approaches motivated by machine learning algorithms as a means of constructing a benchmark for the best attainable level of prediction. We illustrate our methods on the task of prediction of human-generated random sequences. Relative to an atheoretical machine learning algorithm benchmark, we find that existing behavioral models explain roughly 10 to 12 percent of the predictable variation in this problem. This fraction is robust across several variations on the problem. We also consider a version of this approach for analyzing field data from domains in which human perception and generation of randomness has been used as a conceptual framework; these include sequential decision-making and repeated zero-sum games. In these domains, our framework for testing the completeness of theories suggest that existing theoretical models may be more complete in their predictions for some domains than for others, suggesting that our methods can offer a comparative perspective across settings. Overall, our results indicate that (i) there is a significant amount of structure in this problem that existing models have yet to capture and (ii) there are rich domains in which machine learning may provide a viable approach to testing completeness.\",\"PeriodicalId\":287551,\"journal\":{\"name\":\"Proceedings of the 2017 ACM Conference on Economics and Computation\",\"volume\":\"17 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-06-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"15\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2017 ACM Conference on Economics and Computation\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3033274.3084094\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2017 ACM Conference on Economics and Computation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3033274.3084094","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 15

摘要

当我们用数据测试一个理论时,通常关注的是正确性:理论的预测与我们在数据中看到的相符吗?但我们也关心完整性:有多少可预测的数据变化被理论捕获?这个问题很难回答,因为一般来说,我们不知道问题中有多少“可预测的变化”。在本文中,我们考虑由机器学习算法驱动的方法,作为构建可达到的最佳预测水平基准的一种手段。我们举例说明了我们的方法对人类产生的随机序列的预测任务。相对于理论机器学习算法基准,我们发现现有的行为模型可以解释这个问题中大约10%到12%的可预测变化。这个分数在问题的几个变体中都是健壮的。我们还考虑了这种方法的一个版本,用于分析来自人类感知和随机性生成已被用作概念框架的领域的现场数据;这包括顺序决策和重复的零和游戏。在这些领域,我们测试理论完整性的框架表明,现有的理论模型在某些领域的预测可能比其他领域更完整,这表明我们的方法可以提供跨设置的比较视角。总的来说,我们的结果表明:(i)在这个问题中存在大量现有模型尚未捕获的结构,(ii)有丰富的领域,机器学习可以提供一种可行的方法来测试完整性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
The Theory is Predictive, but is it Complete?: An Application to Human Perception of Randomness
When we test a theory using data, it is common to focus on correctness: do the predictions of the theory match what we see in the data? But we also care about completeness: how much of the predictable variation in the data is captured by the theory? This question is difficult to answer, because in general we do not know how much "predictable variation" there is in the problem. In this paper, we consider approaches motivated by machine learning algorithms as a means of constructing a benchmark for the best attainable level of prediction. We illustrate our methods on the task of prediction of human-generated random sequences. Relative to an atheoretical machine learning algorithm benchmark, we find that existing behavioral models explain roughly 10 to 12 percent of the predictable variation in this problem. This fraction is robust across several variations on the problem. We also consider a version of this approach for analyzing field data from domains in which human perception and generation of randomness has been used as a conceptual framework; these include sequential decision-making and repeated zero-sum games. In these domains, our framework for testing the completeness of theories suggest that existing theoretical models may be more complete in their predictions for some domains than for others, suggesting that our methods can offer a comparative perspective across settings. Overall, our results indicate that (i) there is a significant amount of structure in this problem that existing models have yet to capture and (ii) there are rich domains in which machine learning may provide a viable approach to testing completeness.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Potential Function Minimizers of Combinatorial Congestion Games: Efficiency and Computation Making Right Decisions Based on Wrong Opinions Learning in the Repeated Secretary Problem Facilitating the Search for Partners on Matching Platforms: Restricting Agent Actions Deferred-Acceptance Auctions for Multiple Levels of Service
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1