General Belief Revision

J. Delgrande, P. Peppas, S. Woltran
{"title":"General Belief Revision","authors":"J. Delgrande, P. Peppas, S. Woltran","doi":"10.1145/3203409","DOIUrl":null,"url":null,"abstract":"In artificial intelligence, a key question concerns how an agent may rationally revise its beliefs in light of new information. The standard (AGM) approach to belief revision assumes that the underlying logic contains classical propositional logic. This is a significant limitation, since many representation schemes in AI don’t subsume propositional logic. In this article, we consider the question of what the minimal requirements are on a logic, such that the AGM approach to revision may be formulated. We show that AGM-style revision can be obtained even when extremely little is assumed of the underlying language and its semantics; in fact, one requires little more than a language with sentences that are satisfied at models, or possible worlds. The classical AGM postulates are expressed in this framework and a representation result is established between the postulate set and certain preorders on possible worlds. To obtain the representation result, we add a new postulate to the AGM postulates, and we add a constraint to preorders on worlds. Crucially, both of these additions are redundant in the original AGM framework, and so we extend, rather than modify, the AGM approach. As well, iterated revision is addressed and the Darwiche/Pearl postulates are shown to be compatible with our approach. Various examples are given to illustrate the approach, including Horn clause revision, revision in extended logic programs, and belief revision in a very basic logic called literal revision.","PeriodicalId":17199,"journal":{"name":"Journal of the ACM (JACM)","volume":"55 1","pages":"1 - 34"},"PeriodicalIF":0.0000,"publicationDate":"2018-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"15","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the ACM (JACM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3203409","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 15

Abstract

In artificial intelligence, a key question concerns how an agent may rationally revise its beliefs in light of new information. The standard (AGM) approach to belief revision assumes that the underlying logic contains classical propositional logic. This is a significant limitation, since many representation schemes in AI don’t subsume propositional logic. In this article, we consider the question of what the minimal requirements are on a logic, such that the AGM approach to revision may be formulated. We show that AGM-style revision can be obtained even when extremely little is assumed of the underlying language and its semantics; in fact, one requires little more than a language with sentences that are satisfied at models, or possible worlds. The classical AGM postulates are expressed in this framework and a representation result is established between the postulate set and certain preorders on possible worlds. To obtain the representation result, we add a new postulate to the AGM postulates, and we add a constraint to preorders on worlds. Crucially, both of these additions are redundant in the original AGM framework, and so we extend, rather than modify, the AGM approach. As well, iterated revision is addressed and the Darwiche/Pearl postulates are shown to be compatible with our approach. Various examples are given to illustrate the approach, including Horn clause revision, revision in extended logic programs, and belief revision in a very basic logic called literal revision.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
一般信念修正
在人工智能领域,一个关键问题是智能体如何根据新信息理性地修正其信念。信念修正的标准(AGM)方法假定底层逻辑包含经典命题逻辑。这是一个重要的限制,因为AI中的许多表示方案不包含命题逻辑。在本文中,我们考虑的问题是逻辑上的最小需求是什么,这样就可以制定AGM方法进行修订。我们证明,即使对底层语言及其语义的假设很少,也可以获得agm风格的修订;事实上,人们只需要一种语言,它的句子满足模型或可能世界的要求。在此框架中对经典的AGM公设进行了表达,并建立了公设集与可能世界上某些预定值之间的表示结果。为了得到表示结果,我们在AGM公设上增加了一个新的公设,并对世界上的预定量增加了一个约束。至关重要的是,这两个附加内容在原始的AGM框架中都是冗余的,因此我们扩展而不是修改AGM方法。此外,还讨论了迭代修订,并证明了达尔文/珀尔假设与我们的方法是兼容的。给出了各种例子来说明这种方法,包括霍恩子句修正,扩展逻辑程序中的修正,以及称为文字修正的非常基本逻辑中的信念修正。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Synchronization Strings: Codes for Insertions and Deletions Approaching the Singleton Bound The Reachability Problem for Two-Dimensional Vector Addition Systems with States Invited Articles Foreword On Nonconvex Optimization for Machine Learning Exploiting Spontaneous Transmissions for Broadcasting and Leader Election in Radio Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1