AdaCAD:自适应解码以平衡上下文知识和参数知识之间的冲突

Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal
{"title":"AdaCAD:自适应解码以平衡上下文知识和参数知识之间的冲突","authors":"Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal","doi":"arxiv-2409.07394","DOIUrl":null,"url":null,"abstract":"Knowledge conflict arises from discrepancies between information in the\ncontext of a large language model (LLM) and the knowledge stored in its\nparameters. This can hurt performance when using standard decoding techniques,\nwhich tend to ignore the context. Existing test-time contrastive methods seek\nto address this by comparing the LLM's output distribution with and without the\ncontext and adjust the model according to the contrast between them. However,\nwe find that these methods frequently misjudge the degree of conflict and\nstruggle to handle instances that vary in their amount of conflict, with static\nmethods over-adjusting when conflict is absent. We propose a fine-grained,\ninstance-level approach called AdaCAD, which dynamically infers the weight of\nadjustment based on the degree of conflict, as measured by the Jensen-Shannon\ndivergence between distributions representing contextual and parametric\nknowledge. Our experiments across four models on six diverse question-answering\n(QA) datasets and three summarization tasks demonstrate that our training-free\nadaptive method consistently outperforms other decoding methods on QA, with\naverage accuracy gains of 14.21% (absolute) over a static contrastive baseline,\nand improves the factuality of summaries by 5.59 (AlignScore). Furthermore, our\nanalysis shows that while decoding with contrastive baselines hurts performance\nwhen conflict is absent, AdaCAD mitigates these losses, making it more\napplicable to real-world datasets in which some examples have conflict and\nothers do not.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"AdaCAD: Adaptively Decoding to Balance Conflicts between Contextual and Parametric Knowledge\",\"authors\":\"Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal\",\"doi\":\"arxiv-2409.07394\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Knowledge conflict arises from discrepancies between information in the\\ncontext of a large language model (LLM) and the knowledge stored in its\\nparameters. This can hurt performance when using standard decoding techniques,\\nwhich tend to ignore the context. Existing test-time contrastive methods seek\\nto address this by comparing the LLM's output distribution with and without the\\ncontext and adjust the model according to the contrast between them. However,\\nwe find that these methods frequently misjudge the degree of conflict and\\nstruggle to handle instances that vary in their amount of conflict, with static\\nmethods over-adjusting when conflict is absent. We propose a fine-grained,\\ninstance-level approach called AdaCAD, which dynamically infers the weight of\\nadjustment based on the degree of conflict, as measured by the Jensen-Shannon\\ndivergence between distributions representing contextual and parametric\\nknowledge. Our experiments across four models on six diverse question-answering\\n(QA) datasets and three summarization tasks demonstrate that our training-free\\nadaptive method consistently outperforms other decoding methods on QA, with\\naverage accuracy gains of 14.21% (absolute) over a static contrastive baseline,\\nand improves the factuality of summaries by 5.59 (AlignScore). Furthermore, our\\nanalysis shows that while decoding with contrastive baselines hurts performance\\nwhen conflict is absent, AdaCAD mitigates these losses, making it more\\napplicable to real-world datasets in which some examples have conflict and\\nothers do not.\",\"PeriodicalId\":501030,\"journal\":{\"name\":\"arXiv - CS - Computation and Language\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computation and Language\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.07394\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computation and Language","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07394","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

知识冲突源于大型语言模型(LLM)上下文中的信息与其参数中存储的知识之间的差异。这可能会影响使用标准解码技术时的性能,因为标准解码技术往往会忽略上下文。现有的测试时间对比方法试图通过比较有语境和无语境的 LLM 输出分布来解决这一问题,并根据两者之间的对比来调整模型。然而,我们发现这些方法经常误判冲突程度,难以处理冲突程度不同的实例,而静态方法会在没有冲突时过度调整。我们提出了一种名为 AdaCAD 的细粒度实例级方法,它可以根据冲突程度动态推断调整权重,冲突程度由代表上下文知识和参数知识的分布之间的詹森-香农发散度来衡量。我们在六个不同的问题解答(QA)数据集和三个摘要任务上对四个模型进行的实验表明,我们的免训练自适应方法在 QA 上的表现始终优于其他解码方法,与静态对比基线相比,平均准确率提高了 14.21%(绝对值),摘要的事实性提高了 5.59(AlignScore)。此外,我们的分析表明,虽然使用对比基线解码会在没有冲突时损害性能,但 AdaCAD 能减轻这些损失,使其更适用于现实世界中一些例子有冲突而另一些例子没有冲突的数据集。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
AdaCAD: Adaptively Decoding to Balance Conflicts between Contextual and Parametric Knowledge
Knowledge conflict arises from discrepancies between information in the context of a large language model (LLM) and the knowledge stored in its parameters. This can hurt performance when using standard decoding techniques, which tend to ignore the context. Existing test-time contrastive methods seek to address this by comparing the LLM's output distribution with and without the context and adjust the model according to the contrast between them. However, we find that these methods frequently misjudge the degree of conflict and struggle to handle instances that vary in their amount of conflict, with static methods over-adjusting when conflict is absent. We propose a fine-grained, instance-level approach called AdaCAD, which dynamically infers the weight of adjustment based on the degree of conflict, as measured by the Jensen-Shannon divergence between distributions representing contextual and parametric knowledge. Our experiments across four models on six diverse question-answering (QA) datasets and three summarization tasks demonstrate that our training-free adaptive method consistently outperforms other decoding methods on QA, with average accuracy gains of 14.21% (absolute) over a static contrastive baseline, and improves the factuality of summaries by 5.59 (AlignScore). Furthermore, our analysis shows that while decoding with contrastive baselines hurts performance when conflict is absent, AdaCAD mitigates these losses, making it more applicable to real-world datasets in which some examples have conflict and others do not.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
LLMs + Persona-Plug = Personalized LLMs MEOW: MEMOry Supervised LLM Unlearning Via Inverted Facts Extract-and-Abstract: Unifying Extractive and Abstractive Summarization within Single Encoder-Decoder Framework Development and bilingual evaluation of Japanese medical large language model within reasonably low computational resources Human-like Affective Cognition in Foundation Models
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1