{"title":"Joint Reasoning on Hybrid-knowledge sources for Task-Oriented Dialog","authors":"Mayank Mishra, Danish Contractor, Dinesh Raghu","doi":"10.48550/arXiv.2210.07295","DOIUrl":null,"url":null,"abstract":"Traditional systems designed for task oriented dialog utilize knowledge present only in structured knowledge sources to generate responses. However, relevant information required to generate responses may also reside in unstructured sources, such as documents. Recent state of the art models such as HyKnow (Gao et al., 2021b) and SEKNOW (Gao et al., 2021a) aimed at overcoming these challenges make limiting assumptions about the knowledge sources. For instance, these systems assume that certain types of information, such as a phone number, is always present in a structured knowledge base (KB) while information about aspects such as entrance ticket prices, would always be available in documents.In this paper, we create a modified version of the MutliWOZ-based dataset prepared by (Gao et al., 2021a) to demonstrate how current methods have significant degradation in performance when strict assumptions about the source of information are removed. Then, in line with recent work exploiting pre-trained language models, we fine-tune a BART (Lewiset al., 2020) based model using prompts (Brown et al., 2020; Sun et al., 2021) for the tasks of querying knowledge sources, as well as, for response generation, without makingassumptions about the information present in each knowledge source. Through a series of experiments, we demonstrate that our model is robust to perturbations to knowledge modality (source of information), and that it can fuse information from structured as well as unstructured knowledge to generate responses.","PeriodicalId":73025,"journal":{"name":"Findings (Sydney (N.S.W.)","volume":"1 1","pages":"1733-1742"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Findings (Sydney (N.S.W.)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2210.07295","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Traditional systems designed for task oriented dialog utilize knowledge present only in structured knowledge sources to generate responses. However, relevant information required to generate responses may also reside in unstructured sources, such as documents. Recent state of the art models such as HyKnow (Gao et al., 2021b) and SEKNOW (Gao et al., 2021a) aimed at overcoming these challenges make limiting assumptions about the knowledge sources. For instance, these systems assume that certain types of information, such as a phone number, is always present in a structured knowledge base (KB) while information about aspects such as entrance ticket prices, would always be available in documents.In this paper, we create a modified version of the MutliWOZ-based dataset prepared by (Gao et al., 2021a) to demonstrate how current methods have significant degradation in performance when strict assumptions about the source of information are removed. Then, in line with recent work exploiting pre-trained language models, we fine-tune a BART (Lewiset al., 2020) based model using prompts (Brown et al., 2020; Sun et al., 2021) for the tasks of querying knowledge sources, as well as, for response generation, without makingassumptions about the information present in each knowledge source. Through a series of experiments, we demonstrate that our model is robust to perturbations to knowledge modality (source of information), and that it can fuse information from structured as well as unstructured knowledge to generate responses.
为任务导向对话设计的传统系统利用仅存在于结构化知识来源中的知识来生成响应。但是,生成响应所需的相关信息也可能驻留在非结构化源中,例如文档。最近的最先进的模型,如HyKnow (Gao等人,2021b)和SEKNOW (Gao等人,2021a)旨在克服这些挑战,对知识来源做出有限的假设。例如,这些系统假设某些类型的信息,如电话号码,总是存在于结构化知识库(KB)中,而关于某些方面的信息,如门票价格,总是在文档中可用。在本文中,我们创建了(Gao等人,2021a)准备的基于multiwoz的数据集的修改版本,以演示当删除关于信息源的严格假设时,当前方法如何显著降低性能。然后,根据最近利用预训练语言模型的工作,我们使用提示(Brown et al., 2020;Sun et al., 2021)用于查询知识来源的任务,以及响应生成,而无需对每个知识来源中存在的信息进行假设。通过一系列的实验,我们证明了我们的模型对知识模态(信息源)的扰动具有鲁棒性,并且它可以融合结构化和非结构化知识的信息来生成响应。