Felix Schneider, Sebastian Stüker, V. Parthasarathy
{"title":"Team Zoom @ AutoMin 2021: Cross-domain Pretraining for Automatic Minuting","authors":"Felix Schneider, Sebastian Stüker, V. Parthasarathy","doi":"10.21437/automin.2021-11","DOIUrl":null,"url":null,"abstract":"This Paper describes Zoom’s submission to the First Shared Task on Automatic Minuting at Interspeech 2021. We participated in Task A: generating abstractive summaries of meetings. For this task, we use a transformer-based summarization model which is first trained on data from a similar domain and then finetuned for domain transfer. In this configuration, our model does not yet produce usable summaries. We theorize that in the choice of pretraining corpus, the target side is more important than the source.","PeriodicalId":186820,"journal":{"name":"First Shared Task on Automatic Minuting at Interspeech 2021","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"First Shared Task on Automatic Minuting at Interspeech 2021","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21437/automin.2021-11","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
This Paper describes Zoom’s submission to the First Shared Task on Automatic Minuting at Interspeech 2021. We participated in Task A: generating abstractive summaries of meetings. For this task, we use a transformer-based summarization model which is first trained on data from a similar domain and then finetuned for domain transfer. In this configuration, our model does not yet produce usable summaries. We theorize that in the choice of pretraining corpus, the target side is more important than the source.