June-Woo Kim, Miika Toikkanen, Sangmin Bae, Minseok Kim, Ho-Young Jung
{"title":"RepAugment:用于呼吸声分类的输入诊断表征级增强技术","authors":"June-Woo Kim, Miika Toikkanen, Sangmin Bae, Minseok Kim, Ho-Young Jung","doi":"arxiv-2405.02996","DOIUrl":null,"url":null,"abstract":"Recent advancements in AI have democratized its deployment as a healthcare\nassistant. While pretrained models from large-scale visual and audio datasets\nhave demonstrably generalized to this task, surprisingly, no studies have\nexplored pretrained speech models, which, as human-originated sounds,\nintuitively would share closer resemblance to lung sounds. This paper explores\nthe efficacy of pretrained speech models for respiratory sound classification.\nWe find that there is a characterization gap between speech and lung sound\nsamples, and to bridge this gap, data augmentation is essential. However, the\nmost widely used augmentation technique for audio and speech, SpecAugment,\nrequires 2-dimensional spectrogram format and cannot be applied to models\npretrained on speech waveforms. To address this, we propose RepAugment, an\ninput-agnostic representation-level augmentation technique that outperforms\nSpecAugment, but is also suitable for respiratory sound classification with\nwaveform pretrained models. Experimental results show that our approach\noutperforms the SpecAugment, demonstrating a substantial improvement in the\naccuracy of minority disease classes, reaching up to 7.14%.","PeriodicalId":501178,"journal":{"name":"arXiv - CS - Sound","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"RepAugment: Input-Agnostic Representation-Level Augmentation for Respiratory Sound Classification\",\"authors\":\"June-Woo Kim, Miika Toikkanen, Sangmin Bae, Minseok Kim, Ho-Young Jung\",\"doi\":\"arxiv-2405.02996\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent advancements in AI have democratized its deployment as a healthcare\\nassistant. While pretrained models from large-scale visual and audio datasets\\nhave demonstrably generalized to this task, surprisingly, no studies have\\nexplored pretrained speech models, which, as human-originated sounds,\\nintuitively would share closer resemblance to lung sounds. This paper explores\\nthe efficacy of pretrained speech models for respiratory sound classification.\\nWe find that there is a characterization gap between speech and lung sound\\nsamples, and to bridge this gap, data augmentation is essential. However, the\\nmost widely used augmentation technique for audio and speech, SpecAugment,\\nrequires 2-dimensional spectrogram format and cannot be applied to models\\npretrained on speech waveforms. To address this, we propose RepAugment, an\\ninput-agnostic representation-level augmentation technique that outperforms\\nSpecAugment, but is also suitable for respiratory sound classification with\\nwaveform pretrained models. Experimental results show that our approach\\noutperforms the SpecAugment, demonstrating a substantial improvement in the\\naccuracy of minority disease classes, reaching up to 7.14%.\",\"PeriodicalId\":501178,\"journal\":{\"name\":\"arXiv - CS - Sound\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Sound\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2405.02996\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Sound","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.02996","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
RepAugment: Input-Agnostic Representation-Level Augmentation for Respiratory Sound Classification
Recent advancements in AI have democratized its deployment as a healthcare
assistant. While pretrained models from large-scale visual and audio datasets
have demonstrably generalized to this task, surprisingly, no studies have
explored pretrained speech models, which, as human-originated sounds,
intuitively would share closer resemblance to lung sounds. This paper explores
the efficacy of pretrained speech models for respiratory sound classification.
We find that there is a characterization gap between speech and lung sound
samples, and to bridge this gap, data augmentation is essential. However, the
most widely used augmentation technique for audio and speech, SpecAugment,
requires 2-dimensional spectrogram format and cannot be applied to models
pretrained on speech waveforms. To address this, we propose RepAugment, an
input-agnostic representation-level augmentation technique that outperforms
SpecAugment, but is also suitable for respiratory sound classification with
waveform pretrained models. Experimental results show that our approach
outperforms the SpecAugment, demonstrating a substantial improvement in the
accuracy of minority disease classes, reaching up to 7.14%.