F. Alotaibi, K. Alyoubi, Ajay Mittal, Vishal Gupta, Navdeep Kaur
{"title":"TinyCheXReport: Compressed deep neural network for Chest X-ray report generation","authors":"F. Alotaibi, K. Alyoubi, Ajay Mittal, Vishal Gupta, Navdeep Kaur","doi":"10.1145/3676166","DOIUrl":null,"url":null,"abstract":"Increase in chest x-ray (CXR) imaging tests has burdened radiologists, thereby posing significant challenges in writing radiological reports on time. Although several deep learning-based automatic report generation methods have been developed till date, most are over-parameterized. For deployment on edge devices with constrained processing power or limited resources, over-parameterized models are often too large. This paper presents a compressed deep learning-based model that is 30% space efficient as compared to the non-compressed base model while both have comparable performance. The model comprising of VGG19 and hierarchical long short-term memory (LSTM) equipped with contextual word embedding layer is used as the base model. The redundant weight parameters are removed from the base model using unstructured one-shot pruning. To overcome the performance degradation, the lightweight pruned model is fine-tuned over publically available OpenI dataset. The quantitative evaluation metric scores demonstrate that proposed model surpasses the performance of state-of-the-art models. Additionally, the proposed model, being 30% space efficient, is easily deployable in resource-limited settings. Thus, this study serves as baseline for development of compressed models to generate radiologicalreports from CXR images.","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":null,"pages":null},"PeriodicalIF":1.8000,"publicationDate":"2024-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Asian and Low-Resource Language Information Processing","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1145/3676166","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Increase in chest x-ray (CXR) imaging tests has burdened radiologists, thereby posing significant challenges in writing radiological reports on time. Although several deep learning-based automatic report generation methods have been developed till date, most are over-parameterized. For deployment on edge devices with constrained processing power or limited resources, over-parameterized models are often too large. This paper presents a compressed deep learning-based model that is 30% space efficient as compared to the non-compressed base model while both have comparable performance. The model comprising of VGG19 and hierarchical long short-term memory (LSTM) equipped with contextual word embedding layer is used as the base model. The redundant weight parameters are removed from the base model using unstructured one-shot pruning. To overcome the performance degradation, the lightweight pruned model is fine-tuned over publically available OpenI dataset. The quantitative evaluation metric scores demonstrate that proposed model surpasses the performance of state-of-the-art models. Additionally, the proposed model, being 30% space efficient, is easily deployable in resource-limited settings. Thus, this study serves as baseline for development of compressed models to generate radiologicalreports from CXR images.
期刊介绍:
The ACM Transactions on Asian and Low-Resource Language Information Processing (TALLIP) publishes high quality original archival papers and technical notes in the areas of computation and processing of information in Asian languages, low-resource languages of Africa, Australasia, Oceania and the Americas, as well as related disciplines. The subject areas covered by TALLIP include, but are not limited to:
-Computational Linguistics: including computational phonology, computational morphology, computational syntax (e.g. parsing), computational semantics, computational pragmatics, etc.
-Linguistic Resources: including computational lexicography, terminology, electronic dictionaries, cross-lingual dictionaries, electronic thesauri, etc.
-Hardware and software algorithms and tools for Asian or low-resource language processing, e.g., handwritten character recognition.
-Information Understanding: including text understanding, speech understanding, character recognition, discourse processing, dialogue systems, etc.
-Machine Translation involving Asian or low-resource languages.
-Information Retrieval: including natural language processing (NLP) for concept-based indexing, natural language query interfaces, semantic relevance judgments, etc.
-Information Extraction and Filtering: including automatic abstraction, user profiling, etc.
-Speech processing: including text-to-speech synthesis and automatic speech recognition.
-Multimedia Asian Information Processing: including speech, image, video, image/text translation, etc.
-Cross-lingual information processing involving Asian or low-resource languages.
-Papers that deal in theory, systems design, evaluation and applications in the aforesaid subjects are appropriate for TALLIP. Emphasis will be placed on the originality and the practical significance of the reported research.