首页 > 最新文献

Natural Language Processing Journal最新文献

英文 中文
Evaluation of google translate for Mandarin Chinese translation using sentiment and semantic analysis 用情感和语义分析评价谷歌翻译对普通话翻译的影响
Pub Date : 2025-12-01 DOI: 10.1016/j.nlp.2025.100188
Xuechun Wang , Rodney Beard , Rohitash Chandra
Machine translation using large language models (LLMs) is having a significant global impact, making communication easier. Mandarin Chinese is the official language used for communication by the government and media in China. In this study, we provide an automated assessment of the translation quality of Google Translate with human experts using sentiment and semantic analysis. In order to demonstrate our framework, we select the classic early twentieth-century novel ’The True Story of Ah Q’ with selected Mandarin Chinese to English translations. We use Google Translate to translate the given text into English and then conduct a chapter-wise sentiment analysis and semantic analysis to compare the extracted sentiments across the different translations. Our results indicate that the precision of Google Translate differs both in terms of semantic and sentiment analysis when compared to human expert translations. We find that Google Translate is unable to translate some of the specific words or phrases in Chinese, such as Chinese traditional idiomatic expressions. The mistranslations may be due to a lack of contextual significance and historical knowledge of China.
使用大型语言模型(llm)的机器翻译正在产生重大的全球影响,使交流更容易。普通话是中国政府和媒体交流使用的官方语言。在这项研究中,我们与人类专家一起使用情感和语义分析对b谷歌翻译的翻译质量进行了自动评估。为了证明我们的框架,我们选择了20世纪早期的经典小说《阿Q正传》,并选择了中文翻译成英文。我们使用谷歌Translate将给定的文本翻译成英文,然后进行逐章情感分析和语义分析,比较不同翻译中提取的情感。我们的研究结果表明,谷歌翻译在语义和情感分析方面的精度与人类专家翻译相比有所不同。我们发现谷歌翻译无法翻译中文中的一些特定单词或短语,例如中文传统成语。这些误译可能是由于缺乏语境意义和对中国历史的了解。
{"title":"Evaluation of google translate for Mandarin Chinese translation using sentiment and semantic analysis","authors":"Xuechun Wang ,&nbsp;Rodney Beard ,&nbsp;Rohitash Chandra","doi":"10.1016/j.nlp.2025.100188","DOIUrl":"10.1016/j.nlp.2025.100188","url":null,"abstract":"<div><div>Machine translation using large language models (LLMs) is having a significant global impact, making communication easier. Mandarin Chinese is the official language used for communication by the government and media in China. In this study, we provide an automated assessment of the translation quality of Google Translate with human experts using sentiment and semantic analysis. In order to demonstrate our framework, we select the classic early twentieth-century novel ’The True Story of Ah Q’ with selected Mandarin Chinese to English translations. We use Google Translate to translate the given text into English and then conduct a chapter-wise sentiment analysis and semantic analysis to compare the extracted sentiments across the different translations. Our results indicate that the precision of Google Translate differs both in terms of semantic and sentiment analysis when compared to human expert translations. We find that Google Translate is unable to translate some of the specific words or phrases in Chinese, such as Chinese traditional idiomatic expressions. The mistranslations may be due to a lack of contextual significance and historical knowledge of China.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100188"},"PeriodicalIF":0.0,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145736272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bridging gaps in natural language processing for Yorùbá: A systematic review of a decade of progress and prospects 弥合Yorùbá自然语言处理的差距:十年来进展和前景的系统回顾
Pub Date : 2025-11-09 DOI: 10.1016/j.nlp.2025.100194
Toheeb Aduramomi Jimoh, Tabea De Wille, Nikola S. Nikolov
Natural Language Processing (NLP) is becoming a dominant subset of artificial intelligence as the need to help machines understand human language becomes indispensable. Several NLP applications are ubiquitous, partly due to the myriad datasets being churned out daily through mediums like social networking sites. However, the growing development has not been evident in most African languages due to the persisting resource limitations, among other issues. Yorùbá language, a tonal and morphologically rich African language, suffers a similar fate, resulting in limited NLP usage. To encourage further research towards improving this situation, this systematic literature review aims to comprehensively analyse studies addressing NLP development for Yorùbá, identifying challenges, resources, techniques, and applications. A well-defined search string from a structured protocol was employed to search, select, and analyse 105 primary studies between 2014 and 2024 from reputable databases. The review highlights the scarcity of annotated corpora, the limited availability of pre-trained language models (PLMs), and linguistic challenges like tonal complexity and diacritic dependency as significant obstacles. It also revealed the prominent techniques, including rule-based methods, statistical methods, deep learning, and transfer learning, which were implemented alongside datasets of Yorùbá speech corpora, among others. The findings reveal a growing body of multilingual and monolingual resources, even though the field is constrained by socio-cultural factors such as code-switching and the desertion of language for digital usage. This review synthesises existing research, providing a foundation for advancing NLP for Yorùbá and in African languages generally. It aims to guide future research by identifying gaps and opportunities, thereby contributing to the broader inclusion of Yorùbá and other under-resourced African languages in global NLP advancements.
随着帮助机器理解人类语言的需求变得不可或缺,自然语言处理(NLP)正在成为人工智能的一个主要子集。一些NLP应用程序无处不在,部分原因是每天通过社交网站等媒介大量产生数据集。然而,除其他问题外,由于持续的资源限制,大多数非洲语言的发展并不明显。Yorùbá语言,一种音调和形态丰富的非洲语言,遭受类似的命运,导致有限的NLP使用。为了鼓励进一步的研究以改善这种情况,本系统的文献综述旨在全面分析Yorùbá的NLP发展研究,确定挑战,资源,技术和应用。从结构化协议中使用定义良好的搜索字符串来搜索、选择和分析2014年至2024年间来自知名数据库的105项主要研究。该综述强调了标注语料库的稀缺性,预训练语言模型(PLMs)的有限可用性,以及音调复杂性和变音符依赖等语言挑战是重大障碍。它还揭示了突出的技术,包括基于规则的方法、统计方法、深度学习和迁移学习,这些技术与Yorùbá语音语料库等数据集一起实现。研究结果显示,尽管该领域受到社会文化因素(如代码转换和为数字使用而放弃语言)的限制,但多语言和单语言资源的数量仍在不断增长。本综述综合了现有的研究,为推进Yorùbá和非洲语言的自然语言处理提供了基础。它旨在通过确定差距和机会来指导未来的研究,从而促进将Yorùbá和其他资源不足的非洲语言更广泛地纳入全球自然语言处理进程。
{"title":"Bridging gaps in natural language processing for Yorùbá: A systematic review of a decade of progress and prospects","authors":"Toheeb Aduramomi Jimoh,&nbsp;Tabea De Wille,&nbsp;Nikola S. Nikolov","doi":"10.1016/j.nlp.2025.100194","DOIUrl":"10.1016/j.nlp.2025.100194","url":null,"abstract":"<div><div>Natural Language Processing (NLP) is becoming a dominant subset of artificial intelligence as the need to help machines understand human language becomes indispensable. Several NLP applications are ubiquitous, partly due to the myriad datasets being churned out daily through mediums like social networking sites. However, the growing development has not been evident in most African languages due to the persisting resource limitations, among other issues. Yorùbá language, a tonal and morphologically rich African language, suffers a similar fate, resulting in limited NLP usage. To encourage further research towards improving this situation, this systematic literature review aims to comprehensively analyse studies addressing NLP development for Yorùbá, identifying challenges, resources, techniques, and applications. A well-defined search string from a structured protocol was employed to search, select, and analyse 105 primary studies between 2014 and 2024 from reputable databases. The review highlights the scarcity of annotated corpora, the limited availability of pre-trained language models (PLMs), and linguistic challenges like tonal complexity and diacritic dependency as significant obstacles. It also revealed the prominent techniques, including rule-based methods, statistical methods, deep learning, and transfer learning, which were implemented alongside datasets of Yorùbá speech corpora, among others. The findings reveal a growing body of multilingual and monolingual resources, even though the field is constrained by socio-cultural factors such as code-switching and the desertion of language for digital usage. This review synthesises existing research, providing a foundation for advancing NLP for Yorùbá and in African languages generally. It aims to guide future research by identifying gaps and opportunities, thereby contributing to the broader inclusion of Yorùbá and other under-resourced African languages in global NLP advancements.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100194"},"PeriodicalIF":0.0,"publicationDate":"2025-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145571255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Llama3SP: A resource-Efficient large language model for agile story point estimation Llama3SP:用于敏捷故事点估计的资源高效的大型语言模型
Pub Date : 2025-11-08 DOI: 10.1016/j.nlp.2025.100189
Juan Camilo Sepúlveda Montoya , Nicole Tatiana Ríos Gómez , José A. Jaramillo Villegas
Effort estimation remains a major challenge in Agile software development. Inaccurate story point forecasts can lead to budget overruns, schedule delays, and diminished stakeholder trust. Widely used approaches, such as story point estimation, are helpful for planning but rely heavily on subjective human judgment, making them prone to inconsistency and bias. Prior efforts applying machine learning and natural language processing (e.g. Deep-SE, GPT2SP) to automate story point prediction have achieved only limited success, often suffering from accuracy issues, poor cross-project adaptability, and high computational costs. To address these challenges, we introduce Llama3SP, which fine-tunes Meta’s LLaMA 3.2 language model using QLoRA, a resource-efficient adaptation technique. This combination enables training of a high-performance model on standard GPUs without sacrificing prediction quality. Experiments show that Llama3SP provides precise and consistent story point estimates, outperforming or matching previous models like GPT2SP and other comparably sized alternatives, all while operating under significantly lower hardware constraints. These findings highlight how combining advanced NLP models with efficient training techniques can make accurate effort estimation more accessible and practical for agile teams.
工作量估算仍然是敏捷软件开发中的一个主要挑战。不准确的故事点预测会导致预算超支、进度延迟和减少涉众的信任。广泛使用的方法,如故事点估计,对计划有帮助,但严重依赖于主观的人类判断,使它们容易产生不一致和偏见。先前应用机器学习和自然语言处理(例如Deep-SE, GPT2SP)来自动化故事点预测的努力只取得了有限的成功,经常受到准确性问题、跨项目适应性差和高计算成本的困扰。为了应对这些挑战,我们引入了Llama3SP,它使用QLoRA(一种资源高效的自适应技术)对Meta的LLaMA 3.2语言模型进行微调。这种组合可以在不牺牲预测质量的情况下在标准gpu上训练高性能模型。实验表明,Llama3SP提供了精确和一致的故事点估计,优于或匹配以前的模型,如GPT2SP和其他相当大小的替代方案,同时在更低的硬件约束下运行。这些发现强调了如何将先进的NLP模型与有效的培训技术相结合,使敏捷团队更容易获得和实用准确的工作量估计。
{"title":"Llama3SP: A resource-Efficient large language model for agile story point estimation","authors":"Juan Camilo Sepúlveda Montoya ,&nbsp;Nicole Tatiana Ríos Gómez ,&nbsp;José A. Jaramillo Villegas","doi":"10.1016/j.nlp.2025.100189","DOIUrl":"10.1016/j.nlp.2025.100189","url":null,"abstract":"<div><div>Effort estimation remains a major challenge in Agile software development. Inaccurate story point forecasts can lead to budget overruns, schedule delays, and diminished stakeholder trust. Widely used approaches, such as story point estimation, are helpful for planning but rely heavily on subjective human judgment, making them prone to inconsistency and bias. Prior efforts applying machine learning and natural language processing (e.g. Deep-SE, GPT2SP) to automate story point prediction have achieved only limited success, often suffering from accuracy issues, poor cross-project adaptability, and high computational costs. To address these challenges, we introduce <strong>Llama3SP</strong>, which fine-tunes Meta’s LLaMA 3.2 language model using QLoRA, a resource-efficient adaptation technique. This combination enables training of a high-performance model on standard GPUs without sacrificing prediction quality. Experiments show that Llama3SP provides precise and consistent story point estimates, outperforming or matching previous models like GPT2SP and other comparably sized alternatives, all while operating under significantly lower hardware constraints. These findings highlight how combining advanced NLP models with efficient training techniques can make accurate effort estimation more accessible and practical for agile teams.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100189"},"PeriodicalIF":0.0,"publicationDate":"2025-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A systematic review of figurative language detection: Methods, challenges, and multilingual perspectives 比喻语言检测的系统回顾:方法、挑战和多语言视角
Pub Date : 2025-11-04 DOI: 10.1016/j.nlp.2025.100192
Zouheir Banou, Sanaa El Filali, El Habib Benlahmar, Fatima-Zahra Alaoui, Laila El Jiani, Hasnae Sakhi
Figurative language detection has emerged as a critical task in natural language processing (NLP), enabling machines to comprehend non-literal expressions such as metaphor, irony, and sarcasm. This study presents a systematic literature review with a multilevel analytical framework, examining figurative language across lexical, syntactic, semantic, discourse, and pragmatic levels. We investigate the interplay between feature engineering, model architectures, and annotation strategies across different languages, analyzing datasets, linguistic resources, and evaluation metrics. Special attention is given to morphologically rich and low-resource languages, where deep learning dominates but rule-based and hybrid approaches remain relevant. Our findings indicate that deep learning models–especially transformer-based architectures like BERT and RoBERTa–consistently outperform other approaches, particularly in semantic and discourse-level tasks, due to their ability to capture context-rich and abstract patterns. However, these models often lack interpretability, raising concerns about transparency. Additional challenges include inconsistencies in annotation practices, class imbalance between figurative and literal instances, and limited data coverage for under-resourced languages. The absence of standardized evaluation metrics further complicates cross-study comparison, especially when diverse figurative language styles are involved. By structuring our analysis through linguistic and computational dimensions, this review aims to facilitate the development of more robust, inclusive, and explainable figurative language detection systems.
比喻语言检测已经成为自然语言处理(NLP)中的一项关键任务,使机器能够理解非文字表达,如隐喻、反讽和讽刺。本研究采用多层次的分析框架,对比喻语言在词汇、句法、语义、语篇和语用等层面进行了系统的文献综述。我们研究了跨不同语言的特征工程、模型架构和注释策略之间的相互作用,分析了数据集、语言资源和评估指标。特别关注形态丰富和低资源的语言,其中深度学习占主导地位,但基于规则的方法和混合方法仍然相关。我们的研究结果表明,深度学习模型——尤其是像BERT和roberta这样基于转换的架构——一直优于其他方法,特别是在语义和话语级任务中,因为它们能够捕获丰富的上下文和抽象模式。然而,这些模型往往缺乏可解释性,引发了对透明度的担忧。其他的挑战包括注释实践中的不一致、形象化实例和文字实例之间的类不平衡,以及资源不足语言的有限数据覆盖。缺乏标准化的评估指标进一步使交叉研究比较复杂化,特别是当涉及到不同的比喻语言风格时。通过语言和计算维度构建我们的分析,本综述旨在促进更稳健、包容和可解释的比喻语言检测系统的发展。
{"title":"A systematic review of figurative language detection: Methods, challenges, and multilingual perspectives","authors":"Zouheir Banou,&nbsp;Sanaa El Filali,&nbsp;El Habib Benlahmar,&nbsp;Fatima-Zahra Alaoui,&nbsp;Laila El Jiani,&nbsp;Hasnae Sakhi","doi":"10.1016/j.nlp.2025.100192","DOIUrl":"10.1016/j.nlp.2025.100192","url":null,"abstract":"<div><div>Figurative language detection has emerged as a critical task in natural language processing (NLP), enabling machines to comprehend non-literal expressions such as metaphor, irony, and sarcasm. This study presents a systematic literature review with a multilevel analytical framework, examining figurative language across lexical, syntactic, semantic, discourse, and pragmatic levels. We investigate the interplay between feature engineering, model architectures, and annotation strategies across different languages, analyzing datasets, linguistic resources, and evaluation metrics. Special attention is given to morphologically rich and low-resource languages, where deep learning dominates but rule-based and hybrid approaches remain relevant. Our findings indicate that deep learning models–especially transformer-based architectures like BERT and RoBERTa–consistently outperform other approaches, particularly in semantic and discourse-level tasks, due to their ability to capture context-rich and abstract patterns. However, these models often lack interpretability, raising concerns about transparency. Additional challenges include inconsistencies in annotation practices, class imbalance between figurative and literal instances, and limited data coverage for under-resourced languages. The absence of standardized evaluation metrics further complicates cross-study comparison, especially when diverse figurative language styles are involved. By structuring our analysis through linguistic and computational dimensions, this review aims to facilitate the development of more robust, inclusive, and explainable figurative language detection systems.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100192"},"PeriodicalIF":0.0,"publicationDate":"2025-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145571256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on the methodology of personalized recommender systems based on multimodal knowledge graphs 基于多模态知识图的个性化推荐系统方法研究
Pub Date : 2025-11-01 DOI: 10.1016/j.nlp.2025.100193
Shaowu Bao , Jiajia Wang
The exponential increase in learning materials has occasioned a greater need for personalized learning experiences, yet conventional unimodal recommender systems are not effective in addressing students' diversified demands.In this research, a personalized recommendation system is advocated, backed by a multimodal knowledge graph that consolidates text, image, and video knowledge to improve accuracy, interpretability, and adaptability.The system uses different algorithms for relationship and entity extractions and includes a graph attention module with hierarchical subgraphs to build a semantic network among "Knowledge Points–Students–Resources." A dual-path embedding module that fuses Node2vec for structural semantics with LSTM for learning temporal behavior provides explainable recommendations using path confidence.Experimental results demonstrate that the model's core performance comprehensively outperforms traditional methods and newly introduced comparison models: Entity alignment accuracy (Hits@10=62.7 %) improved by 13.4 % over traditional Node2vec, 6.8 % over KGAT, and 4.2 % over M3KGR; cross-modal similarity (0.76) increased by 11.8 % over traditional Node2vec and 5.6 % over M3KGR. Learning engagement (effective duration 65 %, completion rate 78 %) and knowledge acquisition efficiency (coverage 67 %, cycle reduction 30 %) are significantly optimized, improving by 8.3 %-11.4 % and 8.1 %-20 % over M3KGR respectively; Achieved an explainability score of 4.3 (34.4 %-104.8 % improvement over traditional methods, 22.9 % improvement over KGAT, and 13.2 % improvement over M3KGR), with a response time of 98 ms (40.6 % reduction compared to KGAT and 25.8 % reduction compared to M3KGR).This indicates that multimodal knowledge graph significantly improves recommendation performance through structured semantics and dynamic fusion,providing a new path for personalized education.
学习材料的指数级增长引发了对个性化学习体验的更大需求,然而传统的单模推荐系统并不能有效地满足学生的多样化需求。本研究提倡个性化推荐系统,以多模态知识图为支撑,整合文本、图像和视频知识,提高推荐的准确性、可解释性和适应性。该系统采用不同的算法对关系和实体进行提取,并包括一个带有分层子图的图形关注模块,构建了“知识点-学生-资源”之间的语义网络。双路径嵌入模块融合了用于结构语义的Node2vec和用于学习时态行为的LSTM,使用路径置信度提供了可解释的建议。实验结果表明,该模型的核心性能全面优于传统方法和新引入的比较模型:实体对准精度(Hits@10= 62.7%)比传统Node2vec提高13.4%,比KGAT提高6.8%,比M3KGR提高4.2%;跨模态相似性(0.76)比传统Node2vec增加11.8%,比M3KGR增加5.6%。学习参与(有效持续时间65%,完成率78%)和知识获取效率(覆盖率67%,周期减少30%)显著优化,分别比M3KGR提高8.3% - 11.4%和8.1% - 20%;可解释性得分为4.3分(比传统方法提高34.4% - 104.8%,比KGAT提高22.9%,比M3KGR提高13.2%),响应时间为98 ms(比KGAT减少40.6%,比M3KGR减少25.8%)。这表明多模态知识图通过结构化语义和动态融合显著提高了推荐性能,为个性化教育提供了新的路径。
{"title":"Research on the methodology of personalized recommender systems based on multimodal knowledge graphs","authors":"Shaowu Bao ,&nbsp;Jiajia Wang","doi":"10.1016/j.nlp.2025.100193","DOIUrl":"10.1016/j.nlp.2025.100193","url":null,"abstract":"<div><div>The exponential increase in learning materials has occasioned a greater need for personalized learning experiences, yet conventional unimodal recommender systems are not effective in addressing students' diversified demands.In this research, a personalized recommendation system is advocated, backed by a multimodal knowledge graph that consolidates text, image, and video knowledge to improve accuracy, interpretability, and adaptability.The system uses different algorithms for relationship and entity extractions and includes a graph attention module with hierarchical subgraphs to build a semantic network among \"Knowledge Points–Students–Resources.\" A dual-path embedding module that fuses Node2vec for structural semantics with LSTM for learning temporal behavior provides explainable recommendations using path confidence.Experimental results demonstrate that the model's core performance comprehensively outperforms traditional methods and newly introduced comparison models: Entity alignment accuracy (Hits@10=62.7 %) improved by 13.4 % over traditional Node2vec, 6.8 % over KGAT, and 4.2 % over M3KGR; cross-modal similarity (0.76) increased by 11.8 % over traditional Node2vec and 5.6 % over M3KGR. Learning engagement (effective duration 65 %, completion rate 78 %) and knowledge acquisition efficiency (coverage 67 %, cycle reduction 30 %) are significantly optimized, improving by 8.3 %-11.4 % and 8.1 %-20 % over M3KGR respectively; Achieved an explainability score of 4.3 (34.4 %-104.8 % improvement over traditional methods, 22.9 % improvement over KGAT, and 13.2 % improvement over M3KGR), with a response time of 98 ms (40.6 % reduction compared to KGAT and 25.8 % reduction compared to M3KGR).This indicates that multimodal knowledge graph significantly improves recommendation performance through structured semantics and dynamic fusion,providing a new path for personalized education.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100193"},"PeriodicalIF":0.0,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A multi-class cyberbullying classification on image and text in code-mixed Bangla-English social media content 代码混合孟加拉-英语社交媒体内容中图像和文本的多类网络欺凌分类
Pub Date : 2025-10-30 DOI: 10.1016/j.nlp.2025.100191
Animesh Chandra Roy , Tanvir Mahmud , Tahlil Abrar
Social media platforms like Facebook, Instagram, and Twitter are widely used; users frequently share their daily lives by uploading pictures, posts, and videos, which gain significant popularity. However, social media posts often receive a mix of reactions, ranging from positive to negative, and in some instances, negative comments escalate into cyberbullying. Numerous studies have addressed this issue by focusing on cyberbullying classification, primarily through binary classification using multimodal data or targeting either text or image data. This study investigates the identification of multi-class images like No-bullying, Religious, Sexual, and Others using the deep learning pre-trained model MobileNetV2 to detect multiple image labels and achieved an F1-score of 0.86. For categorizing hate comments, we consider multiple classes, including Not Hate, Slang, Sexual, Racial, and Religious-related content. Extensive experiments were conducted on a novel Bengali-English code-mixed dataset, utilizing a combination of advanced transformer models, traditional machine learning techniques, and deep learning approaches to detect multiple hate comment labels. Bangla BERT achieved the highest F1-score of 0.79, followed closely by SVM at 0.78 and BiLSTM with attention at 0.73. These findings underscore the effectiveness of these models in capturing the complexities of code-mixed Bengali-English, offering valuable insights into cyberbullying detection in diverse linguistic contexts. This research contributes essential strategies for improving online safety and fostering respectful digital interactions.
Facebook、Instagram和Twitter等社交媒体平台被广泛使用;用户通过上传图片、帖子、视频等方式,频繁地分享自己的日常生活,获得了极大的人气。然而,社交媒体上的帖子通常会收到褒贬不一的反应,在某些情况下,负面评论会升级为网络欺凌。许多研究通过关注网络欺凌分类来解决这个问题,主要是通过使用多模态数据或针对文本或图像数据的二元分类。本研究使用深度学习预训练模型MobileNetV2检测多个图像标签,研究了No-bullying、Religious、Sexual和Others等多类别图像的识别,并获得了f1 -得分0.86。为了对仇恨评论进行分类,我们考虑了多个类别,包括非仇恨、俚语、性、种族和宗教相关内容。在一个新的孟加拉语-英语代码混合数据集上进行了广泛的实验,利用先进的变压器模型、传统机器学习技术和深度学习方法的组合来检测多个仇恨评论标签。孟加拉语BERT的f1得分最高,为0.79,紧随其后的是SVM,为0.78,BiLSTM的关注度为0.73。这些发现强调了这些模型在捕捉代码混合孟加拉语-英语的复杂性方面的有效性,为在不同语言背景下检测网络欺凌提供了有价值的见解。这项研究为提高网络安全和促进相互尊重的数字互动提供了必要的策略。
{"title":"A multi-class cyberbullying classification on image and text in code-mixed Bangla-English social media content","authors":"Animesh Chandra Roy ,&nbsp;Tanvir Mahmud ,&nbsp;Tahlil Abrar","doi":"10.1016/j.nlp.2025.100191","DOIUrl":"10.1016/j.nlp.2025.100191","url":null,"abstract":"<div><div>Social media platforms like Facebook, Instagram, and Twitter are widely used; users frequently share their daily lives by uploading pictures, posts, and videos, which gain significant popularity. However, social media posts often receive a mix of reactions, ranging from positive to negative, and in some instances, negative comments escalate into cyberbullying. Numerous studies have addressed this issue by focusing on cyberbullying classification, primarily through binary classification using multimodal data or targeting either text or image data. This study investigates the identification of multi-class images like No-bullying, Religious, Sexual, and Others using the deep learning pre-trained model MobileNetV2 to detect multiple image labels and achieved an F1-score of 0.86. For categorizing hate comments, we consider multiple classes, including Not Hate, Slang, Sexual, Racial, and Religious-related content. Extensive experiments were conducted on a novel Bengali-English code-mixed dataset, utilizing a combination of advanced transformer models, traditional machine learning techniques, and deep learning approaches to detect multiple hate comment labels. Bangla BERT achieved the highest F1-score of 0.79, followed closely by SVM at 0.78 and BiLSTM with attention at 0.73. These findings underscore the effectiveness of these models in capturing the complexities of code-mixed Bengali-English, offering valuable insights into cyberbullying detection in diverse linguistic contexts. This research contributes essential strategies for improving online safety and fostering respectful digital interactions.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100191"},"PeriodicalIF":0.0,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145466720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BERT-KAN: Enhancing bilingual sentiment analysis in bangladeshi E-commerce through fine-tuned large language models BERT-KAN:通过微调大型语言模型,加强孟加拉电子商务的双语情感分析
Pub Date : 2025-10-29 DOI: 10.1016/j.nlp.2025.100190
Mohammad Rifat Ahmmad Rashid, Aritra Das, Kazi Ferdous Hasan, Md. Rakibul Hasan, Mithila Sultana, Mahamudul Hasan, Raihan Ul Islam, Rashedul Amin Tuhin, M. Saddam Hossain Khan
Sentiment analysis of code-mixed reviews poses unique challenges due to linguistic variability and contex- tual ambiguity, particularly in multilingual e-commerce environments. In this paper, we introduce BERT- KAN, a novel hybrid architecture that enhances bilingual sentiment analysis in Bangladeshi e-commerce by integrating the deep contextual representations of Bidirectional Encoder Representations from Transform- ers(BERT) with a Kolmogorov-Arnold Network (KAN) layer. The KAN component employs a polynomial expansion to capture complex non-linear relationships within code-mixed Bengali-English text, while an innovative polynomial attention mechanism further refines feature extraction. Extensive ablation studies were conducted on two base models—bert-base-multilingual-uncased and BanglaBERT—using polynomial degrees of 2 and 3. Notably, the best configuration for bert-base-multilingual-uncased (employing KAN, polynomial attention, and feature fusion with polynomial degree 2) achieved a precision of 95.3 %, recall of 97.0 %, and an F1-score of 96.1 %. Comparable performance was observed for polynomial degree 3 (precision 96.2 %, recall 95.8 %, and F1-score 96.0 %), while cross-validation experiments yielded average accuracies exceeding 90 % across multiple folds. Detailed error analyses, supported by confusion matrices and sam- ple predictions, as well as discussions on computational requirements and deployment challenges, further validate the robustness of our approach.
由于语言的可变性和上下文的模糊性,特别是在多语言的电子商务环境中,代码混合评论的情感分析提出了独特的挑战。在本文中,我们介绍了BERT- KAN,这是一种新的混合架构,通过将来自Transform- ers(BERT)的双向编码器表示的深度上下文表示与Kolmogorov-Arnold网络(KAN)层集成,增强了孟加拉国电子商务中的双语情感分析。KAN组件采用多项式展开来捕获代码混合的孟加拉语-英语文本中复杂的非线性关系,而创新的多项式关注机制进一步改进了特征提取。使用多项式度为2和3的两个基本模型(bert-base-multilingual-uncase和banglabert)进行了广泛的消融研究。值得注意的是,bert-base- multilinguo -uncase的最佳配置(使用KAN、多项式注意和多项式度2的特征融合)达到了95.3%的准确率、97.0%的召回率和96.1%的f1分数。在多项式度为3的情况下(准确率为96.2%,召回率为95.8%,f1评分为96.0%),交叉验证实验的平均准确率超过90%。详细的误差分析,由混淆矩阵和样本预测支持,以及对计算需求和部署挑战的讨论,进一步验证了我们方法的鲁棒性。
{"title":"BERT-KAN: Enhancing bilingual sentiment analysis in bangladeshi E-commerce through fine-tuned large language models","authors":"Mohammad Rifat Ahmmad Rashid,&nbsp;Aritra Das,&nbsp;Kazi Ferdous Hasan,&nbsp;Md. Rakibul Hasan,&nbsp;Mithila Sultana,&nbsp;Mahamudul Hasan,&nbsp;Raihan Ul Islam,&nbsp;Rashedul Amin Tuhin,&nbsp;M. Saddam Hossain Khan","doi":"10.1016/j.nlp.2025.100190","DOIUrl":"10.1016/j.nlp.2025.100190","url":null,"abstract":"<div><div>Sentiment analysis of code-mixed reviews poses unique challenges due to linguistic variability and contex- tual ambiguity, particularly in multilingual e-commerce environments. In this paper, we introduce BERT- KAN, a novel hybrid architecture that enhances bilingual sentiment analysis in Bangladeshi e-commerce by integrating the deep contextual representations of Bidirectional Encoder Representations from Transform- ers(BERT) with a Kolmogorov-Arnold Network (KAN) layer. The KAN component employs a polynomial expansion to capture complex non-linear relationships within code-mixed Bengali-English text, while an innovative polynomial attention mechanism further refines feature extraction. Extensive ablation studies were conducted on two base models—bert-base-multilingual-uncased and BanglaBERT—using polynomial degrees of 2 and 3. Notably, the best configuration for bert-base-multilingual-uncased (employing KAN, polynomial attention, and feature fusion with polynomial degree 2) achieved a precision of 95.3 %, recall of 97.0 %, and an F1-score of 96.1 %. Comparable performance was observed for polynomial degree 3 (precision 96.2 %, recall 95.8 %, and F1-score 96.0 %), while cross-validation experiments yielded average accuracies exceeding 90 % across multiple folds. Detailed error analyses, supported by confusion matrices and sam- ple predictions, as well as discussions on computational requirements and deployment challenges, further validate the robustness of our approach.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100190"},"PeriodicalIF":0.0,"publicationDate":"2025-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145466719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
OPT2CODE: A retrieval-augmented framework for solving linear programming problems OPT2CODE:一个用于求解线性规划问题的检索增强框架
Pub Date : 2025-10-16 DOI: 10.1016/j.nlp.2025.100185
Tasnim Ahmed, Salimur Choudhury
Mathematical optimization drives decisions across domains such as supply chains, energy grids, and financial systems, among others. Linear programming (LP), a tool for optimizing objectives under constraints, requires domain expertise to translate real-world problems into executable models. We explore automating this translation using Large Language Models (LLMs), generating solver-ready code from textual descriptions to reduce reliance on specialized knowledge. We propose OPT2CODE, a Retrieval-Augmented Generation (RAG) framework that utilizes compact LLMs to transform problem descriptions into optimization solver executable code. OPT2CODE utilizes code documentation for document retrieval and incorporates multiple LLM-as-a-Judge components to improve baseline performance. In addition, OPT2CODE is solver flexible and LLM flexible, and it can generate code for a broad range of mathematical optimization problems such as linear, integer linear, and mixed-integer linear, across different solvers as long as the corresponding solver documentation is available. We show empirical results on two datasets, NL4Opt and EOR, and across two solvers, Gurobi and FICO Xpress, using Llama-3.1-8B and Qwen-2.5-Coder-7B. OPT2CODE consistently improves code generation accuracy, reaching up to 67.13% on NL4Opt with FICO Xpress and 80.00% on EOR with Gurobi. Finally, our energy analysis shows that these improvements come at reasonable computational cost: OPT2CODE consumes 2,732.91 joules/sample (Llama-3.1-8B) and 1,759.95 joules/sample (Qwen-2.5-Coder-7B).
数学优化驱动跨领域的决策,如供应链、能源网络和金融系统等。线性规划(LP)是一种在约束条件下优化目标的工具,它需要领域的专业知识来将现实世界的问题转化为可执行的模型。我们探索使用大型语言模型(llm)自动化翻译,从文本描述生成求解器就绪的代码,以减少对专业知识的依赖。我们提出了OPT2CODE,这是一个检索增强生成(RAG)框架,它利用紧凑的llm将问题描述转换为优化求解器可执行代码。OPT2CODE利用代码文档进行文档检索,并结合多个LLM-as-a-Judge组件来提高基准性能。此外,OPT2CODE是灵活的求解器和LLM灵活的,只要有相应的求解器文档,它可以跨不同的求解器为广泛的数学优化问题(如线性,整数线性和混合整数线性)生成代码。我们使用Llama-3.1-8B和Qwen-2.5-Coder-7B,在两个数据集(NL4Opt和EOR)以及两个求解器(Gurobi和FICO express)上展示了实证结果。OPT2CODE不断提高代码生成精度,FICO express在NL4Opt上达到67.13%,在Gurobi上达到80.00%的EOR。最后,我们的能量分析表明,这些改进是在合理的计算成本下实现的:OPT2CODE消耗2,732.91焦耳/样本(Llama-3.1-8B)和1,759.95焦耳/样本(Qwen-2.5-Coder-7B)。
{"title":"OPT2CODE: A retrieval-augmented framework for solving linear programming problems","authors":"Tasnim Ahmed,&nbsp;Salimur Choudhury","doi":"10.1016/j.nlp.2025.100185","DOIUrl":"10.1016/j.nlp.2025.100185","url":null,"abstract":"<div><div>Mathematical optimization drives decisions across domains such as supply chains, energy grids, and financial systems, among others. Linear programming (LP), a tool for optimizing objectives under constraints, requires domain expertise to translate real-world problems into executable models. We explore automating this translation using Large Language Models (LLMs), generating solver-ready code from textual descriptions to reduce reliance on specialized knowledge. We propose OPT2CODE, a Retrieval-Augmented Generation (RAG) framework that utilizes compact LLMs to transform problem descriptions into optimization solver executable code. OPT2CODE utilizes code documentation for document retrieval and incorporates multiple LLM-as-a-Judge components to improve baseline performance. In addition, OPT2CODE is solver flexible and LLM flexible, and it can generate code for a broad range of mathematical optimization problems such as linear, integer linear, and mixed-integer linear, across different solvers as long as the corresponding solver documentation is available. We show empirical results on two datasets, NL4Opt and EOR, and across two solvers, Gurobi and FICO Xpress, using Llama-3.1-8B and Qwen-2.5-Coder-7B. OPT2CODE consistently improves code generation accuracy, reaching up to <span><math><mrow><mn>67.13</mn><mspace></mspace><mo>%</mo></mrow></math></span> on NL4Opt with FICO Xpress and <span><math><mrow><mn>80.00</mn><mspace></mspace><mo>%</mo></mrow></math></span> on EOR with Gurobi. Finally, our energy analysis shows that these improvements come at reasonable computational cost: OPT2CODE consumes 2,732.91 joules/sample (Llama-3.1-8B) and 1,759.95 joules/sample (Qwen-2.5-Coder-7B).</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100185"},"PeriodicalIF":0.0,"publicationDate":"2025-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Trusted knowledge extraction for operations and maintenance intelligence 为运维智能提取可信知识
Pub Date : 2025-10-13 DOI: 10.1016/j.nlp.2025.100187
Kathleen P. Mealey, Jonathan A. Karr Jr, Priscila Saboia Moreira, Paul R. Brenner, Charles F. Vardeman II
Deriving operational intelligence from organizational data repositories is a key challenge due to the dichotomy of data confidentiality vs data integration objectives, as well as the limitations of Natural Language Processing (NLP) tools relative to the specific knowledge structure of domains such as operations and maintenance. In this work, we discuss Knowledge Graph construction and break down the Knowledge Extraction process into its Named Entity Recognition, Coreference Resolution, Named Entity Linking, and Relation Extraction functional components. We then evaluate sixteen NLP tools in concert with or in comparison to the rapidly advancing capabilities of Large Language Models (LLMs). We focus on the operational and maintenance intelligence use case for trusted applications in the aircraft industry. A baseline dataset is derived from a rich public domain US Federal Aviation Administration dataset focused on equipment failures or maintenance requirements. We assess the zero-shot performance of NLP and LLM tools that can be operated within a controlled, confidential environment (no data is sent to third parties). Based on our observation of significant performance limitations, we discuss the challenges related to trusted NLP and LLM tools as well as their Technical Readiness Level for wider use in mission-critical industries such as aviation. We conclude with recommendations to enhance trust and provide our open-source curated dataset to support further baseline testing and evaluation.
由于数据机密性与数据集成目标的二分法,以及自然语言处理(NLP)工具相对于操作和维护等领域的特定知识结构的局限性,从组织数据存储库中获取操作智能是一个关键挑战。在这项工作中,我们讨论了知识图谱的构建,并将知识提取过程分解为命名实体识别、共同参考解析、命名实体链接和关系提取等功能组件。然后,我们评估了16种NLP工具,这些工具与大型语言模型(llm)快速发展的能力相一致或比较。我们专注于飞机工业中可信赖应用的操作和维护智能用例。基线数据集来源于美国联邦航空管理局丰富的公共领域数据集,重点关注设备故障或维护需求。我们评估了NLP和LLM工具的零射击性能,这些工具可以在受控的机密环境中运行(没有数据发送给第三方)。基于我们对显著性能限制的观察,我们讨论了与可信NLP和LLM工具相关的挑战,以及它们在关键任务行业(如航空)中更广泛使用的技术准备水平。最后,我们提出了增强信任的建议,并提供了我们的开源管理数据集,以支持进一步的基线测试和评估。
{"title":"Trusted knowledge extraction for operations and maintenance intelligence","authors":"Kathleen P. Mealey,&nbsp;Jonathan A. Karr Jr,&nbsp;Priscila Saboia Moreira,&nbsp;Paul R. Brenner,&nbsp;Charles F. Vardeman II","doi":"10.1016/j.nlp.2025.100187","DOIUrl":"10.1016/j.nlp.2025.100187","url":null,"abstract":"<div><div>Deriving operational intelligence from organizational data repositories is a key challenge due to the dichotomy of data confidentiality vs data integration objectives, as well as the limitations of Natural Language Processing (NLP) tools relative to the specific knowledge structure of domains such as operations and maintenance. In this work, we discuss Knowledge Graph construction and break down the Knowledge Extraction process into its Named Entity Recognition, Coreference Resolution, Named Entity Linking, and Relation Extraction functional components. We then evaluate sixteen NLP tools in concert with or in comparison to the rapidly advancing capabilities of Large Language Models (LLMs). We focus on the operational and maintenance intelligence use case for trusted applications in the aircraft industry. A baseline dataset is derived from a rich public domain US Federal Aviation Administration dataset focused on equipment failures or maintenance requirements. We assess the zero-shot performance of NLP and LLM tools that can be operated within a controlled, confidential environment (no data is sent to third parties). Based on our observation of significant performance limitations, we discuss the challenges related to trusted NLP and LLM tools as well as their Technical Readiness Level for wider use in mission-critical industries such as aviation. We conclude with recommendations to enhance trust and provide our open-source curated dataset to support further baseline testing and evaluation.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100187"},"PeriodicalIF":0.0,"publicationDate":"2025-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145519628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging large language models to predict unplanned ICU readmissions from electronic health records 利用大型语言模型从电子健康记录中预测非计划的ICU再入院情况
Pub Date : 2025-10-06 DOI: 10.1016/j.nlp.2025.100182
Hoda Helmy , Ahmed Ibrahim , Maryam Arabi , Aamenah Sattar , Ahmed Serag
Unplanned readmissions to Intensive Care Units (ICUs) are associated with increased mortality, higher healthcare costs, and significant strain on limited medical resources. Accurate prediction of readmissions can improve patient outcomes and optimize resource allocation. This study investigates the use of large language models (LLMs) for ICU readmission prediction through both classification and explanation tasks. We compare a general-purpose model (Gemma2B) and a medical-domain model (Apollo2B), both open-source and fine-tuned for this task. The models were evaluated on their ability to classify readmission cases and generate clinically meaningful justifications. Gemma2B outperformed Apollo2B, achieving an AUC of 0.9, along with strong performance in explanatory outputs. Its ability to produce accurate, context-aware explanations without hallucinations underscores the value of fine-tuned general-purpose models in healthcare settings. These findings highlight the promise of interpretable LLMs in critical care and support their integration into clinical workflows to enhance patient safety and reduce the burden of unplanned ICU readmissions.
非计划再入住重症监护病房(icu)与死亡率增加、医疗费用增加以及对有限医疗资源的巨大压力有关。准确预测再入院可以改善患者预后并优化资源分配。本研究通过分类和解释任务探讨了大语言模型(llm)在ICU再入院预测中的应用。我们比较了通用模型(Gemma2B)和医疗领域模型(Apollo2B),它们都是开源的,也都是为这项任务进行了微调。评估模型对再入院病例进行分类的能力,并产生有临床意义的理由。Gemma2B的表现优于Apollo2B,实现了0.9的AUC,并且在解释性输出方面表现出色。它能够产生准确的、情境感知的解释,而不会产生幻觉,这强调了在医疗保健环境中微调通用模型的价值。这些发现强调了可解释的法学硕士在重症监护中的前景,并支持将其整合到临床工作流程中,以提高患者安全并减少计划外ICU再入院的负担。
{"title":"Leveraging large language models to predict unplanned ICU readmissions from electronic health records","authors":"Hoda Helmy ,&nbsp;Ahmed Ibrahim ,&nbsp;Maryam Arabi ,&nbsp;Aamenah Sattar ,&nbsp;Ahmed Serag","doi":"10.1016/j.nlp.2025.100182","DOIUrl":"10.1016/j.nlp.2025.100182","url":null,"abstract":"<div><div>Unplanned readmissions to Intensive Care Units (ICUs) are associated with increased mortality, higher healthcare costs, and significant strain on limited medical resources. Accurate prediction of readmissions can improve patient outcomes and optimize resource allocation. This study investigates the use of large language models (LLMs) for ICU readmission prediction through both classification and explanation tasks. We compare a general-purpose model (Gemma2B) and a medical-domain model (Apollo2B), both open-source and fine-tuned for this task. The models were evaluated on their ability to classify readmission cases and generate clinically meaningful justifications. Gemma2B outperformed Apollo2B, achieving an AUC of 0.9, along with strong performance in explanatory outputs. Its ability to produce accurate, context-aware explanations without hallucinations underscores the value of fine-tuned general-purpose models in healthcare settings. These findings highlight the promise of interpretable LLMs in critical care and support their integration into clinical workflows to enhance patient safety and reduce the burden of unplanned ICU readmissions.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"13 ","pages":"Article 100182"},"PeriodicalIF":0.0,"publicationDate":"2025-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145363596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Natural Language Processing Journal
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1