Automated ontology annotation of scientific literature plays a critical role in knowledge management, particularly in fields like biology and biomedicine, where accurate concept tagging can enhance information retrieval, semantic search, and knowledge integration. Traditional models for ontology annotation, such as Recurrent Neural Networks (RNNs) and Bidirectional Gated Recurrent Units (Bi-GRUs), have been effective but limited in handling complex biomedical terminologies and semantic nuances. This study explores the potential of large language models (LLMs), including MPT-7B, Phi, BiomedLM, and Meditron, for improving ontology annotation, specifically with Gene Ontology (GO) concepts. We fine-tuned these models on the CRAFT dataset, assessing their performance in terms of F1 score, semantic similarity, memory usage, and inference speed. Our results show that while Bi-GRU baselines remain competitive in raw accuracy, LLMs offer complementary strengths. LLMs exhibit qualitatively higher semantic consistency in some cases, particularly when handling complex or multi-word ontology terms. However, these observations are exploratory and not statistically verified across all model types. However, resource requirements for LLMs are notably high, raising considerations about computational efficiency. Techniques like parameter-efficient fine-tuning (PEFT) and advanced prompting were explored to address these challenges, demonstrating potential in reducing computational demands while maintaining performance. Our findings suggest that while LLMs offer advantages in annotation accuracy, practical deployment should balance these benefits with resource costs. This research highlights the need for further optimization and domain-specific training to make LLMs a feasible choice for real-world biomedical ontology annotation tasks.
扫码关注我们
求助内容:
应助结果提醒方式:
