{"title":"在合理的低计算资源条件下开发日本医学大语言模型并进行双语评估","authors":"Issey Sukeda","doi":"arxiv-2409.11783","DOIUrl":null,"url":null,"abstract":"The recent success of large language models (LLMs) and the scaling law has\nled to a widespread adoption of larger models. Particularly in the healthcare\nindustry, there is an increasing demand for locally operated LLMs due to\nsecurity concerns. However, the majority of high quality open-source LLMs have\na size of 70B parameters, imposing significant financial burdens on users for\nGPU preparation and operation. To overcome these issues, we present a medical\nadaptation based on the recent 7B models, which enables the operation in low\ncomputational resources. We compare the performance on medical\nquestion-answering benchmarks in two languages (Japanese and English),\ndemonstrating that its scores reach parity with or surpass those of currently\nexisting medical LLMs that are ten times larger. We find that fine-tuning an\nEnglish-centric base model on Japanese medical dataset improves the score in\nboth language, supporting the effect of cross-lingual knowledge transfer. We\nhope that this study will alleviate financial challenges, serving as a stepping\nstone for clinical institutions to practically utilize LLMs locally. Our\nevaluation code is available at\nhttps://huggingface.co/stardust-coder/jmedllm-7b-v1.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"11 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Development and bilingual evaluation of Japanese medical large language model within reasonably low computational resources\",\"authors\":\"Issey Sukeda\",\"doi\":\"arxiv-2409.11783\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The recent success of large language models (LLMs) and the scaling law has\\nled to a widespread adoption of larger models. Particularly in the healthcare\\nindustry, there is an increasing demand for locally operated LLMs due to\\nsecurity concerns. However, the majority of high quality open-source LLMs have\\na size of 70B parameters, imposing significant financial burdens on users for\\nGPU preparation and operation. To overcome these issues, we present a medical\\nadaptation based on the recent 7B models, which enables the operation in low\\ncomputational resources. We compare the performance on medical\\nquestion-answering benchmarks in two languages (Japanese and English),\\ndemonstrating that its scores reach parity with or surpass those of currently\\nexisting medical LLMs that are ten times larger. We find that fine-tuning an\\nEnglish-centric base model on Japanese medical dataset improves the score in\\nboth language, supporting the effect of cross-lingual knowledge transfer. We\\nhope that this study will alleviate financial challenges, serving as a stepping\\nstone for clinical institutions to practically utilize LLMs locally. Our\\nevaluation code is available at\\nhttps://huggingface.co/stardust-coder/jmedllm-7b-v1.\",\"PeriodicalId\":501030,\"journal\":{\"name\":\"arXiv - CS - Computation and Language\",\"volume\":\"11 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computation and Language\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11783\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computation and Language","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11783","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Development and bilingual evaluation of Japanese medical large language model within reasonably low computational resources
The recent success of large language models (LLMs) and the scaling law has
led to a widespread adoption of larger models. Particularly in the healthcare
industry, there is an increasing demand for locally operated LLMs due to
security concerns. However, the majority of high quality open-source LLMs have
a size of 70B parameters, imposing significant financial burdens on users for
GPU preparation and operation. To overcome these issues, we present a medical
adaptation based on the recent 7B models, which enables the operation in low
computational resources. We compare the performance on medical
question-answering benchmarks in two languages (Japanese and English),
demonstrating that its scores reach parity with or surpass those of currently
existing medical LLMs that are ten times larger. We find that fine-tuning an
English-centric base model on Japanese medical dataset improves the score in
both language, supporting the effect of cross-lingual knowledge transfer. We
hope that this study will alleviate financial challenges, serving as a stepping
stone for clinical institutions to practically utilize LLMs locally. Our
evaluation code is available at
https://huggingface.co/stardust-coder/jmedllm-7b-v1.