A multimodal expert system for the intelligent monitoring and maintenance of transformers enhanced by multimodal language large model fine-tuning and digital twins
{"title":"A multimodal expert system for the intelligent monitoring and maintenance of transformers enhanced by multimodal language large model fine-tuning and digital twins","authors":"Xuedong Zhang, Wenlei Sun, Ke Chen, Renben Jiang","doi":"10.1049/cim2.70007","DOIUrl":null,"url":null,"abstract":"<p>The development of multimodal large models and digital twin technology is set to revolutionise the methods of intelligent monitoring and maintenance for transformers. To address the issues of low intelligence level, single application mode, and poor human–machine collaboration in traditional transformer monitoring and maintenance methods, an intelligent monitoring and maintenance digital twin multimodal expert reasoning system, fine-tuned on visual language-based large models, is proposed. The paper explores the modes and methods for implementing intelligent monitoring and maintenance of transformers based on multimodal data, large models, and digital twin technology. A multimodal language large model (MLLM) framework for intelligent transformer maintenance, grounded on the Large Language and Vision Assistant model, has been designed. To enable large models to understand and reason about image annotation areas, an adaptive grid-based positional information processor has been designed. To facilitate the compatibility and learning of large models with transformer Dissolved Gas Analysis data, a heterogeneous modality converter based on the Gram–Schmidt angular field has been developed. For the unified modelling and management of multimodal reasoning and comprehensive resource integration in human–machine dialogue, a central linker based on an identity resolution asset management shell has been designed. Subsequently, a visual-language multimodal dataset for transformer monitoring and maintenance was constructed. Finally, by fine-tuning parameters, a multimodal expert reasoning system for intelligent transformer monitoring and maintenance was developed. This system not only achieves real-time monitoring of the transformer's operational status but also generates maintenance strategies intelligently based on operational conditions. The expert system possesses robust human–machine dialogue capabilities and reasoning generation abilities. This research provides a reference for the deep integration of MLLM and digital twin in industrial scenarios, particularly in the application modes of intelligent operation and maintenance for transformers.</p>","PeriodicalId":33286,"journal":{"name":"IET Collaborative Intelligent Manufacturing","volume":"6 4","pages":""},"PeriodicalIF":2.5000,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cim2.70007","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Collaborative Intelligent Manufacturing","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/cim2.70007","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, INDUSTRIAL","Score":null,"Total":0}
引用次数: 0
Abstract
The development of multimodal large models and digital twin technology is set to revolutionise the methods of intelligent monitoring and maintenance for transformers. To address the issues of low intelligence level, single application mode, and poor human–machine collaboration in traditional transformer monitoring and maintenance methods, an intelligent monitoring and maintenance digital twin multimodal expert reasoning system, fine-tuned on visual language-based large models, is proposed. The paper explores the modes and methods for implementing intelligent monitoring and maintenance of transformers based on multimodal data, large models, and digital twin technology. A multimodal language large model (MLLM) framework for intelligent transformer maintenance, grounded on the Large Language and Vision Assistant model, has been designed. To enable large models to understand and reason about image annotation areas, an adaptive grid-based positional information processor has been designed. To facilitate the compatibility and learning of large models with transformer Dissolved Gas Analysis data, a heterogeneous modality converter based on the Gram–Schmidt angular field has been developed. For the unified modelling and management of multimodal reasoning and comprehensive resource integration in human–machine dialogue, a central linker based on an identity resolution asset management shell has been designed. Subsequently, a visual-language multimodal dataset for transformer monitoring and maintenance was constructed. Finally, by fine-tuning parameters, a multimodal expert reasoning system for intelligent transformer monitoring and maintenance was developed. This system not only achieves real-time monitoring of the transformer's operational status but also generates maintenance strategies intelligently based on operational conditions. The expert system possesses robust human–machine dialogue capabilities and reasoning generation abilities. This research provides a reference for the deep integration of MLLM and digital twin in industrial scenarios, particularly in the application modes of intelligent operation and maintenance for transformers.
期刊介绍:
IET Collaborative Intelligent Manufacturing is a Gold Open Access journal that focuses on the development of efficient and adaptive production and distribution systems. It aims to meet the ever-changing market demands by publishing original research on methodologies and techniques for the application of intelligence, data science, and emerging information and communication technologies in various aspects of manufacturing, such as design, modeling, simulation, planning, and optimization of products, processes, production, and assembly.
The journal is indexed in COMPENDEX (Elsevier), Directory of Open Access Journals (DOAJ), Emerging Sources Citation Index (Clarivate Analytics), INSPEC (IET), SCOPUS (Elsevier) and Web of Science (Clarivate Analytics).