Rapid and accurate damage assessment of structures is critical for post-earthquake recovery and emergency response. Current evaluations are heavily reliant on on-site visual inspections conducted by engineering experts, which are time-consuming and resource-intensive. To this end, the large vision-language model (VLM) for multitask structural damage assessment chatbot (MT-SDAChat) is developed in this paper. It can perform both image-level and regional-level inference analysis, accurately locating and providing specific information about various structural components and damage locations. With the MT-SDAChat, a two-stage automated assessment framework that transitions from a global perspective to a component-specific perspective is proposed. A dataset containing 3348 image-text pairs of seismic structural damage with multiple attributes has been constructed. Experimental results show that MT-SDAChat performs well in multitask evaluation. It achieves a question-and-answer accuracy of 82.92 % and a localization accuracy of 78.6 %. These results highlight its strong zero-shot capability across various damage assessments in building construction.
扫码关注我们
求助内容:
应助结果提醒方式:
